LocalAI [bot]
8d9a452e4b
⬆️ Update ggerganov/llama.cpp ( #2671 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-28 10:09:01 +02:00
Ettore Di Giacinto
466eb82845
ci: add latest tag for vulkan images
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-28 09:04:33 +02:00
LocalAI [bot]
7e562d10a3
⬆️ Update ggerganov/llama.cpp ( #2652 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-28 01:30:37 +00:00
Ettore Di Giacinto
7b1e792732
deps(llama.cpp): bump to latest, update build variables ( #2669 )
...
* arrow_up: Update ggerganov/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* deps(llama.cpp): update build variables to follow upstream
Update build recipes with https://github.com/ggerganov/llama.cpp/pull/8006
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Disable shared libs by default in llama.cpp
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Disable shared libs in llama.cpp Makefile
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Disable metal embedding for now, until it is tested
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix(mac): explicitly enable metal
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* debug
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix typo
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-27 23:10:04 +02:00
Ettore Di Giacinto
30b883affe
models(gallery): add gemma2-9b-it and gemma2-27b-it ( #2670 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-27 23:08:22 +02:00
Tachi Koma
20ec4d0342
telegram-bot example: Update LocalAI version ( fixes #2638 ) ( #2640 )
...
* Update LocalAI version for telegram-bot example (fixes #2638 )
* Update examples/telegram-bot/docker-compose.yml
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-27 14:35:50 +02:00
Ettore Di Giacinto
a9f8460086
fix: typo in gallery model
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-27 08:54:13 +02:00
LocalAI [bot]
98b3b2b1ab
feat(swagger): update swagger ( #2666 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-26 23:12:53 +00:00
Ettore Di Giacinto
e8bc0a789b
models(gallery): add arcee-spark ( #2665 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-26 20:06:40 +02:00
Ettore Di Giacinto
2b6a2c7dde
models(gallery): add einstein v7 qwen2 ( #2664 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-26 17:52:52 +02:00
Ettore Di Giacinto
c8c8238f9d
models(gallery): add dark-idol-1.2 ( #2663 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-26 17:52:17 +02:00
Ettore Di Giacinto
3eaf59021c
feat(grammar): expose properties_order ( #2662 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-26 14:59:02 +02:00
Ettore Di Giacinto
a8bfb6f9c2
feat(options): add repeat_last_n ( #2660 )
...
feat(options): add repeat_last_n
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-26 14:58:50 +02:00
Ettore Di Giacinto
b783c811db
feat(build): only build llama.cpp relevant targets ( #2659 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-26 14:58:38 +02:00
Ettore Di Giacinto
59af0e77af
feat(ui): allow to select between all the available models in the chat ( #2657 )
...
feat(ui): let the chat to select from all the detected models
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-25 20:22:17 +02:00
Ettore Di Giacinto
5d83c8d3a2
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-25 19:23:58 +02:00
Ettore Di Giacinto
8f968d0341
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-25 19:18:43 +02:00
Ettore Di Giacinto
f93fe30350
ci: vulkan not ready for arm64 yet
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-25 18:11:57 +02:00
Ettore Di Giacinto
784ccf97ba
ci: adjust max-parallel
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-25 15:14:43 +02:00
LocalAI [bot]
a0163dafce
feat(swagger): update swagger ( #2651 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-25 10:50:52 +02:00
Ettore Di Giacinto
f072cb3cd0
fix(cli): remove duplicate alias ( #2654 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-25 10:08:13 +02:00
Ettore Di Giacinto
e84b31935c
feat(vulkan): add vulkan support to the llama.cpp backend ( #2648 )
...
feat(vulkan): add vulkan support to llama.cpp
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-24 20:04:58 +02:00
Ettore Di Giacinto
03b1cf51fd
feat(whisper): add translate option ( #2649 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-24 19:21:22 +02:00
Ettore Di Giacinto
9e6dec0bc4
fix(install.sh): not all systems have nproc
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-24 18:21:20 +02:00
Ettore Di Giacinto
04b01cd62c
ci: put a cap on parallel runs
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-24 18:08:09 +02:00
Ettore Di Giacinto
a181dd0ebc
refactor: gallery inconsistencies ( #2647 )
...
* refactor(gallery): move under core/
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix(unarchive): do not allow symlinks
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-24 17:32:12 +02:00
Ettore Di Giacinto
69206fcd4b
fix(install.sh): move ARCH detection so it works also for mac ( #2646 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-24 10:34:35 +02:00
Ettore Di Giacinto
2c94e15746
fix(install.sh): fix version typo ( #2645 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-24 10:30:17 +02:00
Dave
12513ebae0
rf: centralize base64 image handling ( #2595 )
...
contains simple fixes to warnings and errors, removes a broken / outdated test, runs go mod tidy, and as the actual change, centralizes base64 image handling
Signed-off-by: Dave Lee <dave@gray101.com >
2024-06-24 08:34:36 +02:00
LocalAI [bot]
4156a4f15f
⬆️ Update ggerganov/llama.cpp ( #2632 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-23 22:21:38 +00:00
Ettore Di Giacinto
491bb4f174
Update hermes-2-pro-mistral.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-23 15:17:41 +02:00
Sertaç Özercan
5866fc8ded
chore: fix go.mod module ( #2635 )
...
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
2024-06-23 08:24:36 +00:00
Ettore Di Giacinto
eb4cd78ca6
ci: run master jobs on self-hosted
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-23 10:15:53 +02:00
Ettore Di Giacinto
40ce71855a
ci: disable max-parallelism on master
2024-06-22 23:28:09 +02:00
Ettore Di Giacinto
9c0d0afd09
ci: bump parallel jobs ( #2633 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-22 23:24:46 +02:00
Ettore Di Giacinto
0f9aa1ef91
fix(install.sh): install CUDA toolkit only if CUDA is detected
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-22 12:21:59 +02:00
Ettore Di Giacinto
3ee5ceb9fa
Update kubernetes.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-22 12:16:55 +02:00
Ettore Di Giacinto
1bd72a3be5
Update kubernetes.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-22 12:16:27 +02:00
Ettore Di Giacinto
fbd14118bf
Update kubernetes.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-22 12:14:53 +02:00
Ettore Di Giacinto
515d98b978
Update model-gallery.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-22 12:10:49 +02:00
Ettore Di Giacinto
789cf6c599
Update model-gallery.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-22 12:10:27 +02:00
Ettore Di Giacinto
0bc82d7270
fix(install.sh): properly detect suse distros
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-22 12:08:48 +02:00
Ettore Di Giacinto
9a7ad75bff
docs: update to include installer and update advanced YAML options ( #2631 )
...
* docs: update quickstart and advanced sections
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* docs: improvements
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* examples(kubernete): add nvidia example
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-22 12:00:38 +02:00
Ettore Di Giacinto
9fb3e4040b
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-22 10:29:46 +02:00
Ettore Di Giacinto
070fd1b9da
Update distributed_inferencing.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-22 10:06:09 +02:00
Ettore Di Giacinto
dda5b9f260
Update distributed_inferencing.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-22 10:05:48 +02:00
Ettore Di Giacinto
8d84dd4f88
fix(worker): use dynaload for single binaries ( #2620 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-22 09:33:18 +02:00
Ettore Di Giacinto
f569237a50
feat(oci): support OCI images and Ollama models ( #2628 )
...
* Support specifying oci:// and ollama:// for model URLs
Fixes: https://github.com/mudler/LocalAI/issues/2527
Fixes: https://github.com/mudler/LocalAI/issues/1028
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Lower watcher warnings
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Allow to install ollama models from CLI
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fixup tests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Do not keep file ownership
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Skip test on darwin
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-22 08:17:41 +02:00
LocalAI [bot]
e265a618d9
models(gallery): ⬆️ update checksum ( #2630 )
...
⬆️ Checksum updates in gallery/index.yaml
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-22 04:45:41 +00:00
LocalAI [bot]
533343c84f
⬆️ Update ggerganov/llama.cpp ( #2629 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-22 02:28:06 +00:00
Ettore Di Giacinto
260f2e1d94
fix(install.sh): correctly handle systemd service installation ( #2627 )
...
Fixup install.sh systemd service installation
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-21 23:56:06 +02:00
Ettore Di Giacinto
964732590d
models(gallery): add hermes-2-theta-llama-3-70b ( #2626 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-21 19:41:49 +02:00
LocalAI [bot]
70a2bfe82e
⬆️ Update ggerganov/llama.cpp ( #2617 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-21 06:41:34 +00:00
Ettore Di Giacinto
ba2d969c44
models(gallery): add qwen2-1.5b-ita ( #2615 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-20 20:35:53 +02:00
Ettore Di Giacinto
d3c78cf4d7
models(gallery): add magnum-72b-v1 ( #2614 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-20 20:31:23 +02:00
Ettore Di Giacinto
34afd891a6
models(gallery): add llama3-8b-darkidol-1.1-iq-imatrix ( #2613 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-20 20:30:47 +02:00
Ettore Di Giacinto
d3137775a1
models(gallery): add llama-3-cursedstock-v1.8-8b-iq-imatrix ( #2612 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-20 20:14:48 +02:00
Ettore Di Giacinto
e1772026a1
models(gallery): add llama-3-sec-chat ( #2611 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-20 20:14:03 +02:00
LocalAI [bot]
d0423254dd
⬆️ Update ggerganov/llama.cpp ( #2606 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-20 00:58:40 +00:00
LocalAI [bot]
db0e52ae9d
⬆️ Update docs version mudler/LocalAI ( #2605 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-20 00:05:19 +00:00
LocalAI [bot]
4f030f9cd3
models(gallery): ⬆️ update checksum ( #2607 )
...
⬆️ Checksum updates in gallery/index.yaml
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-19 22:20:17 +02:00
Ettore Di Giacinto
60fb45eb97
models(gallery): add l3-umbral-mind-rp-v1.0-8b-iq-imatrix ( #2608 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-19 22:19:40 +02:00
Rene Leonhardt
43f0688a95
feat: Upgrade to CUDA 12.5 ( #2601 )
...
Signed-off-by: Rene Leonhardt <65483435+reneleonhardt@users.noreply.github.com >
2024-06-19 17:50:49 +02:00
LocalAI [bot]
8142bdc48f
⬆️ Update ggerganov/llama.cpp ( #2603 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-19 00:28:50 +00:00
Ettore Di Giacinto
89a11e15e7
fix(single-binary): bundle ld.so ( #2602 )
...
* debug
* fix copy command/silly muscle memory
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* remove tmate
* Debugging
* Start binary with ld.so if present in libdir
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* small refactor
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-18 22:43:43 +02:00
Ettore Di Giacinto
06de542032
feat(talk): display an informative box, better colors ( #2600 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-18 15:10:01 +02:00
Ettore Di Giacinto
ecbb61cbf4
feat(sd-3): add stablediffusion 3 support ( #2591 )
...
* feat(sd-3): add stablediffusion 3 support
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* deps(diffusers): add sentencepiece
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* models(gallery): add stablediffusion-3
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-18 15:09:39 +02:00
Ettore Di Giacinto
7f13e3a783
docs(models): fixup top message
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-18 08:42:30 +02:00
LocalAI [bot]
c926469b9c
⬆️ Update ggerganov/llama.cpp ( #2594 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-18 03:06:31 +00:00
LocalAI [bot]
c30b57a629
⬆️ Update docs version mudler/LocalAI ( #2593 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-18 01:47:04 +00:00
LocalAI [bot]
2f297979a7
⬆️ Update ggerganov/llama.cpp ( #2587 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-17 15:28:19 +00:00
Ettore Di Giacinto
2437a2769d
models(gallery): add gemma-1.1-7b-it ( #2588 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-17 14:13:27 +02:00
Ettore Di Giacinto
b58b7cad94
models(gallery): add samantha-qwen2 ( #2586 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-17 10:08:29 +02:00
LocalAI [bot]
68148f2a1a
⬆️ Update ggerganov/llama.cpp ( #2584 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-17 00:18:44 +00:00
Ettore Di Giacinto
4897eb0ba2
ci: pack less libs inside the binary ( #2579 )
...
The binary grew up to 1.8GB quickly - rocm at least raises +800MB by
itself - so we might just want to manage the GPU libs separately.
Adds a comment to list all the libraries found so far that we are
depending on, but will likely follow up in a way to bundle these
separately.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-16 22:10:28 +02:00
Ettore Di Giacinto
1b43966c48
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-16 20:27:37 +02:00
Ettore Di Giacinto
c5f2f11503
models(gallery): add hathor_stable-v0.2-l3-8b ( #2582 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-16 20:24:36 +02:00
Ettore Di Giacinto
895443d1b5
models(gallery): add tess-v2.5-phi-3-medium-128k-14b ( #2581 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-16 20:22:08 +02:00
Ettore Di Giacinto
6a0802e8e6
models(gallery): add dolphin-qwen ( #2580 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-16 20:11:21 +02:00
Ettore Di Giacinto
94cfaad7f4
feat(libpath): refactor and expose functions for external library paths ( #2578 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-16 13:58:28 +02:00
Ettore Di Giacinto
ac4a94dd44
feat(build): bundle libs for arm64 and x86 linux binaries ( #2572 )
...
This PR bundles further libs into the arm64 and x86_64 binaries
This can be improved by a lot - it's far from perfect, however in this PR I wanted to collect the required libs, and give a simple baseline to improve later upon. It is quite challenging to do this exercise with CI only - but it's the fastest way I see now.
I hope that after the list is initially built we can further improve this down the line and remove some of the technical debt left here to speedup things and do not get stuck in the middle of CI cycles.
In this PR:
- The x86_64 binary now bundles hipblas, nvidia and intel libraries too to avoid any dependency to be installed in the host
- Similarly, for the arm64 we now bundle all the required assets
## What's left
We should be also able to cross-compile Nvidia for arm64 - however I didn't succeed so far so I've left that open. Similarly I might have missed some libraries, but we will see with bug reports and testing around with the new binaries. I've tested on my arm64 board and I could finally start things up.
An open point still is shipping libraries for e.g. tts and stablediffusion. this is not done yet, however with the same methodology we should be able to extend support also for these two backends in the binary.
2024-06-16 09:10:44 +02:00
LocalAI [bot]
58bf8614d9
⬆️ Update ggerganov/llama.cpp ( #2575 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-15 23:45:10 +00:00
Ettore Di Giacinto
3764e50b35
models(gallery): add firefly-gemma-7b ( #2576 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-15 23:07:20 +02:00
Nate Harris
3f464d2d9e
Fix standard image latest Docker tags ( #2574 )
...
- Fix standard image latest Docker tags
Signed-off-by: Nate Harris <nwithan8@users.noreply.github.com >
2024-06-15 22:08:30 +02:00
LocalAI [bot]
5116d561e1
⬆️ Update ggerganov/llama.cpp ( #2570 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-14 23:39:20 +00:00
Ettore Di Giacinto
96a7a3b59f
fix(Makefile): enable STATIC on dist ( #2569 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-14 12:28:46 +02:00
Ettore Di Giacinto
112d0ffa45
feat(darwin): embed grpc libs ( #2567 )
...
* debug
* feat(makefile): allow to bundle libs into binary
* ci: bundle protobuf into single-binary
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci: tests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix(assets): correctly reference extract folder
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* bundle also abseil
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* bundle more libs
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-14 08:51:25 +02:00
LocalAI [bot]
25f45827ab
⬆️ Update ggerganov/whisper.cpp ( #2565 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-14 00:26:51 +00:00
LocalAI [bot]
f322f7c62d
⬆️ Update ggerganov/llama.cpp ( #2564 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-13 23:47:50 +00:00
Ettore Di Giacinto
06351cbbb4
feat(binary): support extracted bundled libs on darwin ( #2563 )
...
When offering fallback libs, use the proper env var for darwin
Note: this does not include the libraries itself, but only sets the
proper env var for the libs to be picked up on darwin.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-13 22:59:42 +02:00
Ettore Di Giacinto
8f952d90b0
feat(guesser): identify gemma models ( #2561 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-13 19:12:37 +02:00
Ettore Di Giacinto
7b205510f9
feat(gallery): uniform download from CLI ( #2559 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-13 16:12:46 +02:00
LocalAI [bot]
f183fec232
⬆️ Update ggerganov/llama.cpp ( #2554 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-13 08:34:32 +00:00
Ettore Di Giacinto
91f48b2143
docs(gallery): lazy-load images ( #2557 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-13 01:05:24 +02:00
Ettore Di Giacinto
f404580256
docs: bump go version
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-13 00:49:51 +02:00
Ettore Di Giacinto
882556d4db
feat(gallery): show available models in website, allow local-ai models install to install from galleries ( #2555 )
...
* WIP
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* gen a static page instead (we force DNS redirects to it)
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(gallery): install models from CLI, unify install
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Uniform graphic of model page
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Makefile: update targets
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Slightly enhance gallery view
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-13 00:47:16 +02:00
LocalAI [bot]
f8382adbf7
⬆️ Update ggerganov/llama.cpp ( #2551 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-12 08:54:00 +00:00
LocalAI [bot]
80298f94fa
⬆️ Update ggerganov/whisper.cpp ( #2552 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-12 07:39:21 +00:00
Ettore Di Giacinto
0f8b489346
models(gallery): add badger-lambda-llama-3-8b ( #2550 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-11 19:11:42 +02:00
Ettore Di Giacinto
154694462e
models(gallery): add duloxetine ( #2549 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-11 19:06:26 +02:00
Ettore Di Giacinto
347317d5d2
models(gallery): add average_normie_v3.69_8b-iq-imatrix ( #2548 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-11 19:05:27 +02:00
Ettore Di Giacinto
d40722d2fa
models(gallery): add llama-salad-8x8b ( #2547 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-11 18:40:16 +02:00
Ettore Di Giacinto
7b12300f15
models(gallery): add l3-aethora-15b ( #2546 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-11 18:31:13 +02:00
Ettore Di Giacinto
3c50abffdd
models(gallery): add hathor-l3-8b-v.01-iq-imatrix ( #2545 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-11 16:37:27 +02:00
Ettore Di Giacinto
2eb2ed84ab
models(gallery): add llama3-8B-aifeifei-1.2-iq-imatrix ( #2544 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-11 10:54:21 +02:00
LocalAI [bot]
5da10fb769
⬆️ Update ggerganov/llama.cpp ( #2540 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-11 00:59:17 +00:00
LocalAI [bot]
bec883e3ff
⬆️ Update ggerganov/whisper.cpp ( #2539 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-10 23:32:32 +00:00
Ettore Di Giacinto
14b41be057
feat(detection): detect by template in gguf file, add qwen2, phi, mistral and chatml ( #2536 )
...
feat(detection): detect by template in gguf file, add qwen and chatml
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-10 22:58:04 +02:00
reid41
aff2acacf9
Add integrations ( #2535 )
...
* update integrations
* update integrations1
2024-06-10 19:18:47 +02:00
Rene Leonhardt
b4d4c0a18f
chore(deps): Update Dockerfile ( #2532 )
...
Signed-off-by: Rene Leonhardt <65483435+reneleonhardt@users.noreply.github.com >
2024-06-10 08:40:02 +00:00
LocalAI [bot]
3a5f2283ea
⬆️ Update ggerganov/llama.cpp ( #2531 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-09 23:15:59 +00:00
Ettore Di Giacinto
d9109ffafb
feat(defaults): add defaults for Command-R models ( #2529 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-09 20:00:16 +02:00
Ettore Di Giacinto
d7e137295a
feat(util): add util command to print GGUF informations ( #2528 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-09 19:27:42 +02:00
Ettore Di Giacinto
6c087ae743
feat(arm64): enable single-binary builds ( #2490 )
...
* ci: try to build for arm64
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Allow to skip hipblas on make dist
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* use arm64 cross compiler
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* correctly target go arm64
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* create a separate target
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* cross-compile grpc
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add Protobuf include dirs
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* temp disable CUDA build
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* aarch64 builds: Reduce backends
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Even less backends
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Even less backends
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(startup): allow to load libs from extracted assets
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* makefile: set arch
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-09 15:11:37 +02:00
LocalAI [bot]
88af1033d6
⬆️ Update ggerganov/llama.cpp ( #2524 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-08 23:53:35 +00:00
Ettore Di Giacinto
e96d2d7667
feat(ui): add page to talk with voice, transcription, and tts ( #2520 )
...
* feat(ui): add page to talk with voice, transcription, and tts
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Enhance graphics and status reporting
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Better UX by blocking unvalid actions
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-09 00:03:26 +02:00
Ettore Di Giacinto
aae7ad9d73
feat(llama.cpp): guess model defaults from file ( #2522 )
...
* wip: guess informations from gguf file
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* update go mod
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Small fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Identify llama3
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Do not try to guess the name, as reading gguf files can be expensive
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Allow to disable guessing
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-08 22:13:02 +02:00
LocalAI [bot]
23b3d22525
⬆️ Update ggerganov/llama.cpp ( #2518 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-07 23:35:16 +00:00
Ettore Di Giacinto
603d81dda1
feat(install): add install.sh for quick installs ( #2489 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-07 22:30:41 +02:00
LocalAI [bot]
a21a52d384
models(gallery): ⬆️ update checksum ( #2519 )
...
⬆️ Checksum updates in gallery/index.yaml
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-07 22:17:25 +02:00
Dave
219078a5e0
test: e2e /reranker endpoint ( #2211 )
...
Create a simple e2e test for the /reranker api \\ go mod tidy
Signed-off-by: Dave Lee <dave@gray101.com >
2024-06-07 18:45:52 +00:00
Ettore Di Giacinto
3b7a78adda
fix(stream): do not break channel consumption ( #2517 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-07 17:20:42 +02:00
Sertaç Özercan
0d62594099
fix: fix chat webui response parsing ( #2515 )
...
fix: fix chat webui
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
2024-06-07 17:20:31 +02:00
Dave
d38e9090df
experiment: -j4 for build-linux: ( #2514 )
...
experiment: set -j4 to see if things go faster, while we wait for a proper fix from mudler
Signed-off-by: Dave Lee <dave@gray101.com >
2024-06-07 11:22:28 +02:00
Ettore Di Giacinto
b049805c9b
ci: run release build on self-hosted runners ( #2505 )
2024-06-06 22:16:34 -04:00
LocalAI [bot]
0f9b58f2cf
⬆️ Update ggerganov/llama.cpp ( #2508 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-06 23:48:17 +00:00
LocalAI [bot]
0f134d557e
⬆️ Update ggerganov/whisper.cpp ( #2507 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-06 23:21:25 +00:00
Ettore Di Giacinto
2676e127ae
models(gallery): add llama3-8b-feifei-1.0-iq-imatrix ( #2511 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-07 00:17:59 +02:00
Ettore Di Giacinto
270d4f8413
models(gallery): add rawr_llama3_8b-iq-imatrix ( #2510 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-07 00:12:11 +02:00
Ettore Di Giacinto
2d79cee8cb
models(gallery): add llama3-8B-aifeifei-1.0-iq-imatrix ( #2509 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-07 00:07:20 +02:00
Ettore Di Giacinto
4c9623f50d
deps(whisper): update, add libcufft-dev ( #2501 )
...
* arrow_up: Update ggerganov/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
* fix(build): add libcufft-dev
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-06 08:41:04 +02:00
Ettore Di Giacinto
596cf76135
build(intel): bundle intel variants in single-binary ( #2494 )
...
* wip: try to build also intel variants
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add dependencies
* Select automatically intel backend
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-06 08:40:51 +02:00
LocalAI [bot]
a293aa1b79
⬆️ Update ggerganov/llama.cpp ( #2493 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-06 00:02:51 +00:00
Ettore Di Giacinto
c4eb02c80f
models(gallery): add l3-8b-stheno-v3.2-iq-imatrix ( #2500 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-05 23:46:59 +02:00
Ettore Di Giacinto
9c9198ff08
models(gallery): add Llama-3-Yggdrasil-2.0-8B ( #2499 )
...
models(gallery): add Llama-3-Yggdrasil-2.0-8B-GGUF
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-05 23:42:23 +02:00
Ettore Di Giacinto
83c79d5453
models(gallery): add llama-3-instruct-8b-SimPO-ExPO ( #2498 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-05 23:37:59 +02:00
Ettore Di Giacinto
88fd000065
models(gallery): add phi-3-4x4b ( #2497 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-05 23:29:15 +02:00
Ettore Di Giacinto
956d652314
models(gallery): add nyun ( #2496 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-05 23:22:58 +02:00
Ettore Di Giacinto
9ce2b4d71f
models(gallery): add dolphin-2.9.2-phi-3-Medium-abliterated ( #2495 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-05 23:14:43 +02:00
Ettore Di Giacinto
4e974cb4fc
models(gallery): add dolphin-2.9.2-Phi-3-Medium ( #2492 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-05 19:17:20 +02:00
Dave
d072835796
feat:OpaqueErrors to hide error information ( #2486 )
...
* adds a new configuration option to hide all error message information from http requests
---------
Signed-off-by: Dave Lee <dave@gray101.com >
2024-06-05 08:45:24 +02:00
Ettore Di Giacinto
17cf6c4a4d
feat(amdgpu): try to build in single binary ( #2485 )
...
* feat(amdgpu): try to build in single binary
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Release space from worker
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-05 08:44:15 +02:00
LocalAI [bot]
fab3e711ff
⬆️ Update ggerganov/llama.cpp ( #2487 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-04 23:11:28 +00:00
Dave
4e1463fec2
feat: fiber CSRF ( #2482 )
...
new config option - enables or disables the fiber csrf middleware
Signed-off-by: Dave Lee <dave@gray101.com >
2024-06-04 19:43:46 +00:00
Dave
2fc6fe806b
fix: pkg/downloader should respect basePath for file:// urls ( #2481 )
...
* pass basePath down to pkg/downloader
Signed-off-by: Dave Lee <dave@gray101.com >
* enforce
Signed-off-by: Dave Lee <dave@gray101.com >
---------
Signed-off-by: Dave Lee <dave@gray101.com >
2024-06-04 14:32:47 +00:00
Ettore Di Giacinto
bdd6769b2d
feat(default): use number of physical cores as default ( #2483 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-04 15:23:29 +02:00
Ettore Di Giacinto
1ffee9989f
README: update sponsors list ( #2476 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-04 15:23:00 +02:00
Dave
34ab442ce9
toil: bump grpc version ( #2480 )
...
bump the grpc package version
---------
Signed-off-by: Dave Lee <dave@gray101.com >
2024-06-04 08:39:19 +02:00
LocalAI [bot]
67aa31faad
⬆️ Update ggerganov/llama.cpp ( #2477 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-03 23:09:24 +00:00
fakezeta
6ef78ef7f6
bugfix: CUDA acceleration not working ( #2475 )
...
* bugfix: CUDA acceleration not working
CUDA not working after #2286 .
Refactored the code to be more polish
* Update requirements.txt
Missing imports
Signed-off-by: fakezeta <fakezeta@gmail.com >
* Update requirements.txt
Signed-off-by: fakezeta <fakezeta@gmail.com >
---------
Signed-off-by: fakezeta <fakezeta@gmail.com >
2024-06-03 22:41:42 +02:00
Ettore Di Giacinto
daa7544d9c
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-03 19:55:01 +02:00
Ettore Di Giacinto
34527737bb
feat(webui): enhance card visibility ( #2473 )
...
Do not let the description text to clutter, also highlight the model
names
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-03 17:07:26 +02:00
Ettore Di Giacinto
148adebe16
docs: fix p2p commands ( #2472 )
...
Also change icons on GPT vision page
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-03 16:58:53 +02:00
Ettore Di Giacinto
bae2a649fd
models(gallery): add new poppy porpoise versions ( #2471 )
...
models(gallery): add new poppy purpoise versions
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-03 15:44:52 +02:00
Ettore Di Giacinto
90945ebab3
models(gallery): add fimbulvetr iqmatrix version ( #2470 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-03 15:44:27 +02:00
fakezeta
4a239a4bff
feat(transformers): various enhancements to the transformers backend ( #2468 )
...
update transformers
*Handle Temperature = 0 as greedy search
*Handle custom works as stop words
*Implement KV cache
*Phi 3 no more requires trust_remote_code: true
2024-06-03 08:52:55 +02:00
LocalAI [bot]
5ddaa19914
⬆️ Update ggerganov/llama.cpp ( #2467 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-02 21:34:29 +00:00
Ettore Di Giacinto
77d752a481
fix(gemma): correctly format the template
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-02 10:51:58 +02:00
Ettore Di Giacinto
29ff51c12a
Update gemma stopwords
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-02 01:26:41 +02:00
Ettore Di Giacinto
c0744899c9
models(gallery): add gemma-2b ( #2466 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-02 01:15:06 +02:00
LocalAI [bot]
c9092ad39c
models(gallery): ⬆️ update checksum ( #2463 )
...
⬆️ Checksum updates in gallery/index.yaml
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-01 23:13:02 +00:00
LocalAI [bot]
b588cae70e
⬆️ Update ggerganov/llama.cpp ( #2465 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-01 22:31:32 +00:00
LocalAI [bot]
fb0f188c93
feat(swagger): update swagger ( #2464 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-01 22:04:01 +00:00
Chakib Benziane
b99182c8d4
TTS API improvements ( #2308 )
...
* update doc on COQUI_LANGUAGE env variable
Signed-off-by: blob42 <contact@blob42.xyz >
* return errors from tts gRPC backend
Signed-off-by: blob42 <contact@blob42.xyz >
* handle speaker_id and language in coqui TTS backend
Signed-off-by: blob42 <contact@blob42.xyz >
* TTS endpoint: add optional language paramter
Signed-off-by: blob42 <contact@blob42.xyz >
* tts fix: empty language string breaks non-multilingual models
Signed-off-by: blob42 <contact@blob42.xyz >
* allow tts param definition in config file
- consolidate TTS options under `tts` config entry
Signed-off-by: blob42 <contact@blob42.xyz >
* tts: update doc
Signed-off-by: blob42 <contact@blob42.xyz >
---------
Signed-off-by: blob42 <contact@blob42.xyz >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-06-01 18:26:27 +00:00
Ettore Di Giacinto
95c65d67f5
models(gallery): add all whisper variants ( #2462 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-01 20:04:03 +02:00
Ettore Di Giacinto
c603b95ac7
ci: pin build-time protoc ( #2461 )
...
ci: pin protoc
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-01 18:59:15 +02:00
Ettore Di Giacinto
13cfa6de0a
models(gallery): add Neural SOVLish Devil ( #2460 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-01 12:54:58 +02:00
Ettore Di Giacinto
0560c6fd57
models(gallery): add poppy porpoise 1.0 ( #2459 )
...
modekls(gallery): add poppy porpoise 1.0
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-01 12:54:37 +02:00
Ettore Di Giacinto
f24dddae42
models(gallery): add ultron ( #2456 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-01 00:09:51 +02:00
LocalAI [bot]
06b461b061
⬆️ Update ggerganov/llama.cpp ( #2453 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-06-01 00:09:26 +02:00
Ettore Di Giacinto
e50a7ba879
models(gallery): add llama3-11b ( #2455 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-01 00:03:57 +02:00
Ettore Di Giacinto
3b2bce1fc9
models(gallery): add anjir ( #2454 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-06-01 00:03:46 +02:00
LocalAI [bot]
3fe7e9f678
⬆️ Update ggerganov/whisper.cpp ( #2452 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-31 21:59:48 +00:00
LocalAI [bot]
654b661688
models(gallery): ⬆️ update checksum ( #2451 )
...
⬆️ Checksum updates in gallery/index.yaml
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-31 21:58:54 +00:00
Ettore Di Giacinto
7f387fb238
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-31 22:59:51 +02:00
Ettore Di Giacinto
5d31e5269d
feat(functions): allow response_regex to be a list ( #2447 )
...
feat(functions): allow regex match to be a list
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-31 22:52:02 +02:00
Ettore Di Giacinto
ff8a6962cd
build(Makefile): add back single target to build native llama-cpp ( #2448 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-31 18:35:33 +02:00
Ettore Di Giacinto
10c64dbb55
models(gallery): add mopeymule ( #2449 )
...
* models(gallery): add mopeymule
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci: try to fix workflow
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-31 18:08:39 +02:00
Ettore Di Giacinto
3f7212c660
feat(functions): better free string matching, allow to expect strings after JSON ( #2445 )
...
Allow now any non-character, both as suffix and prefix when mixed grammars are enabled
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-31 09:36:27 +02:00
LocalAI [bot]
5dc6bace49
⬆️ Update ggerganov/whisper.cpp ( #2443 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-30 22:18:55 +00:00
LocalAI [bot]
3cd5918ae6
⬆️ Update ggerganov/llama.cpp ( #2444 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-30 22:09:42 +00:00
Ettore Di Giacinto
5b75bf16c7
models(gallery): add Codestral ( #2442 )
...
models(gallery): add Coderstral
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-30 18:50:26 +02:00
LocalAI [bot]
0c40f545d4
feat(swagger): update swagger ( #2436 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-30 08:11:05 +00:00
LocalAI [bot]
b2fc92daa7
⬆️ Update ggerganov/whisper.cpp ( #2438 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-30 06:07:28 +00:00
LocalAI [bot]
0787797961
⬆️ Update ggerganov/llama.cpp ( #2437 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-29 23:15:36 +00:00
Ettore Di Giacinto
2ba9e27bcf
models(gallery): add neuraldaredevil ( #2439 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-30 00:15:52 +02:00
Prajwal S Nayak
4d98dd9ce7
feat(image): support response_type in the OpenAI API request ( #2347 )
...
* Change response_format type to string to match OpenAI Spec
Signed-off-by: prajwal <prajwalnayak7@gmail.com >
* updated response_type type to interface
Signed-off-by: prajwal <prajwalnayak7@gmail.com >
* feat: correctly parse generic struct
Signed-off-by: mudler <mudler@localai.io >
* add tests
Signed-off-by: mudler <mudler@localai.io >
---------
Signed-off-by: prajwal <prajwalnayak7@gmail.com >
Signed-off-by: mudler <mudler@localai.io >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: mudler <mudler@localai.io >
2024-05-29 14:40:54 +02:00
LocalAI [bot]
087bceccac
⬆️ Update ggerganov/llama.cpp ( #2433 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-28 21:55:03 +00:00
Ettore Di Giacinto
7064697ce5
models(gallery): add halu ( #2434 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-28 23:13:50 +02:00
Ettore Di Giacinto
0b99be73b3
models(gallery): add una-thepitbull ( #2435 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-28 23:13:28 +02:00
Ettore Di Giacinto
669cd06dd9
feat(functions): allow parallel calls with mixed/no grammars ( #2432 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-28 21:06:09 +02:00
Ettore Di Giacinto
2bbc52fcc8
feat(build): add arm64 core containers ( #2421 )
...
ci: add arm64 container images
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-28 10:34:59 +02:00
LocalAI [bot]
577888f3c0
⬆️ Update ggerganov/llama.cpp ( #2428 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-27 22:02:49 +00:00
LocalAI [bot]
1c80f628ff
⬆️ Update ggerganov/whisper.cpp ( #2427 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-27 21:28:36 +00:00
Ettore Di Giacinto
10430a00bd
feat(hipblas): extend default hipblas GPU_TARGETS ( #2426 )
...
Makefile: extend default hipblas GPU_TARGETS
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-27 22:35:11 +02:00
Ettore Di Giacinto
9f5c274321
feat(images): do not install python deps in the core image ( #2425 )
...
do not install python deps in the core image
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-27 22:07:48 +02:00
Ettore Di Giacinto
d075dc44dd
ci: push test images when building PRs ( #2424 )
...
ci: try to push image
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-27 22:07:35 +02:00
Ettore Di Giacinto
be8ffbdfcf
ci(grpc-cache): also arm64 ( #2423 )
...
grpc-cache: also arm64
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-27 17:23:34 +02:00
Ettore Di Giacinto
eaf653f3d3
models(gallery): add iterative-dpo, fix minicpm ( #2422 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-27 17:17:04 +02:00
LocalAI [bot]
e9c28a1ed7
⬆️ Update ggerganov/llama.cpp ( #2419 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-26 21:32:05 +00:00
cryptk
ba984c7097
fix: pin version of setuptools for intel builds to work around #2406 ( #2414 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-05-26 18:27:07 +00:00
Ettore Di Giacinto
ff1f9125ed
models(gallery): add stheno-mahou ( #2418 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-26 20:12:40 +02:00
Ettore Di Giacinto
2c82058548
models(gallery): add cream-phi-13b ( #2417 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-26 20:11:57 +02:00
cryptk
16433d2e8e
fix: install pytorch from proper index for hipblas builds ( #2413 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-05-26 18:05:52 +00:00
Ettore Di Giacinto
345047ed7c
models(gallery): add alpha centauri ( #2416 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-26 20:04:26 +02:00
Ettore Di Giacinto
6343758f9c
models(gallery): add poppy porpoise 0.85 ( #2415 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-26 19:59:49 +02:00
Ettore Di Giacinto
135208806c
models(gallery): add minicpm ( #2412 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-26 15:58:19 +02:00
Ettore Di Giacinto
3280de7adf
models(gallery): add Mahou ( #2411 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-26 15:43:31 +02:00
Ettore Di Giacinto
db3113c5c8
fix(watcher): do not emit fatal errors ( #2410 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-26 14:48:30 +02:00
LocalAI [bot]
593fb62bf0
⬆️ Update ggerganov/llama.cpp ( #2409 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-26 08:43:50 +00:00
LocalAI [bot]
480834f75b
⬆️ Update ggerganov/whisper.cpp ( #2408 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-26 08:05:15 +00:00
Sertaç Özercan
3200a6655e
fix: gpu fetch device info ( #2403 )
...
* fix: gpu fetch device info
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* use pciutils package
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
---------
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
2024-05-26 09:56:06 +02:00
Ettore Di Giacinto
b90cdced59
docs: rewording
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-25 20:18:25 +02:00
Ettore Di Giacinto
fc3502b56f
docs: rewording
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-25 20:17:04 +02:00
Ettore Di Giacinto
785adc1ed5
docs: updaet title
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-25 16:13:48 +02:00
Ettore Di Giacinto
e25fc656c9
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-25 16:13:04 +02:00
Ettore Di Giacinto
bb3ec56de3
docs: add distributed inferencing docs
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-25 16:12:08 +02:00
Ettore Di Giacinto
785c54e7b0
models(gallery): add Mirai Nova ( #2405 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-25 16:11:01 +02:00
Ettore Di Giacinto
003b43f6fc
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-25 10:18:20 +02:00
LocalAI [bot]
663488b6bd
⬆️ Update docs version mudler/LocalAI ( #2398 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-25 10:08:35 +02:00
Ettore Di Giacinto
e1d6b706f4
Update quickstart.md ( #2404 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-25 10:08:23 +02:00
Sertaç Özercan
29615576fb
ci: fix sd release ( #2400 )
...
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
2024-05-25 09:33:50 +02:00
LocalAI [bot]
f8cea16c03
⬆️ Update ggerganov/llama.cpp ( #2399 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-24 21:52:13 +00:00
Ettore Di Giacinto
e0187c2a1a
ci: do not tag latest on AIO automatically
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-24 09:41:13 +02:00
Ettore Di Giacinto
b76d2fe68a
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-24 09:02:59 +02:00
Ettore Di Giacinto
ee4f722bf8
models(gallery): add aya-35b ( #2391 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-23 23:51:34 +02:00
LocalAI [bot]
dce63237f2
⬆️ Update ggerganov/llama.cpp ( #2360 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-23 21:02:13 +00:00
Dave
0b637465d9
refactor: Minor improvements to BackendConfigLoader ( #2353 )
...
some minor renames and refactorings within BackendConfigLoader - make things more consistent, remove underused code, rename things for clarity
Signed-off-by: Dave Lee <dave@gray101.com >
2024-05-23 22:48:12 +02:00
Mauro Morales
114f549f5e
Add warning for running the binary on MacOS ( #2389 )
2024-05-23 22:40:55 +02:00
Ettore Di Giacinto
ea330d452d
models(gallery): add mistral-0.3 and command-r, update functions ( #2388 )
...
* models(gallery): add mistral-0.3 and command-r, update functions
Add also disable_parallel_new_lines to disable newlines in the JSON
output when forcing parallel tools. Some models (like mistral) might be
very sensible to that when being used for function calling.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* models(gallery): add aya-23-8b
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-23 19:16:08 +02:00
Valentin Fröhlich
eb11a46a73
Add Home Assistant Integration ( #2387 )
...
Add https://github.com/valentinfrlch/ha-gpt4vision to Home Assistant Integration section
gpt4vision uses LocalAI's API to send images along with a prompt and return the models output.
Signed-off-by: Valentin Fröhlich <85313672+valentinfrlch@users.noreply.github.com >
2024-05-23 15:21:01 +02:00
LocalAI [bot]
b57e14d65c
models(gallery): ⬆️ update checksum ( #2386 )
...
⬆️ Checksum updates in gallery/index.yaml
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-23 08:42:45 +02:00
Sertaç Özercan
7efa8e75d4
fix: stablediffusion binary ( #2385 )
...
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
2024-05-23 08:34:37 +02:00
Ettore Di Giacinto
7551369abe
Update checksum_checker.sh
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-23 08:33:58 +02:00
LocalAI [bot]
79915bcd11
models(gallery): ⬆️ update checksum ( #2383 )
...
⬆️ Checksum updates in gallery/index.yaml
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-23 01:10:15 +00:00
LocalAI [bot]
c8d7d14a37
⬆️ Update go-skynet/go-bert.cpp ( #1225 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-22 23:42:38 +00:00
LocalAI [bot]
c56bc0de98
⬆️ Update ggerganov/whisper.cpp ( #2361 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-23 01:02:57 +02:00
Ettore Di Giacinto
3a9408363b
deps(llama.cpp): update and adapt API changes ( #2381 )
...
deps(llama.cpp): update and rename function
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-23 01:02:11 +02:00
Ettore Di Giacinto
21a12c2cdd
ci(checksum_checker): do get sha from hf API when available ( #2380 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-22 23:51:02 +02:00
Ettore Di Giacinto
371d0cc1f7
ci: generate specific image for intel builds ( #2374 )
...
ci: fix intel images until are fixed upstream
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-22 23:35:39 +02:00
Ettore Di Giacinto
23fa92bec0
models(gallery): add hercules and helpingAI ( #2376 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-22 22:42:41 +02:00
Ettore Di Giacinto
f91e4e5c03
ci: correctly build p2p in GO_TAGS ( #2369 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-22 10:15:36 +02:00
Ettore Di Giacinto
6cbe6a4f99
models(gallery): add phi-3-medium-4k-instruct ( #2367 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-22 08:32:30 +02:00
Ettore Di Giacinto
491e1d752b
feat(functions): relax mixedgrammars ( #2365 )
...
* feat(functions): relax mixedgrammars
Extend even more the functionalities and when mixed mode is enabled,
tolerate also both strings and JSON in the result - in this case we make
sure that the JSON can be correctly parsed.
This also updates the examples and the gallery model to configure the
grammar.
The changeset also breaks current function/grammar configuration as it
reserves now a stanza in the YAML config.
For example:
```yaml
function:
grammar:
# This allows the grammar to also return messages
mixed_mode: true
# Suffix to add to the grammar
# prefix: '<tool_call>\n'
# Force parallel calls in the grammar
# parallel_calls: true
```
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* refactor, add a way to disable mixed json and freestring
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Fix linting issues
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-22 00:14:16 +02:00
nold
1542c58466
fix(gallery): checksum Meta-Llama-3-70B-Instruct.Q4_K_M.gguf - #2364 ( #2366 )
...
Signed-off-by: Gerrit Pannek <nold@gnu.one >
2024-05-21 21:51:48 +02:00
Ettore Di Giacinto
1a3dedece0
dependencies(grpcio): bump to fix CI issues ( #2362 )
...
feat(grpcio): bump to fix CI issues
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-21 14:33:47 +02:00
Ettore Di Giacinto
a58ff00ab1
models(gallery): add stheno ( #2358 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-20 19:18:14 +02:00
Ettore Di Giacinto
fdb45153fe
feat(llama.cpp): Totally decentralized, private, distributed, p2p inference ( #2343 )
...
* feat(llama.cpp): Enable decentralized, distributed inference
As https://github.com/mudler/LocalAI/pull/2324 introduced distributed inferencing thanks to
@rgerganov implementation in https://github.com/ggerganov/llama.cpp/pull/6829 in upstream llama.cpp, now
it is possible to distribute the workload to remote llama.cpp gRPC server.
This changeset now uses mudler/edgevpn to establish a secure, distributed network between the nodes using a shared token.
The token is generated automatically when starting the server with the `--p2p` flag, and can be used by starting the workers
with `local-ai worker p2p-llama-cpp-rpc` by passing the token via environment variable (TOKEN) or with args (--token).
As per how mudler/edgevpn works, a network is established between the server and the workers with dht and mdns discovery protocols,
the llama.cpp rpc server is automatically started and exposed to the underlying p2p network so the API server can connect on.
When the HTTP server is started, it will discover the workers in the network and automatically create the port-forwards to the service locally.
Then llama.cpp is configured to use the services.
This feature is behind the "p2p" GO_FLAGS
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* go mod tidy
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci: add p2p tag
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* better message
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-20 19:17:59 +02:00
Ettore Di Giacinto
16474bfb40
build: add sha ( #2356 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-20 18:02:19 +02:00
Ettore Di Giacinto
5a6d120a56
feat(functions): don't use yaml.MapSlice ( #2354 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-20 08:31:06 +02:00
Ettore Di Giacinto
7a480bb16f
models(gallery): add LocalAI-Llama3-8b-Function-Call-v0.2-GGUF ( #2355 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-20 00:59:17 +02:00
LocalAI [bot]
053531e434
⬆️ Update ggerganov/whisper.cpp ( #2352 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-19 22:23:02 +00:00
LocalAI [bot]
b7ab4f25d9
⬆️ Update ggerganov/llama.cpp ( #2351 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-19 22:22:03 +00:00
Ettore Di Giacinto
73566a2bb2
feat(functions): allow to use JSONRegexMatch unconditionally ( #2349 )
...
* feat(functions): allow to use JSONRegexMatch unconditionally
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(functions): make json_regex_match a list
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-19 18:24:49 +02:00
Ettore Di Giacinto
8ccd5ab040
feat(webui): statically embed js/css assets ( #2348 )
...
* feat(webui): statically embed js/css assets
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* update font assets
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-19 18:24:27 +02:00
Ettore Di Giacinto
5a3db730b9
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-19 16:37:10 +02:00
Ettore Di Giacinto
8ad669339e
add openvoice backend ( #2334 )
...
Wip openvoice
2024-05-19 16:27:08 +02:00
Ettore Di Giacinto
a10a952085
models(gallery): update poppy porpoise mmproj ( #2346 )
...
models(gallery): update poppy porpose mmproj
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-19 13:26:02 +02:00
Ettore Di Giacinto
b37447cac5
models(gallery): add master-yi ( #2345 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-19 13:25:29 +02:00
Ettore Di Giacinto
f2d182a2eb
models(gallery): add anita ( #2344 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-19 13:25:16 +02:00
lenaxia
6b6c8cdd5f
feat(functions): Enable true regex replacement for the regexReplacement option ( #2341 )
...
* Adding regex capabilities to ParseFunctionCall replacement
Signed-off-by: Lenaxia <github@47north.lat >
* Adding tests for the regex replace in ParseFunctionCall
Signed-off-by: Lenaxia <github@47north.lat >
* Fixing tests and adding a test case to validate double quote replacement works
Signed-off-by: Lenaxia <github@47north.lat >
* Make Regex replacement stable, drop lookaheads
Signed-off-by: mudler <mudler@localai.io >
---------
Signed-off-by: Lenaxia <github@47north.lat >
Signed-off-by: mudler <mudler@localai.io >
Co-authored-by: Lenaxia <github@47north.lat >
Co-authored-by: mudler <mudler@localai.io >
2024-05-19 01:29:10 +02:00
LocalAI [bot]
5f35e85e86
⬆️ Update ggerganov/llama.cpp ( #2342 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-18 21:06:29 +00:00
Ettore Di Giacinto
02f1b477df
feat(functions): simplify parsing, read functions as list ( #2340 )
...
Signed-off-by: mudler <mudler@localai.io >
2024-05-18 09:35:28 +02:00
LocalAI [bot]
9ab8f8f5e0
⬆️ Update ggerganov/llama.cpp ( #2339 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-17 21:13:01 +00:00
LocalAI [bot]
9a255d6453
⬆️ Update ggerganov/llama.cpp ( #2337 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-16 21:53:19 +00:00
Ettore Di Giacinto
e0ef9e2bb9
models(gallery): add yi 6/9b, sqlcoder, sfr-iterative-dpo ( #2335 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-16 20:05:20 +02:00
cryptk
86627b27f7
fix: add setuptools to all requirements-intel.txt files for python backends ( #2333 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-05-16 19:15:46 +02:00
LocalAI [bot]
4e92569d45
⬆️ Update ggerganov/whisper.cpp ( #2329 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-15 22:24:06 +00:00
Ettore Di Giacinto
f7508e3888
models(gallery): add hermes-2-theta-llama-3-8b ( #2331 )
...
Signed-off-by: mudler <mudler@localai.io >
2024-05-16 00:22:32 +02:00
Aleksandr Oleinikov
badfc16df1
fix(gallery) Correct llama3-8b-instruct model file ( #2330 )
...
Correct llama3-8b-instruct model file
This must be a mistake because the config tries to use a model file that is different from the one actually being downloaded.
I assumed the downloaded file is what should be used so I corrected the specified model file to that
Signed-off-by: Aleksandr Oleinikov <10602045+tannisroot@users.noreply.github.com >
2024-05-16 00:22:05 +02:00
LocalAI [bot]
b584dcf18a
⬆️ Update ggerganov/llama.cpp ( #2316 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-15 22:20:37 +00:00
Ettore Di Giacinto
4c845fb47d
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-15 23:56:52 +02:00
Ettore Di Giacinto
07c0559d06
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-15 23:56:22 +02:00
Ettore Di Giacinto
beb598e4f9
feat(functions): mixed JSON BNF grammars ( #2328 )
...
feat(functions): support mixed JSON BNF grammar
This PR provides new options to control how functions are extracted from
the LLM, and also provides more control on how JSON grammars can be used
(also in conjunction).
New YAML settings introduced:
- `grammar_message`: when enabled, the generated grammar can also decide
to push strings and not only JSON objects. This allows the LLM to pick
to either respond freely or using JSON.
- `grammar_prefix`: Allows to prefix a string to the JSON grammar
definition.
- `replace_results`: Is a map that allows to replace strings in the LLM
result.
As an example, consider the following settings for Hermes-2-Pro-Mistral,
which allow extracting both JSON results coming from the model, and the
ones coming from the grammar:
```yaml
function:
# disable injecting the "answer" tool
disable_no_action: true
# This allows the grammar to also return messages
grammar_message: true
# Suffix to add to the grammar
grammar_prefix: '<tool_call>\n'
return_name_in_function_response: true
# Without grammar uncomment the lines below
# Warning: this is relying only on the capability of the
# LLM model to generate the correct function call.
# no_grammar: true
# json_regex_match: "(?s)<tool_call>(.*?)</tool_call>"
replace_results:
"<tool_call>": ""
"\'": "\""
```
Note: To disable entirely grammars usage in the example above, uncomment the
`no_grammar` and `json_regex_match`.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-15 20:03:18 +02:00
Ettore Di Giacinto
c89271b2e4
feat(llama.cpp): add distributed llama.cpp inferencing ( #2324 )
...
* feat(llama.cpp): support distributed llama.cpp
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat: let tweak how chat messages are merged together
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* refactor
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Makefile: register to ALL_GRPC_BACKENDS
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* refactoring, allow disable auto-detection of backends
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* minor fixups
Signed-off-by: mudler <mudler@localai.io >
* feat: add cmd to start rpc-server from llama.cpp
Signed-off-by: mudler <mudler@localai.io >
* ci: add ccache
Signed-off-by: mudler <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: mudler <mudler@localai.io >
2024-05-15 01:17:02 +02:00
Ettore Di Giacinto
29909666c3
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-15 00:33:16 +02:00
LocalAI [bot]
566b5cf2ee
⬆️ Update ggerganov/whisper.cpp ( #2326 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-14 21:17:46 +00:00
Sertaç Özercan
a670318a9f
feat: auto select llama-cpp cuda runtime ( #2306 )
...
* auto select cpu variant
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* remove cuda target for now
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* fix metal
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* fix path
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* cuda
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* auto select cuda
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* update test
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* select CUDA backend only if present
Signed-off-by: mudler <mudler@localai.io >
* ci: keep cuda bin in path
Signed-off-by: mudler <mudler@localai.io >
* Makefile: make dist now builds also cuda
Signed-off-by: mudler <mudler@localai.io >
* Keep pushing fallback in case auto-flagset/nvidia fails
There could be other reasons for which the default binary may fail. For example we might have detected an Nvidia GPU,
however the user might not have the drivers/cuda libraries installed in the system, and so it would fail to start.
We keep the fallback of llama.cpp at the end of the llama.cpp backends to try to fallback loading in case things go wrong
Signed-off-by: mudler <mudler@localai.io >
* Do not build cuda on MacOS
Signed-off-by: mudler <mudler@localai.io >
* cleanup
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* Apply suggestions from code review
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Signed-off-by: mudler <mudler@localai.io >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: mudler <mudler@localai.io >
2024-05-14 19:40:18 +02:00
Ettore Di Giacinto
84e2407afa
feat(functions): allow to set JSON matcher ( #2319 )
...
Signed-off-by: mudler <mudler@localai.io >
2024-05-14 09:39:20 +02:00
Ettore Di Giacinto
c4186f13c3
feat(functions): support models with no grammar and no regex ( #2315 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-14 00:32:32 +02:00
LocalAI [bot]
4ac7956f68
⬆️ Update ggerganov/whisper.cpp ( #2317 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-13 22:25:14 +00:00
Ettore Di Giacinto
e49ea0123b
feat(llama.cpp): add flash_attention and no_kv_offloading ( #2310 )
...
feat(llama.cpp): add flash_attn and no_kv_offload
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-13 19:07:51 +02:00
Ettore Di Giacinto
7123d07456
models(gallery): add orthocopter ( #2313 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-13 18:45:58 +02:00
Ettore Di Giacinto
2db22087ae
models(gallery): add lumimaidv2 ( #2312 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-13 18:44:44 +02:00
Ettore Di Giacinto
fa7b2aee9c
models(gallery): add Bunny-llama ( #2311 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-13 18:44:25 +02:00
Ettore Di Giacinto
4d70b6fb2d
models(gallery): add aura-llama-Abliterated ( #2309 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-13 18:44:10 +02:00
Sertaç Özercan
e2c3ffb09b
feat: auto select llama-cpp cpu variant ( #2305 )
...
* auto select cpu variant
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* remove cuda target for now
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* fix metal
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* fix path
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
---------
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
2024-05-13 11:37:52 +02:00
LocalAI [bot]
b4cb22f444
⬆️ Update ggerganov/llama.cpp ( #2303 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-12 21:18:59 +00:00
LocalAI [bot]
5534b13903
feat(swagger): update swagger ( #2302 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-12 21:00:18 +00:00
fakezeta
5b79bd04a7
add setuptools for openvino ( #2301 )
2024-05-12 19:31:43 +00:00
Ettore Di Giacinto
9d8c705fd9
feat(ui): display number of available models for installation ( #2298 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-12 14:24:36 +02:00
Ettore Di Giacinto
310b2171be
models(gallery): add llama-3-refueled ( #2297 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-12 09:39:58 +02:00
Ettore Di Giacinto
98af0b5d85
models(gallery): add jsl-medllama-3-8b-v2.0 ( #2296 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-12 09:38:05 +02:00
Ettore Di Giacinto
ca14f95d2c
models(gallery): add l3-chaoticsoliloquy-v1.5-4x8b ( #2295 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-12 09:37:55 +02:00
Ikko Eltociear Ashimine
1b69b338c0
docs: Update semantic-todo/README.md ( #2294 )
...
seperate -> separate
Signed-off-by: Ikko Eltociear Ashimine <eltociear@gmail.com >
2024-05-12 09:02:11 +02:00
cryptk
88942e4761
fix: add missing openvino/optimum/etc libraries for Intel, fixes #2289 ( #2292 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-05-12 09:01:45 +02:00
Ettore Di Giacinto
efa32a2677
feat(grammar): support models with specific construct ( #2291 )
...
When enabling grammar with functions, it might be useful to
allow more flexibility to support models that are fine-tuned against returning
function calls of the form of { "name": "function_name", "arguments" {...} }
rather then { "function": "function_name", "arguments": {..} }.
This might call out to a more generic approach later on, but for the moment being we can easily support both
as we have just to specific different types.
If needed we can expand on this later on
Signed-off-by: mudler <mudler@localai.io >
2024-05-12 01:13:22 +02:00
LocalAI [bot]
dfc420706c
⬆️ Update ggerganov/llama.cpp ( #2290 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-11 21:16:34 +00:00
cryptk
e2de8a88f7
feat: create bash library to handle install/run/test of python backends ( #2286 )
...
* feat: create bash library to handle install/run/test of python backends
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* chore: minor cleanup
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: remove incorrect LIMIT_TARGETS from parler-tts
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: update runUnitests to handle running tests from a custom test file
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* chore: document runUnittests
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-05-11 18:32:46 +02:00
Ettore Di Giacinto
7f4febd6c2
models(gallery): add Llama-3-8B-Instruct-abliterated ( #2288 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-11 10:10:57 +02:00
LocalAI [bot]
93e581dfd0
⬆️ Update ggerganov/llama.cpp ( #2285 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-10 21:09:22 +00:00
Ettore Di Giacinto
cf513efa78
Update openai-functions.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-10 17:09:51 +02:00
Ettore Di Giacinto
9e8b34427a
Update openai-functions.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-10 17:05:16 +02:00
Ettore Di Giacinto
88d0aa1e40
docs: update function docs
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-10 17:03:56 +02:00
Ettore Di Giacinto
9b09eb005f
build: do not specify a BUILD_ID by default ( #2284 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-10 16:01:55 +02:00
Ettore Di Giacinto
4db41b71f3
models(gallery): add aloe ( #2283 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-10 16:01:47 +02:00
cryptk
28a421cb1d
feat: migrate python backends from conda to uv ( #2215 )
...
* feat: migrate diffusers backend from conda to uv
- replace conda with UV for diffusers install (prototype for all
extras backends)
- add ability to build docker with one/some/all extras backends
instead of all or nothing
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate autogtpq bark coqui from conda to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: convert exllama over to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate exllama2 to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate mamba to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate parler to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate petals to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: fix tests
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate rerankers to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate sentencetransformers to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: install uv for tests-linux
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: make sure file exists before installing on intel images
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate transformers backend to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate transformers-musicgen to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate vall-e-x to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: migrate vllm to uv
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: add uv install to the rest of test-extra.yml
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: adjust file perms on all install/run/test scripts
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: add missing acclerate dependencies
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: add some more missing dependencies to python backends
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: parler tests venv py dir fix
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: correct filename for transformers-musicgen tests
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: adjust the pwd for valle tests
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: cleanup and optimization work for uv migration
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: add setuptools to requirements-install for mamba
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: more size optimization work
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: make installs and tests more consistent, cleanup some deps
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: cleanup
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: mamba backend is cublas only
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: uncomment lines in makefile
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-05-10 15:08:08 +02:00
LocalAI [bot]
e6768097f4
⬆️ Update docs version mudler/LocalAI ( #2280 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-10 09:10:00 +02:00
LocalAI [bot]
18a04246fa
⬆️ Update ggerganov/llama.cpp ( #2281 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-09 22:18:49 +00:00
LocalAI [bot]
f69de3be0d
models(gallery): ⬆️ update checksum ( #2278 )
...
⬆️ Checksum updates in gallery/index.yaml
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-09 12:21:24 +00:00
Ettore Di Giacinto
650ae620c5
ci: get latest git version
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 11:33:16 +02:00
Ettore Di Giacinto
6a209cbef6
ci: get file name correctly in checksum_checker.sh
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 10:57:23 +02:00
Ettore Di Giacinto
9786bb826d
ci: try to fix checksum_checker.sh
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 09:34:07 +02:00
Ettore Di Giacinto
9b4c6f348a
Update checksum_checker.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 00:57:22 +02:00
Ettore Di Giacinto
cb6ddb21ec
Update checksum_checker.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 00:55:48 +02:00
Ettore Di Giacinto
0baacca605
Update checksum_checker.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 00:54:35 +02:00
Ettore Di Giacinto
222d714ec7
Update checksum_checker.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 00:51:57 +02:00
Ettore Di Giacinto
fd2d89d37b
Update checksum_checker.sh
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 00:43:16 +02:00
Ettore Di Giacinto
6440b608dc
Update checksum_checker.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 00:42:48 +02:00
Ettore Di Giacinto
1937118eab
Update checksum_checker.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-09 00:34:56 +02:00
Ettore Di Giacinto
bc272d1e4b
ci: add checksum checker pipeline ( #2274 )
...
Signed-off-by: mudler <mudler@localai.io >
2024-05-09 00:31:27 +02:00
LocalAI [bot]
d651f390cd
⬆️ Update ggerganov/whisper.cpp ( #2273 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-08 22:11:10 +00:00
Ettore Di Giacinto
ea777f8716
models(gallery): update SHA for einstein
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-08 23:40:58 +02:00
LocalAI [bot]
eca5200fbd
⬆️ Update ggerganov/llama.cpp ( #2272 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-08 21:34:56 +00:00
Ettore Di Giacinto
0809e9e7a0
models(gallery): fix openbiollm typo
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-08 23:19:43 +02:00
LocalAI [bot]
b66baa3db6
⬆️ Update docs version mudler/LocalAI ( #2271 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-08 21:10:30 +00:00
Ettore Di Giacinto
6eb77f0d3a
models(gallery): add tiamat ( #2269 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-08 19:59:42 +02:00
Ettore Di Giacinto
b20354b3ad
models(gallery): add aurora ( #2270 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-08 19:55:39 +02:00
Ettore Di Giacinto
d6f76c75e1
models(gallery): add kunocchini ( #2268 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-08 19:52:08 +02:00
Ettore Di Giacinto
ed4f412f1c
models(gallery): add lumimaid variant ( #2267 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-08 19:51:53 +02:00
Ettore Di Giacinto
5bf56e01aa
models(gallery): add tess ( #2266 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-08 19:51:44 +02:00
Ettore Di Giacinto
5ff5f0b393
fix(ux): fix small glitches ( #2265 )
...
also drop duplicates for displaying in-progress model ops
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-08 19:34:33 +02:00
Ettore Di Giacinto
6559ac11b1
feat(ui): prompt for chat, support vision, enhancements ( #2259 )
...
* feat(ui): allow to set system prompt for chat
Make also the models in the index clickable, and display as table
Fixes #2257
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(vision): support also png with base64 input
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(ui): support vision and upload of files
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* display the processed image
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* make trust remote code stand out
Signed-off-by: mudler <mudler@localai.io >
* feat(ui): track in progress job across index/model gallery
Signed-off-by: mudler <mudler@localai.io >
* minor fixups
Signed-off-by: mudler <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: mudler <mudler@localai.io >
2024-05-08 00:42:34 +02:00
Ettore Di Giacinto
02ec546dd6
models(gallery): Add Soliloquy ( #2260 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-08 00:14:19 +02:00
LocalAI [bot]
995aa5ed21
⬆️ Update ggerganov/llama.cpp ( #2263 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-07 21:39:12 +00:00
Michael Mior
e28ba4b807
Add missing Homebrew dependencies ( #2256 )
...
Signed-off-by: Michael Mior <michael.mior@gmail.com >
Signed-off-by: Michael Mior <mmior@mail.rit.edu >
2024-05-07 16:34:30 +00:00
Daniel
d1e3436de5
Update readme: add ShellOracle to community integrations ( #2254 )
...
Signed-off-by: Daniel Copley <djcopley@users.noreply.github.com >
2024-05-07 08:39:58 +02:00
Dave
d3ddc9e4aa
UI: flag trust_remote_code to users // favicon support ( #2253 )
...
* attempt to indicate trust_remote_code in some way
* bonus: favicon support!
---------
Signed-off-by: Dave Lee <dave@gray101.com >
2024-05-07 08:39:23 +02:00
fakezeta
fea9522982
fix: OpenVINO winograd always disabled ( #2252 )
...
Winograd convolutions were always disabled giving error when inference device was CPU.
This commit implement logic to disable Winograd convolutions only if CPU or NPU are declared.
2024-05-07 08:38:58 +02:00
Ettore Di Giacinto
fe055d4b36
feat(webui): ux improvements ( #2247 )
...
* ux: change welcome when there are no models installed
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ux: filter
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ux: show tags in filter
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* wip
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* make tags clickable
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* allow to delete models from the list
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ui: display icon of installed models
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* gallery: remove gallery file when removing model
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(gallery): show a re-install button
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* make filter buttons, rename Gallery field
Signed-off-by: mudler <mudler@localai.io >
* show again buttons at end of operations
Signed-off-by: mudler <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: mudler <mudler@localai.io >
2024-05-07 01:17:07 +02:00
LocalAI [bot]
581b894789
⬆️ Update ggerganov/llama.cpp ( #2255 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-06 21:28:07 +00:00
Ettore Di Giacinto
477655f6e6
models(gallery): average_norrmie reupload
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-06 19:56:24 +02:00
fakezeta
169d8d21ff
gallery: Added some OpenVINO models ( #2249 )
...
* Added some OpenVINO models
Added Phi-3 trust_remote_code: true
Added Hermes 2 Pro Llama3
Added Multilingual-E5-base embedding model with OpenVINO acceleration (CPU and XPU)
Added all-MiniLM-L6-v2 with OpenVINO acceleration (CPU and XPU)
* Added Remote Code for phi, fixed error on Yamllint
* update openvino.yaml
I need to go to rest: today is not my day...
2024-05-06 10:52:05 +02:00
LocalAI [bot]
c5475020fe
⬆️ Update ggerganov/llama.cpp ( #2251 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-05 21:16:00 +00:00
Dave
b52ff1249f
test: check the response URL during image gen in app_test.go ( #2248 )
...
test: actually check the response URL from image gen
Signed-off-by: Dave Lee <dave@gray101.com >
2024-05-05 18:46:33 +00:00
Ettore Di Giacinto
c5798500cb
feat(single-build): generate single binaries for releases ( #2246 )
...
* feat(single-build): generate single binaries for releases
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* drop old targets
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-05 17:20:51 +02:00
Ettore Di Giacinto
67ad3532ec
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-05 15:45:55 +02:00
Ettore Di Giacinto
5cb96fe7df
models(gallery): add openbiollm ( #2245 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-05 15:19:46 +02:00
Ettore Di Giacinto
810e8e5855
models(gallery): add lumimaid ( #2244 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-05 15:19:33 +02:00
Ettore Di Giacinto
f3bcc648e7
models(gallery): add icon for instruct-coder
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-05 12:20:06 +02:00
Ettore Di Giacinto
3096566333
models(gallery): poppy porpoise fix
...
correct mmproj URL
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-05 11:56:07 +02:00
Ettore Di Giacinto
f50c6a4e88
models(gallery): update poppy porpoise ( #2243 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-05 11:19:09 +02:00
Ettore Di Giacinto
ab4ee54855
models(gallery): add llama3-instruct-coder ( #2242 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-05 11:18:50 +02:00
Ettore Di Giacinto
f2d35062d4
models(gallery): moondream2 fixups
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-05 10:49:04 +02:00
Ettore Di Giacinto
b69ff46c7e
feat(startup): show CPU/GPU information with --debug ( #2241 )
...
Signed-off-by: mudler <mudler@localai.io >
2024-05-05 09:10:23 +02:00
Ettore Di Giacinto
117c9873e1
fix(webui): display small navbar with smaller screens ( #2240 )
...
Signed-off-by: mudler <mudler@localai.io >
2024-05-04 23:38:39 +02:00
LocalAI [bot]
17e94fbcb1
⬆️ Update ggerganov/llama.cpp ( #2239 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-04 21:26:22 +00:00
Ettore Di Giacinto
92f7feb874
models(gallery): add llama3-llava ( #2238 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-04 22:43:11 +02:00
Ettore Di Giacinto
b70e2bffa3
models(gallery): add moondream2 ( #2237 )
...
* models(gallery): add moondream2
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* models(gallery): fix typo for TTS models
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* models(gallery): add base config for moondream2 and icon
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* linter fixes
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-04 18:29:04 +02:00
nold
06c43ca285
fix(gallery): hermes-2-pro-llama3 models checksum changed ( #2236 )
...
fix(gallery): hermes-2-pro-llama3 models checksum
Signed-off-by: Gerrit Pannek <nold@gnu.one >
2024-05-04 17:59:54 +02:00
Ettore Di Giacinto
530bec9c64
feat(llama.cpp): do not specify backends to autoload and add llama.cpp variants ( #2232 )
...
* feat(initializer): do not specify backends to autoload
We can simply try to autoload the backends extracted in the asset dir.
This will allow to build variants of the same backend (for e.g. with different instructions sets),
so to have a single binary for all the variants.
Signed-off-by: mudler <mudler@localai.io >
* refactor(prepare): refactor out llama.cpp prepare steps
Make it so are idempotent and that we can re-build
Signed-off-by: mudler <mudler@localai.io >
* [TEST] feat(build): build noavx version along
Signed-off-by: mudler <mudler@localai.io >
* build: make build parallel
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* build: do not override CMAKE_ARGS
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* build: add fallback variant
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix(huggingface-langchain): fail if no token is set
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix(huggingface-langchain): rename
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: do not autoload local-store
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: give priority between the listed backends
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: mudler <mudler@localai.io >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-04 17:56:12 +02:00
fakezeta
fa10302dd2
docs: updated Transformer parameters description ( #2234 )
...
updated Transformer parameters
2024-05-04 10:45:25 +02:00
Ettore Di Giacinto
54faaa87ea
fix(webui): correct documentation URL for text2img ( #2233 )
...
Signed-off-by: mudler <mudler@localai.io >
Co-authored-by: Dave <dave@gray101.com >
2024-05-04 00:25:13 +00:00
dependabot[bot]
daba8a85f9
build(deps): bump tqdm from 4.65.0 to 4.66.3 in /examples/langchain/langchainpy-localai-example in the pip group across 1 directory ( #2231 )
...
build(deps): bump tqdm
Bumps the pip group with 1 update in the /examples/langchain/langchainpy-localai-example directory: [tqdm](https://github.com/tqdm/tqdm ).
Updates `tqdm` from 4.65.0 to 4.66.3
- [Release notes](https://github.com/tqdm/tqdm/releases )
- [Commits](https://github.com/tqdm/tqdm/compare/v4.65.0...v4.66.3 )
---
updated-dependencies:
- dependency-name: tqdm
dependency-type: direct:production
dependency-group: pip
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-05-03 23:15:06 +00:00
LocalAI [bot]
ac0f3d6e82
⬆️ Update ggerganov/whisper.cpp ( #2230 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-03 22:16:26 +00:00
LocalAI [bot]
da0b6a89ae
⬆️ Update ggerganov/llama.cpp ( #2229 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-03 21:39:28 +00:00
LocalAI [bot]
929a68c06d
⬆️ Update docs version mudler/LocalAI ( #2228 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-03 21:18:11 +00:00
cryptk
a0aa5d01a1
feat: update ROCM and use smaller image ( #2196 )
...
* feat: update ROCM and use smaller image
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: add call to ldconfig to fix AMDs broken library packages
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-05-03 18:46:49 +02:00
Ettore Di Giacinto
dc834cc9d2
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-03 09:42:35 +02:00
Ettore Di Giacinto
b58274b8a2
feat(ui): support multilineand style ul ( #2226 )
...
* feat(ui/chat): handle multiline in the input field
Signed-off-by: mudler <mudler@localai.io >
* feat(ui/chat): correctly display multiline messages
Signed-off-by: mudler <mudler@localai.io >
* feat(ui/chat): add list style
Signed-off-by: mudler <mudler@localai.io >
---------
Signed-off-by: mudler <mudler@localai.io >
2024-05-03 00:43:02 +02:00
Ettore Di Giacinto
a31d00d904
feat(aio): switch to llama3-based for LLM ( #2225 )
...
Signed-off-by: mudler <mudler@localai.io >
2024-05-03 00:41:45 +02:00
LocalAI [bot]
2cc1bd85af
⬆️ Update ggerganov/llama.cpp ( #2224 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-02 21:23:40 +00:00
Ettore Di Giacinto
2c5a46bc34
feat(ux): Add chat, tts, and image-gen pages to the WebUI ( #2222 )
...
* feat(webui): Add chat page
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(webui): Add image-gen page
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(webui): Add tts page
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-02 21:14:10 +02:00
Ettore Di Giacinto
f7f8b4804b
models(gallery): Add Hermes-2-Pro-Llama-3-8B-GGUF ( #2218 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-02 18:31:13 +02:00
Ettore Di Giacinto
e5bd9a76c7
models(gallery): add wizardlm2 ( #2209 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-05-02 18:31:02 +02:00
fakezeta
4690b534e0
feat: user defined inference device for CUDA and OpenVINO ( #2212 )
...
user defined inference device
configuration via main_gpu parameter
2024-05-02 09:54:29 +02:00
LocalAI [bot]
6a7a7996bb
⬆️ Update ggerganov/llama.cpp ( #2213 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-05-01 21:19:44 +00:00
Ettore Di Giacinto
962ebbaf77
models(gallery): fixup phi-3 sha
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-05-01 23:06:58 +02:00
LocalAI [bot]
f90d56d371
⬆️ Update ggerganov/llama.cpp ( #2203 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-30 21:53:31 +00:00
Ettore Di Giacinto
445cfd4db3
models(gallery): add guillaumetell ( #2195 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-30 23:24:41 +02:00
Ettore Di Giacinto
b24d44dc56
models(gallery): add suzume-llama-3-8B-multilingual-gguf ( #2194 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-30 23:24:28 +02:00
Ettore Di Giacinto
cd31f8d865
models(gallery): add lexifun ( #2193 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-30 23:24:13 +02:00
Chris Jowett
970cb3a219
chore: update go-stablediffusion to latest commit with Make jobserver fix
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-30 15:59:28 -05:00
cryptk
f7aabf1b50
fix: bring everything onto the same GRPC version to fix tests ( #2199 )
...
fix: more places where we are installing grpc that need a version specified
fix: attempt to fix metal tests
fix: metal/brew is forcing an update, they don't have 1.58 available anymore
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-30 19:12:15 +00:00
fakezeta
e38610e521
feat: OpenVINO acceleration for embeddings in transformer backend ( #2190 )
...
OpenVINO acceleration for embeddings
New argument type: OVModelForFeatureExtraction
2024-04-30 10:13:04 +02:00
cryptk
3754f154ee
feat: organize Dockerfile into distinct sections ( #2181 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-30 10:12:19 +02:00
LocalAI [bot]
29d7812344
⬆️ Update ggerganov/whisper.cpp ( #2188 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-29 22:16:04 +00:00
cryptk
5fd46175dc
fix: ensure GNUMake jobserver is passed through to whisper.cpp build ( #2187 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-29 16:40:50 -05:00
LocalAI [bot]
52a268c38c
⬆️ Update ggerganov/llama.cpp ( #2189 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-29 21:36:30 +00:00
dependabot[bot]
53c3842bc2
build(deps): bump dependabot/fetch-metadata from 2.0.0 to 2.1.0 ( #2186 )
...
Bumps [dependabot/fetch-metadata](https://github.com/dependabot/fetch-metadata ) from 2.0.0 to 2.1.0.
- [Release notes](https://github.com/dependabot/fetch-metadata/releases )
- [Commits](https://github.com/dependabot/fetch-metadata/compare/v2.0.0...v2.1.0 )
---
updated-dependencies:
- dependency-name: dependabot/fetch-metadata
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-29 21:12:37 +00:00
Dave
c4f958e11b
refactor(application): introduce application global state ( #2072 )
...
* start breaking up the giant channel refactor now that it's better understood - easier to merge bites
Signed-off-by: Dave Lee <dave@gray101.com >
* add concurrency and base64 back in, along with new base64 tests.
Signed-off-by: Dave Lee <dave@gray101.com >
* Automatic rename of whisper.go's Result to TranscriptResult
Signed-off-by: Dave Lee <dave@gray101.com >
* remove pkg/concurrency - significant changes coming in split 2
Signed-off-by: Dave Lee <dave@gray101.com >
* fix comments
Signed-off-by: Dave Lee <dave@gray101.com >
* add list_model service as another low-risk service to get it out of the way
Signed-off-by: Dave Lee <dave@gray101.com >
* split backend config loader into seperate file from the actual config struct. No changes yet, just reduce cognative load with smaller files of logical blocks
Signed-off-by: Dave Lee <dave@gray101.com >
* rename state.go ==> application.go
Signed-off-by: Dave Lee <dave@gray101.com >
* fix lost import?
Signed-off-by: Dave Lee <dave@gray101.com >
---------
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-29 17:42:37 +00:00
Ettore Di Giacinto
147440b39b
docs: add reference for concurrent requests
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-29 18:31:50 +02:00
Ettore Di Giacinto
baff5ff8c2
models(gallery): add openvino models ( #2184 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-29 18:17:47 +02:00
Ettore Di Giacinto
ea13863221
models(gallery): add llama3-32k ( #2183 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-29 18:17:39 +02:00
cryptk
93ca56086e
update go-tinydream to latest commit ( #2182 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-29 15:17:09 +02:00
Dave
11c48a0004
fix: security scanner warning noise: error handlers part 2 ( #2145 )
...
check off a few more error handlers
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-29 15:11:42 +02:00
fakezeta
b7ea9602f5
fix: undefined symbol: iJIT_NotifyEvent in import torch ##2153 ( #2179 )
...
* add extra index to Intel repository
* Update install.sh
2024-04-29 15:11:09 +02:00
Dave
982dc6a2bd
fix: github bump_docs.sh regex to drop emoji and other text ( #2180 )
...
fix: bump_docs regex
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-29 03:55:29 +00:00
Sijia Lu
74d903acca
[Documentations] Removed invalid numberings from troubleshooting mac ( #2174 )
...
* updated troubleshooting mac
Signed-off-by: LeonSijiaLu <leonsijialu1@gmail.com >
* prepend -
Signed-off-by: LeonSijiaLu <leonsijialu1@gmail.com >
---------
Signed-off-by: LeonSijiaLu <leonsijialu1@gmail.com >
2024-04-29 02:21:51 +00:00
LocalAI [bot]
5fef3b0ff1
⬆️ Update ggerganov/whisper.cpp ( #2177 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-28 22:32:45 +00:00
Ettore Di Giacinto
0674893649
Update .env
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-28 23:56:10 +02:00
Ettore Di Giacinto
e8d44447ad
feat(gallery): support model deletion ( #2173 )
...
* feat(gallery): op now supports deletion of models
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Wire things with WebUI(WIP)
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* minor improvements
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-28 23:42:46 +02:00
Ettore Di Giacinto
a24cd4fda0
docs: enhance and condense few sections ( #2178 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-28 23:41:59 +02:00
LocalAI [bot]
01860674c4
⬆️ Update ggerganov/llama.cpp ( #2176 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-28 21:41:12 +00:00
cryptk
987b7ad42d
feat: only keep the build artifacts from the grpc build ( #2172 )
...
* feat: only keep the build artifacts from the grpc build
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: remove separate Cache GRPC build step
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: remove docker inspect step, it is leftover from previous debugging
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-28 19:24:16 +00:00
cryptk
21974fe1d3
fix: swap to WHISPER_CUDA per deprecation message from whisper.cpp ( #2170 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-28 17:51:53 +00:00
Sijia Lu
26e1892521
Issue-1720: Updated Build on mac documentations ( #2171 )
...
updated build on macs documentation
Signed-off-by: LeonSijiaLu <leonsijialu1@gmail.com >
2024-04-28 19:38:02 +02:00
Ettore Di Giacinto
a78cd67737
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-28 19:30:23 +02:00
Ettore Di Giacinto
5e243ceaeb
docs: update gallery, add rerankers ( #2166 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-28 15:54:15 +02:00
QuinnPiers
1a0a6f60a7
docs: update model-gallery.md with correct gallery file ( #2163 )
...
* Update model-gallery.md with correct gallery file
The readme points to a file that hasn't been updated in months so when there are announcements about new models, user's won't get them pointing to the old file. Point to the updated files instead.
Signed-off-by: QuinnPiers <167640194+QuinnPiers@users.noreply.github.com >
* Update model-gallery.md
second pass with more understanding
Signed-off-by: QuinnPiers <167640194+QuinnPiers@users.noreply.github.com >
* Update model-gallery.md
Signed-off-by: QuinnPiers <167640194+QuinnPiers@users.noreply.github.com >
* Update model-gallery.md
Signed-off-by: QuinnPiers <167640194+QuinnPiers@users.noreply.github.com >
---------
Signed-off-by: QuinnPiers <167640194+QuinnPiers@users.noreply.github.com >
2024-04-28 12:34:15 +02:00
Ettore Di Giacinto
3179c019af
Revert " ⬆️ Update docs version mudler/LocalAI" ( #2165 )
...
* Revert "⬆️ Update docs version mudler/LocalAI (#2149 )"
This reverts commit 56d843c263 .
* Apply suggestions from code review
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-28 12:32:49 +02:00
Ettore Di Giacinto
a8089494fd
models(gallery): add biomistral-7b ( #2161 )
...
* models(gallery): add biomistral-7b
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* add <|end_of_text|> to llama3 as stopword
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-28 12:04:04 +02:00
Ettore Di Giacinto
a248ede222
models(gallery): add Undi95/Llama-3-LewdPlay-8B-evo-GGUF ( #2160 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-28 12:02:50 +02:00
Ettore Di Giacinto
0f0ae13ad0
models(gallery): add poppy porpoise ( #2158 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-28 12:01:01 +02:00
Ettore Di Giacinto
773d5d23d5
models(gallery): add solana ( #2157 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-28 11:57:22 +02:00
LocalAI [bot]
c3982212f9
⬆️ Update ggerganov/llama.cpp ( #2159 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-27 21:32:43 +00:00
Ettore Di Giacinto
7e6bf6e7a1
ci: add auto-label rule for gallery in labeler.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-27 19:52:26 +02:00
cryptk
9fc0135991
feat: cleanup Dockerfile and make final image a little smaller ( #2146 )
...
* feat: cleanup Dockerfile and make final image a little smaller
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: add build-essential to final stage
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: more GRPC cache misses
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: correct for another cause of GRPC cache misses
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: generate new GRPC cache automatically if needed
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: use new GRPC_MAKEFLAGS build arg in GRPC cache generation
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-27 19:48:20 +02:00
Ettore Di Giacinto
164be58445
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-27 18:10:58 +02:00
Ettore Di Giacinto
1f8461767d
models(gallery): add average_normie ( #2155 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-27 17:59:31 +02:00
Ettore Di Giacinto
935f4c23f6
models(gallery): add SOVL ( #2154 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-27 17:53:56 +02:00
Ettore Di Giacinto
4c97406f2b
models(gallery): add Einstein v6.1 ( #2152 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-27 12:30:15 +02:00
Ettore Di Giacinto
fb2a05ff43
feat(gallery): display job status also during navigation ( #2151 )
...
* feat(gallery): keep showing progress also when refreshing
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix(intel-gpu): better defaults
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat: make it thread-safe
Signed-off-by: mudler <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: mudler <mudler@localai.io >
2024-04-27 09:08:33 +02:00
LocalAI [bot]
030d555995
⬆️ Update ggerganov/llama.cpp ( #2150 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-27 02:18:28 +00:00
LocalAI [bot]
56d843c263
⬆️ Update docs version mudler/LocalAI ( #2149 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-26 23:03:10 +00:00
Dave
2dc1fa2474
fix: config_file_watcher.go - root all file reads for safety ( #2144 )
...
callHandler() now has all file access rooted within DynamicConfigDir
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-26 16:46:35 +00:00
fakezeta
c9451cb604
Bump oneapi-basekit, optimum and openvino ( #2139 )
...
* Bump oneapi-basekit, optimum and openvino
* Changed PERFORMANCE HINT to CUMULATIVE_THROUGHPUT
Minor latency change for first token but about 10-15% speedup on token generation.
2024-04-26 16:20:43 +02:00
Dave
006306b183
fix: use bluemonday as recommended by blackfriday ( #2142 )
...
use bluemonday as recommended by blackfriday
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-26 10:34:50 +02:00
Dave
2cd4936c99
fix: security scanner warning noise: error handlers part 1 ( #2141 )
...
first group of error handlers to reduce security scanner warning noise level
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-26 10:34:31 +02:00
Dave
44bc540bb5
fix: security scanner dislikes runCommand function arguments ( #2140 )
...
runCommand ==> ffmpegCommand. No functional changes, but makes it clear to the security scanner and future developers that this function cannot run arbitrary commands
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-26 10:33:12 +02:00
Ettore Di Giacinto
6b411ae212
models(gallery): add variants of llama3 70b ( #2138 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-26 00:48:06 +02:00
Dave
eed285f9de
fix: update langchainjs ( #2136 )
...
quick update of the langchainjs example to quiet down some dependency security scanner noise
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-26 00:47:35 +02:00
Dave
c8dd8e5ef4
fix: reduce chmod permissions for created files and directories ( #2137 )
...
quiet more security scanner issues: pass one of chmod restriction to remove group and other permissions
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-26 00:47:06 +02:00
LocalAI [bot]
365ef92530
⬆️ Update mudler/go-stable-diffusion ( #2134 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-25 21:41:38 +00:00
LocalAI [bot]
5fceb876c4
⬆️ Update ggerganov/llama.cpp ( #2133 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-25 21:40:41 +00:00
cryptk
d98063e80e
fix: api key polling was not using correct filepath ( #2132 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-25 20:06:22 +00:00
Dave
45761f8be2
fix: yamlint warnings and errors ( #2131 )
...
fix yamlint warnings and errors
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-25 17:25:56 +00:00
LocalAI [bot]
4ae4e44506
feat(swagger): update swagger ( #2128 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-25 16:10:08 +02:00
Ettore Di Giacinto
2ada13b1ad
models(gallery): add more models ( #2129 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-25 16:06:18 +02:00
Ettore Di Giacinto
5d170e9264
Update yaml-check.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-25 16:05:02 +02:00
Ettore Di Giacinto
1b0a64aa46
Update yaml-check.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-25 15:57:06 +02:00
Ettore Di Giacinto
aa8e1c63d5
Create yaml-check.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-25 15:52:52 +02:00
Ettore Di Giacinto
60690c9fc4
ci: add swagger pipeline
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-25 15:11:01 +02:00
dependabot[bot]
758b0c9042
build(deps): bump pydantic from 1.10.7 to 1.10.13 in /examples/langchain/langchainpy-localai-example in the pip group across 1 directory ( #2125 )
...
build(deps): bump pydantic
Bumps the pip group with 1 update in the /examples/langchain/langchainpy-localai-example directory: [pydantic](https://github.com/pydantic/pydantic ).
Updates `pydantic` from 1.10.7 to 1.10.13
- [Release notes](https://github.com/pydantic/pydantic/releases )
- [Changelog](https://github.com/pydantic/pydantic/blob/main/HISTORY.md )
- [Commits](https://github.com/pydantic/pydantic/compare/v1.10.7...v1.10.13 )
---
updated-dependencies:
- dependency-name: pydantic
dependency-type: direct:production
dependency-group: pip
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-25 06:49:29 +00:00
Ettore Di Giacinto
48d0aa2f6d
models(gallery): add new models to the gallery ( #2124 )
...
* models: add reranker and parler-tts-mini
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: chatml im_end should not have a newline
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* models(noromaid): add
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* models(llama3): add 70b, add dolphin2.9
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* models(llama3): add unholy-8b
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* models(llama3): add therapyllama3, aura
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-25 01:28:02 +02:00
Ettore Di Giacinto
b664edde29
feat(rerankers): Add new backend, support jina rerankers API ( #2121 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-25 00:19:02 +02:00
LocalAI [bot]
e16658b7ec
⬆️ Update ggerganov/llama.cpp ( #2123 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-24 22:00:17 +00:00
LocalAI [bot]
d30280ed23
⬆️ Update ggerganov/whisper.cpp ( #2122 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-24 21:55:30 +00:00
Ettore Di Giacinto
9dbd217c59
docs(integrations): add Wave terminal
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-24 19:56:51 +02:00
Ettore Di Giacinto
23eac98b3c
docs: update hot topics
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-24 19:43:07 +02:00
Ettore Di Giacinto
4fffc47e77
deps(llama.cpp): update, use better model for function call tests ( #2119 )
...
deps(llama.cpp): update
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-24 18:44:04 +02:00
LocalAI [bot]
d65214a234
⬆️ Update docs version mudler/LocalAI ( #2113 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-24 11:11:41 +02:00
jtwolfe
2fb34b00b5
Incl ocv pkg for diffsusers utils ( #2115 )
...
* Update diffusers.yml
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
* Update diffusers-rocm.yml
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
---------
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
2024-04-24 09:17:49 +02:00
fakezeta
f718a391c0
fix missing TrustRemoteCode in OpenVINO model load ( #2114 )
2024-04-24 00:45:37 +00:00
Ettore Di Giacinto
ac56ac2b2d
fix(gallery): show a fake image if no there is no icon ( #2111 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-23 20:10:58 +02:00
Ettore Di Giacinto
34c3f563fd
fix(gallery): fixup dreamshaper icon
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-23 20:05:59 +02:00
Ettore Di Giacinto
d2bea6f9e3
fix(gallery): fixup hermes q8 entry
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-23 20:01:56 +02:00
Ettore Di Giacinto
a09fe1b9ba
fix(gallery): set margin for images
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-23 20:00:20 +02:00
Ettore Di Giacinto
55778b35ff
fix(gallery): move metadata where it belongs
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-23 19:47:45 +02:00
Ettore Di Giacinto
8b169f1dac
feat(gallery): add llama3, hermes, phi-3, and others ( #2110 )
...
Also adds embeddings and llava models
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-23 19:35:45 +02:00
Ettore Di Giacinto
d344daf129
feat(models-ui): minor visual enhancements ( #2109 )
...
Show image if present, URL, tags, and better display buttons
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-23 18:43:25 +02:00
cryptk
3411e072ca
Fix cleanup sonarqube findings ( #2106 )
...
* fix: update dockerignore and gitignore to exclude sonarqube work dir
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: remove useless equality check
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: use sonarqube Dockerfile recommendations
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-23 18:43:00 +02:00
fakezeta
8e36fe9b6f
Transformers Backend: max_tokens adherence to OpenAI API ( #2108 )
...
max token adherence to OpenAI API
improve adherence to OpenAI API when max tokens is omitted or equal to 0 in the request
2024-04-23 18:42:17 +02:00
Ettore Di Giacinto
0d8bf91699
feat: Galleries UI ( #2104 )
...
* WIP: add models to webui
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Register routes
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: don't cache models
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* small fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: fixup multiple installs (strings.Clone)
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-23 09:22:58 +02:00
LocalAI [bot]
bd507678be
⬆️ Update docs version mudler/LocalAI ( #2105 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-22 22:04:57 +00:00
Taikono-Himazin
b6f0e80d54
Update text-generation.md ( #2095 )
...
Signed-off-by: Taikono-Himazin <kazu@po.harenet.ne.jp >
2024-04-22 16:37:13 +02:00
jtwolfe
729378ca98
AMD/ROCm Documentation update + formatting fix ( #2100 )
...
* Update aio-images.md
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
* Update aio-images.md
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
* Update aio-images.md
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
* Update GPU-acceleration.md
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
* Update GPU-acceleration.md
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
* Update GPU-acceleration.md
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
* Update GPU-acceleration.md
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
* Update GPU-acceleration.md
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
* Update GPU-acceleration.md
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
---------
Signed-off-by: jtwolfe <jamie.t.wolfe@gmail.com >
2024-04-22 15:47:51 +02:00
Ikko Eltociear Ashimine
220958a87c
fix: typo in models.go ( #2099 )
2024-04-22 04:34:59 +00:00
Ettore Di Giacinto
f3f6535aad
fix: rename fiber entrypoint from http/api to http/app ( #2096 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: Dave <dave@gray101.com >
2024-04-21 22:39:28 +02:00
Dave
228bc4903f
fix: action-tmate detached ( #2092 )
...
connect-timeout-seconds works best with `detached: true`
Signed-off-by: Dave <dave@gray101.com >
2024-04-21 22:39:17 +02:00
LocalAI [bot]
38c9abed8b
⬆️ Update ggerganov/llama.cpp ( #2089 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-21 16:35:30 +00:00
fakezeta
66b002458d
Transformer Backend: Implementing use_tokenizer_template and stop_prompts options ( #2090 )
...
* fix regression #1971
fixes regression #1971 introduced by intel_extension_for_transformers==1.4
* UseTokenizerTemplate and StopPrompt
Implementation of use_tokenizer_template and stopwords options
2024-04-21 16:20:25 +00:00
Ettore Di Giacinto
39814cab32
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-21 16:46:13 +02:00
Ettore Di Giacinto
180cd4ccda
fix(llama.cpp-ggml): fixup max_tokens for old backend ( #2094 )
...
fix(llama.cpp-ggml): set 0 as default for `max_tokens`
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-21 16:34:00 +02:00
Ettore Di Giacinto
284ad026b1
refactor(routes): split routes registration ( #2077 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-21 01:19:57 +02:00
Ettore Di Giacinto
afa1bca1e3
fix(llama.cpp): set -1 as default for max tokens ( #2087 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-20 20:20:10 +02:00
Taikono-Himazin
03adc1f60d
Add tensor_parallel_size setting to vllm setting items ( #2085 )
...
Signed-off-by: Taikono-Himazin <kazu@po.harenet.ne.jp >
2024-04-20 14:37:02 +00:00
Ettore Di Giacinto
b319ed58b0
models(gallery): add gallery ( #2078 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-20 15:22:54 +02:00
cryptk
8d30b39811
feat: fiber logs with zerlog and add trace level ( #2082 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-20 10:43:37 +02:00
Dave
1038f7469c
fix: action-tmate: use connect-timeout-sections and limit-access-to-actor ( #2083 )
...
fix for action-tmate: connect-timeout-sections and limit-access-to-actor
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-20 08:42:02 +00:00
cryptk
b9e7708643
feat: enable polling configs for systems with broken fsnotify (docker volumes on windows) ( #2081 )
...
* feat: enable polling configs for systems with broken fsnotify (docker volumes on windows)
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: update logging to make it clear that the config file is being polled
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-19 19:31:15 -05:00
LocalAI [bot]
1e37101930
⬆️ Update ggerganov/llama.cpp ( #2080 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-20 00:05:16 +00:00
Ettore Di Giacinto
b2772509b4
models(llama3): add llama3 to embedded models ( #2074 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-19 18:23:44 +02:00
Ettore Di Giacinto
27ec84827c
refactor(template): isolate and add tests ( #2069 )
...
* refactor(template): isolate and add tests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: Dave <dave@gray101.com >
Co-authored-by: Dave <dave@gray101.com >
2024-04-19 02:40:18 +00:00
cryptk
852316c5a6
fix: move the GRPC cache generation workflow into it's own concurrency group ( #2071 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-18 20:52:34 -04:00
LocalAI [bot]
e9448005a5
⬆️ Update ggerganov/llama.cpp ( #2051 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-18 21:30:55 +00:00
Ettore Di Giacinto
bbea62b907
feat(functions): support models with no grammar, add tests ( #2068 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-18 22:43:12 +02:00
cryptk
13012cfa70
feat: better control of GRPC docker cache ( #2070 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-18 16:19:36 -04:00
dependabot[bot]
8f2681f904
build(deps): bump aiohttp from 3.9.2 to 3.9.4 in /examples/langchain/langchainpy-localai-example in the pip group across 1 directory ( #2067 )
...
build(deps): bump aiohttp
Bumps the pip group with 1 update in the /examples/langchain/langchainpy-localai-example directory: [aiohttp](https://github.com/aio-libs/aiohttp ).
Updates `aiohttp` from 3.9.2 to 3.9.4
- [Release notes](https://github.com/aio-libs/aiohttp/releases )
- [Changelog](https://github.com/aio-libs/aiohttp/blob/master/CHANGES.rst )
- [Commits](https://github.com/aio-libs/aiohttp/compare/v3.9.2...v3.9.4 )
---
updated-dependencies:
- dependency-name: aiohttp
dependency-type: direct:production
dependency-group: pip
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-18 17:17:33 +00:00
Ettore Di Giacinto
f9c75d4878
tests: add template tests ( #2063 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-18 10:57:24 +02:00
cryptk
502c1eedaa
feat: refactor the dynamic json configs for api_keys and external_backends ( #2055 )
...
* feat: refactor the dynamic json configs for api_keys and external_backends
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: remove commented code
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-18 03:21:55 +00:00
cryptk
e9f090257c
fix: adjust some sources names to match the naming of their repositories ( #2061 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-18 01:59:05 +00:00
Ettore Di Giacinto
af9e5a2d05
Revert #1963 ( #2056 )
...
* Revert "fix(fncall): fix regression introduced in #1963 (#2048 )"
This reverts commit 6b06d4e0af .
* Revert "fix: action-tmate back to upstream, dead code removal (#2038 )"
This reverts commit fdec8a9d00 .
* Revert "feat(grpc): return consumed token count and update response accordingly (#2035 )"
This reverts commit e843d7df0e .
* Revert "refactor: backend/service split, channel-based llm flow (#1963 )"
This reverts commit eed5706994 .
* feat(grpc): return consumed token count and update response accordingly
Fixes : #1920
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-17 23:33:49 +02:00
LocalAI [bot]
af8c705ecd
⬆️ Update ggerganov/whisper.cpp ( #2060 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-17 21:17:25 +00:00
LocalAI [bot]
5763dc1613
⬆️ Update ggerganov/whisper.cpp ( #2050 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-16 21:37:50 +00:00
Ettore Di Giacinto
6b06d4e0af
fix(fncall): fix regression introduced in #1963 ( #2048 )
...
Signed-off-by: Dave <dave@gray101.com >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: Dave <dave@gray101.com >
Co-authored-by: Dave <dave@gray101.com >
2024-04-16 21:20:11 +00:00
dependabot[bot]
bcaa320f36
build(deps): bump the pip group across 4 directories with 8 updates ( #2049 )
...
Bumps the pip group with 1 update in the /examples/functions directory: [langchain](https://github.com/langchain-ai/langchain ).
Bumps the pip group with 2 updates in the /examples/langchain-chroma directory: [langchain](https://github.com/langchain-ai/langchain ) and [llama-index](https://github.com/run-llama/llama_index ).
Bumps the pip group with 6 updates in the /examples/langchain/langchainpy-localai-example directory:
| Package | From | To |
| --- | --- | --- |
| [langchain](https://github.com/langchain-ai/langchain ) | `0.0.159` | `0.1.0` |
| [aiohttp](https://github.com/aio-libs/aiohttp ) | `3.8.4` | `3.9.2` |
| [certifi](https://github.com/certifi/python-certifi ) | `2022.12.7` | `2023.7.22` |
| [idna](https://github.com/kjd/idna ) | `3.4` | `3.7` |
| [requests](https://github.com/psf/requests ) | `2.29.0` | `2.31.0` |
| [urllib3](https://github.com/urllib3/urllib3 ) | `1.26.15` | `1.26.18` |
Bumps the pip group with 1 update in the /examples/streamlit-bot directory: [streamlit](https://github.com/streamlit/streamlit ).
Updates `langchain` from 0.0.234 to 0.1.0
- [Release notes](https://github.com/langchain-ai/langchain/releases )
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.234...v0.1.0 )
Updates `langchain` from 0.0.160 to 0.1.0
- [Release notes](https://github.com/langchain-ai/langchain/releases )
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.234...v0.1.0 )
Updates `llama-index` from 0.6.2 to 0.9.36
- [Release notes](https://github.com/run-llama/llama_index/releases )
- [Changelog](https://github.com/run-llama/llama_index/blob/main/CHANGELOG.md )
- [Commits](https://github.com/run-llama/llama_index/compare/v0.6.2...v0.9.36 )
Updates `langchain` from 0.0.159 to 0.1.0
- [Release notes](https://github.com/langchain-ai/langchain/releases )
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.234...v0.1.0 )
Updates `aiohttp` from 3.8.4 to 3.9.2
- [Release notes](https://github.com/aio-libs/aiohttp/releases )
- [Changelog](https://github.com/aio-libs/aiohttp/blob/master/CHANGES.rst )
- [Commits](https://github.com/aio-libs/aiohttp/compare/v3.8.4...v3.9.2 )
Updates `certifi` from 2022.12.7 to 2023.7.22
- [Commits](https://github.com/certifi/python-certifi/compare/2022.12.07...2023.07.22 )
Updates `idna` from 3.4 to 3.7
- [Release notes](https://github.com/kjd/idna/releases )
- [Changelog](https://github.com/kjd/idna/blob/master/HISTORY.rst )
- [Commits](https://github.com/kjd/idna/compare/v3.4...v3.7 )
Updates `requests` from 2.29.0 to 2.31.0
- [Release notes](https://github.com/psf/requests/releases )
- [Changelog](https://github.com/psf/requests/blob/main/HISTORY.md )
- [Commits](https://github.com/psf/requests/compare/v2.29.0...v2.31.0 )
Updates `urllib3` from 1.26.15 to 1.26.18
- [Release notes](https://github.com/urllib3/urllib3/releases )
- [Changelog](https://github.com/urllib3/urllib3/blob/main/CHANGES.rst )
- [Commits](https://github.com/urllib3/urllib3/compare/1.26.15...1.26.18 )
Updates `streamlit` from 1.26.0 to 1.30.0
- [Release notes](https://github.com/streamlit/streamlit/releases )
- [Commits](https://github.com/streamlit/streamlit/compare/1.26.0...1.30.0 )
---
updated-dependencies:
- dependency-name: langchain
dependency-type: direct:production
dependency-group: pip
- dependency-name: langchain
dependency-type: direct:production
dependency-group: pip
- dependency-name: llama-index
dependency-type: direct:production
dependency-group: pip
- dependency-name: langchain
dependency-type: direct:production
dependency-group: pip
- dependency-name: aiohttp
dependency-type: direct:production
dependency-group: pip
- dependency-name: certifi
dependency-type: direct:production
dependency-group: pip
- dependency-name: idna
dependency-type: direct:production
dependency-group: pip
- dependency-name: requests
dependency-type: direct:production
dependency-group: pip
- dependency-name: urllib3
dependency-type: direct:production
dependency-group: pip
- dependency-name: streamlit
dependency-type: direct:production
dependency-group: pip
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-16 19:49:54 +00:00
Ettore Di Giacinto
33c78d2228
feat(store): add Golang client ( #1977 )
...
This adds a basic store client for Go
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-16 15:54:14 +02:00
Adrien Brault
df4a13a08b
docs: fix stores link ( #2044 )
...
Signed-off-by: Adrien Brault <adrien.brault@gmail.com >
2024-04-16 11:10:23 +02:00
Dave
fdec8a9d00
fix: action-tmate back to upstream, dead code removal ( #2038 )
...
cleanup: upstream action-tmate has taken my PR, drop master reference. Also remove dead code from api.go
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-16 01:46:36 +00:00
LocalAI [bot]
0cc1ad2188
⬆️ Update ggerganov/whisper.cpp ( #2042 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-15 23:27:52 +00:00
LocalAI [bot]
cdece3879f
⬆️ Update ggerganov/llama.cpp ( #2043 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-15 22:47:29 +00:00
dependabot[bot]
320d8a48d9
build(deps): bump github/codeql-action from 2 to 3 ( #2041 )
...
Bumps [github/codeql-action](https://github.com/github/codeql-action ) from 2 to 3.
- [Release notes](https://github.com/github/codeql-action/releases )
- [Changelog](https://github.com/github/codeql-action/blob/main/CHANGELOG.md )
- [Commits](https://github.com/github/codeql-action/compare/v2...v3 )
---
updated-dependencies:
- dependency-name: github/codeql-action
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-15 22:02:44 +00:00
dependabot[bot]
46609e936e
build(deps): bump dependabot/fetch-metadata from 1.3.4 to 2.0.0 ( #2040 )
...
Bumps [dependabot/fetch-metadata](https://github.com/dependabot/fetch-metadata ) from 1.3.4 to 2.0.0.
- [Release notes](https://github.com/dependabot/fetch-metadata/releases )
- [Commits](https://github.com/dependabot/fetch-metadata/compare/v1.3.4...v2.0.0 )
---
updated-dependencies:
- dependency-name: dependabot/fetch-metadata
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-15 21:37:06 +00:00
dependabot[bot]
b72c6cc9fc
build(deps): bump softprops/action-gh-release from 1 to 2 ( #2039 )
...
Bumps [softprops/action-gh-release](https://github.com/softprops/action-gh-release ) from 1 to 2.
- [Release notes](https://github.com/softprops/action-gh-release/releases )
- [Changelog](https://github.com/softprops/action-gh-release/blob/master/CHANGELOG.md )
- [Commits](https://github.com/softprops/action-gh-release/compare/v1...v2 )
---
updated-dependencies:
- dependency-name: softprops/action-gh-release
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-15 20:52:39 +00:00
cryptk
538a086309
fix: previous CLI rework broke debug logging ( #2036 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
Co-authored-by: Dave <dave@gray101.com >
2024-04-15 22:13:59 +02:00
cryptk
c751a4ac06
fix: remove build path from help text documentation ( #2037 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-15 19:47:51 +00:00
Ettore Di Giacinto
e843d7df0e
feat(grpc): return consumed token count and update response accordingly ( #2035 )
...
Fixes : #1920
2024-04-15 19:47:11 +02:00
LocalAI [bot]
de3a1a0a8e
⬆️ Update ggerganov/llama.cpp ( #2033 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-14 23:35:44 +00:00
LocalAI [bot]
57bd365d87
⬆️ Update docs version mudler/LocalAI ( #2032 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-14 23:31:43 +00:00
Ettore Di Giacinto
b739cbb86b
Revert "build(deps): bump the pip group across 4 directories with 8 updates" ( #2030 )
...
Revert "build(deps): bump the pip group across 4 directories with 8 updates (…"
This reverts commit e0dee52a2a .
2024-04-14 10:57:07 +02:00
Ettore Di Giacinto
4486db912b
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-13 22:57:46 +02:00
Ettore Di Giacinto
6b07ded119
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-13 19:12:54 +02:00
Dave
d5699dbf4f
fix - correct checkout versions ( #2029 )
...
minor fix - bump some checkout@v3 to checkout@v4 to match and clean up warnings
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-13 19:01:17 +02:00
Ettore Di Giacinto
0fdff26924
feat(parler-tts): Add new backend ( #2027 )
...
* feat(parler-tts): Add new backend
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(parler-tts): try downgrade protobuf
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(parler-tts): add parler conda env
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Revert "feat(parler-tts): try downgrade protobuf"
This reverts commit bd5941d5cfc00676b45a99f71debf3c34249cf3c.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* deps: add grpc
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: try to gen proto with same environment
* workaround
* Revert "fix: try to gen proto with same environment"
This reverts commit 998c745e2f .
* Workaround fixup
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: Dave <dave@gray101.com >
2024-04-13 18:59:21 +02:00
LocalAI [bot]
619f2517a4
⬆️ Update ggerganov/llama.cpp ( #2028 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-13 13:47:39 +00:00
Ettore Di Giacinto
b91820b7f8
Update localaibot_automerge.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-13 13:46:07 +02:00
Ettore Di Giacinto
4e74560649
ci: fix release pipeline missing dependencies ( #2025 )
2024-04-13 13:30:40 +02:00
Ettore Di Giacinto
95244ed6e7
Update localaibot_automerge.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-13 10:03:15 +02:00
Ettore Di Giacinto
f1f39eea3f
Create localaibot_automerge.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-13 09:47:33 +02:00
Dave
eed5706994
refactor: backend/service split, channel-based llm flow ( #1963 )
...
Refactor: channel based llm flow and services split
---------
Signed-off-by: Dave Lee <dave@gray101.com >
2024-04-13 09:45:34 +02:00
cryptk
1981154f49
fix: dont commit generated files to git ( #1993 )
...
* fix: initial work towards not committing generated files to the repository
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: improve build docs
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: remove unused folder from .dockerignore and .gitignore
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: attempt to fix extra backend tests
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: attempt to fix other tests
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: more test fixes
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: fix apple tests
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: more extras tests fixes
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: add GOBIN to PATH in docker build
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: extra tests and Dockerfile corrections
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: remove build dependency checks
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: add golang protobuf compilers to tests-linux action
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: ensure protogen is run for extra backend installs
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: use newer protobuf
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: more missing protoc binaries
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: missing dependencies during docker build
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: don't install grpc compilers in the final stage if they aren't needed
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: python-grpc-tools in 22.04 repos is too old
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: add a couple of extra build dependencies to Makefile
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: unbreak container rebuild functionality
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-13 09:37:32 +02:00
cryptk
a8ebf6f575
fix: respect concurrency from parent build parameters when building GRPC ( #2023 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-13 09:14:32 +02:00
LocalAI [bot]
912d2dccfa
⬆️ Update ggerganov/llama.cpp ( #2024 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-13 09:13:00 +02:00
dependabot[bot]
fcb63aed8a
build(deps): bump follow-redirects from 1.15.2 to 1.15.6 in /examples/langchain/langchainjs-localai-example ( #2020 )
...
build(deps): bump follow-redirects
Bumps [follow-redirects](https://github.com/follow-redirects/follow-redirects ) from 1.15.2 to 1.15.6.
- [Release notes](https://github.com/follow-redirects/follow-redirects/releases )
- [Commits](https://github.com/follow-redirects/follow-redirects/compare/v1.15.2...v1.15.6 )
---
updated-dependencies:
- dependency-name: follow-redirects
dependency-type: indirect
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-12 15:36:46 +00:00
Ettore Di Giacinto
0e549424e7
Update dependabot_auto.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-12 15:59:25 +02:00
Ettore Di Giacinto
69d638268b
Update dependabot_auto.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-12 15:57:13 +02:00
Ettore Di Giacinto
18eea9088a
Update dependabot_auto.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-12 15:38:34 +02:00
Ettore Di Giacinto
fb105837ba
Update secscan.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-12 15:37:56 +02:00
Ettore Di Giacinto
7e52c8e21a
Update CONTRIBUTING.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-12 15:27:40 +02:00
LocalAI [bot]
d068839896
⬆️ Update docs version mudler/LocalAI ( #2013 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-12 08:40:19 +02:00
dependabot[bot]
e0dee52a2a
build(deps): bump the pip group across 4 directories with 8 updates ( #2017 )
...
* build(deps): bump the pip group across 4 directories with 8 updates
Bumps the pip group with 1 update in the /examples/functions directory: [langchain](https://github.com/langchain-ai/langchain ).
Bumps the pip group with 2 updates in the /examples/langchain-chroma directory: [langchain](https://github.com/langchain-ai/langchain ) and [llama-index](https://github.com/run-llama/llama_index ).
Bumps the pip group with 6 updates in the /examples/langchain/langchainpy-localai-example directory:
| Package | From | To |
| --- | --- | --- |
| [langchain](https://github.com/langchain-ai/langchain ) | `0.0.159` | `0.1.0` |
| [aiohttp](https://github.com/aio-libs/aiohttp ) | `3.8.4` | `3.9.2` |
| [certifi](https://github.com/certifi/python-certifi ) | `2022.12.7` | `2023.7.22` |
| [idna](https://github.com/kjd/idna ) | `3.4` | `3.7` |
| [requests](https://github.com/psf/requests ) | `2.29.0` | `2.31.0` |
| [urllib3](https://github.com/urllib3/urllib3 ) | `1.26.15` | `1.26.18` |
Bumps the pip group with 1 update in the /examples/streamlit-bot directory: [streamlit](https://github.com/streamlit/streamlit ).
Updates `langchain` from 0.0.234 to 0.1.0
- [Release notes](https://github.com/langchain-ai/langchain/releases )
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.234...v0.1.0 )
Updates `langchain` from 0.0.160 to 0.1.0
- [Release notes](https://github.com/langchain-ai/langchain/releases )
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.234...v0.1.0 )
Updates `llama-index` from 0.6.2 to 0.9.36
- [Release notes](https://github.com/run-llama/llama_index/releases )
- [Changelog](https://github.com/run-llama/llama_index/blob/main/CHANGELOG.md )
- [Commits](https://github.com/run-llama/llama_index/compare/v0.6.2...v0.9.36 )
Updates `langchain` from 0.0.159 to 0.1.0
- [Release notes](https://github.com/langchain-ai/langchain/releases )
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.234...v0.1.0 )
Updates `aiohttp` from 3.8.4 to 3.9.2
- [Release notes](https://github.com/aio-libs/aiohttp/releases )
- [Changelog](https://github.com/aio-libs/aiohttp/blob/master/CHANGES.rst )
- [Commits](https://github.com/aio-libs/aiohttp/compare/v3.8.4...v3.9.2 )
Updates `certifi` from 2022.12.7 to 2023.7.22
- [Commits](https://github.com/certifi/python-certifi/compare/2022.12.07...2023.07.22 )
Updates `idna` from 3.4 to 3.7
- [Release notes](https://github.com/kjd/idna/releases )
- [Changelog](https://github.com/kjd/idna/blob/master/HISTORY.rst )
- [Commits](https://github.com/kjd/idna/compare/v3.4...v3.7 )
Updates `requests` from 2.29.0 to 2.31.0
- [Release notes](https://github.com/psf/requests/releases )
- [Changelog](https://github.com/psf/requests/blob/main/HISTORY.md )
- [Commits](https://github.com/psf/requests/compare/v2.29.0...v2.31.0 )
Updates `urllib3` from 1.26.15 to 1.26.18
- [Release notes](https://github.com/urllib3/urllib3/releases )
- [Changelog](https://github.com/urllib3/urllib3/blob/main/CHANGES.rst )
- [Commits](https://github.com/urllib3/urllib3/compare/1.26.15...1.26.18 )
Updates `streamlit` from 1.26.0 to 1.30.0
- [Release notes](https://github.com/streamlit/streamlit/releases )
- [Commits](https://github.com/streamlit/streamlit/compare/1.26.0...1.30.0 )
---
updated-dependencies:
- dependency-name: langchain
dependency-type: direct:production
dependency-group: pip
- dependency-name: langchain
dependency-type: direct:production
dependency-group: pip
- dependency-name: llama-index
dependency-type: direct:production
dependency-group: pip
- dependency-name: langchain
dependency-type: direct:production
dependency-group: pip
- dependency-name: aiohttp
dependency-type: direct:production
dependency-group: pip
- dependency-name: certifi
dependency-type: direct:production
dependency-group: pip
- dependency-name: idna
dependency-type: direct:production
dependency-group: pip
- dependency-name: requests
dependency-type: direct:production
dependency-group: pip
- dependency-name: urllib3
dependency-type: direct:production
dependency-group: pip
- dependency-name: streamlit
dependency-type: direct:production
dependency-group: pip
...
Signed-off-by: dependabot[bot] <support@github.com >
* Update version.json
PR appears stuck on a check, needs any arbitrary commit to run the security check workflow with write permissions. Bumping docs versions to match latest release as said useless change.
Signed-off-by: Dave <dave@gray101.com >
---------
Signed-off-by: dependabot[bot] <support@github.com >
Signed-off-by: Dave <dave@gray101.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Co-authored-by: Dave <dave@gray101.com >
2024-04-12 00:53:43 -04:00
LocalAI [bot]
677e20756b
⬆️ Update ggerganov/llama.cpp ( #2014 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2024-04-12 00:49:41 +02:00
Ettore Di Giacinto
b2785ff06e
feat(gallery): support ConfigURLs ( #2012 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-12 00:49:23 +02:00
dependabot[bot]
da82ce81b5
build(deps): bump github.com/opencontainers/runc from 1.1.5 to 1.1.12 ( #2000 )
...
Bumps [github.com/opencontainers/runc](https://github.com/opencontainers/runc ) from 1.1.5 to 1.1.12.
- [Release notes](https://github.com/opencontainers/runc/releases )
- [Changelog](https://github.com/opencontainers/runc/blob/main/CHANGELOG.md )
- [Commits](https://github.com/opencontainers/runc/compare/v1.1.5...v1.1.12 )
---
updated-dependencies:
- dependency-name: github.com/opencontainers/runc
dependency-type: indirect
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 18:57:33 +00:00
Ettore Di Giacinto
70c4f110a4
Update overview.md
2024-04-11 20:18:05 +02:00
Ettore Di Giacinto
099bd54ff2
ci: try to build on macos14 ( #2011 )
...
* ci: try to build on macos14
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci: fixup artifact name
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-11 19:22:30 +02:00
Ludovic Leroux
12c0d9443e
feat: use tokenizer.apply_chat_template() in vLLM ( #1990 )
...
Use tokenizer.apply_chat_template() in vLLM
Signed-off-by: Ludovic LEROUX <ludovic@inpher.io >
2024-04-11 19:20:22 +02:00
dependabot[bot]
cbda06fb96
build(deps): bump github.com/gofiber/fiber/v2 from 2.52.0 to 2.52.4 ( #2008 )
...
Bumps [github.com/gofiber/fiber/v2](https://github.com/gofiber/fiber ) from 2.52.0 to 2.52.4.
- [Release notes](https://github.com/gofiber/fiber/releases )
- [Commits](https://github.com/gofiber/fiber/compare/v2.52.0...v2.52.4 )
---
updated-dependencies:
- dependency-name: github.com/gofiber/fiber/v2
dependency-type: direct:production
update-type: version-update:semver-patch
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 16:52:54 +00:00
Ettore Di Giacinto
b1a242251c
ci: fixup upload artifact name
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-11 18:26:03 +02:00
dependabot[bot]
fce606fc0f
build(deps): bump github.com/charmbracelet/glamour from 0.6.0 to 0.7.0 ( #2004 )
...
Bumps [github.com/charmbracelet/glamour](https://github.com/charmbracelet/glamour ) from 0.6.0 to 0.7.0.
- [Release notes](https://github.com/charmbracelet/glamour/releases )
- [Commits](https://github.com/charmbracelet/glamour/compare/v0.6.0...v0.7.0 )
---
updated-dependencies:
- dependency-name: github.com/charmbracelet/glamour
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 15:41:58 +00:00
dependabot[bot]
b606c7b768
build(deps): bump actions/upload-artifact from 3 to 4 ( #2007 )
...
Bumps [actions/upload-artifact](https://github.com/actions/upload-artifact ) from 3 to 4.
- [Release notes](https://github.com/actions/upload-artifact/releases )
- [Commits](https://github.com/actions/upload-artifact/compare/v3...v4 )
---
updated-dependencies:
- dependency-name: actions/upload-artifact
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 14:44:02 +00:00
dependabot[bot]
0a6956b029
build(deps): bump actions/cache from 3 to 4 ( #2006 )
...
Bumps [actions/cache](https://github.com/actions/cache ) from 3 to 4.
- [Release notes](https://github.com/actions/cache/releases )
- [Changelog](https://github.com/actions/cache/blob/main/RELEASES.md )
- [Commits](https://github.com/actions/cache/compare/v3...v4 )
---
updated-dependencies:
- dependency-name: actions/cache
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 14:35:27 +00:00
dependabot[bot]
821cf0e3fd
build(deps): bump peter-evans/create-pull-request from 5 to 6 ( #2005 )
...
Bumps [peter-evans/create-pull-request](https://github.com/peter-evans/create-pull-request ) from 5 to 6.
- [Release notes](https://github.com/peter-evans/create-pull-request/releases )
- [Commits](https://github.com/peter-evans/create-pull-request/compare/v5...v6 )
---
updated-dependencies:
- dependency-name: peter-evans/create-pull-request
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 13:58:04 +00:00
dependabot[bot]
11a0418510
build(deps): bump actions/setup-go from 4 to 5 ( #2003 )
...
Bumps [actions/setup-go](https://github.com/actions/setup-go ) from 4 to 5.
- [Release notes](https://github.com/actions/setup-go/releases )
- [Commits](https://github.com/actions/setup-go/compare/v4...v5 )
---
updated-dependencies:
- dependency-name: actions/setup-go
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 13:10:32 +00:00
dependabot[bot]
40781ac013
build(deps): bump actions/checkout from 3 to 4 ( #2002 )
...
Bumps [actions/checkout](https://github.com/actions/checkout ) from 3 to 4.
- [Release notes](https://github.com/actions/checkout/releases )
- [Changelog](https://github.com/actions/checkout/blob/main/CHANGELOG.md )
- [Commits](https://github.com/actions/checkout/compare/v3...v4 )
---
updated-dependencies:
- dependency-name: actions/checkout
dependency-type: direct:production
update-type: version-update:semver-major
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 12:48:30 +00:00
dependabot[bot]
fdfd868953
build(deps): bump github.com/gofiber/fiber/v2 from 2.52.0 to 2.52.1 ( #2001 )
...
Bumps [github.com/gofiber/fiber/v2](https://github.com/gofiber/fiber ) from 2.52.0 to 2.52.1.
- [Release notes](https://github.com/gofiber/fiber/releases )
- [Commits](https://github.com/gofiber/fiber/compare/v2.52.0...v2.52.1 )
---
updated-dependencies:
- dependency-name: github.com/gofiber/fiber/v2
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 12:21:52 +00:00
dependabot[bot]
0795975486
build(deps): bump github.com/docker/docker from 20.10.7+incompatible to 24.0.9+incompatible ( #1999 )
...
build(deps): bump github.com/docker/docker
Bumps [github.com/docker/docker](https://github.com/docker/docker ) from 20.10.7+incompatible to 24.0.9+incompatible.
- [Release notes](https://github.com/docker/docker/releases )
- [Commits](https://github.com/docker/docker/compare/v20.10.7...v24.0.9 )
---
updated-dependencies:
- dependency-name: github.com/docker/docker
dependency-type: indirect
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 11:44:34 +00:00
dependabot[bot]
a49248d29f
build(deps): bump google.golang.org/protobuf from 1.31.0 to 1.33.0 ( #1998 )
...
Bumps google.golang.org/protobuf from 1.31.0 to 1.33.0.
---
updated-dependencies:
- dependency-name: google.golang.org/protobuf
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-04-11 11:07:45 +00:00
Ettore Di Giacinto
182fef339d
Create dependabot_auto.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-11 12:13:06 +02:00
Ettore Di Giacinto
c74dec7e38
Add dependabot.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-11 11:47:54 +02:00
Ludovic Leroux
b4548ad72d
feat: add flash-attn in nvidia and rocm envs ( #1995 )
...
Signed-off-by: Ludovic LEROUX <ludovic@inpher.io >
2024-04-11 09:44:39 +02:00
LocalAI [bot]
e152b07b74
⬆️ Update ggerganov/llama.cpp ( #1991 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-11 09:22:07 +02:00
LocalAI [bot]
0e44a4e664
⬆️ Update docs version mudler/LocalAI ( #1988 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-11 09:19:46 +02:00
cryptk
24d7dadfed
feat: kong cli refactor fixes #1955 ( #1974 )
...
* feat: migrate to alecthomas/kong for CLI
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: bring in new flag for granular log levels
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* chore: go mod tidy
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: allow loading cli flag values from ["./localai.yaml", "~/.config/localai.yaml", "/etc/localai.yaml"] in that order
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: load from .env file instead of a yaml file
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: better loading for environment files
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat(doc): add initial documentation about configuration
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: remove test log lines
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: integrate new documentation into existing pages
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: add documentation on .env files
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* fix: cleanup some documentation table errors
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
* feat: refactor CLI logic out to it's own package under core/cli
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
---------
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-11 09:19:24 +02:00
Ettore Di Giacinto
92005b9c02
Update openai-functions.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-10 16:30:57 +02:00
Ettore Di Giacinto
636d487dc8
Update gpt-vision.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-10 16:30:03 +02:00
Ettore Di Giacinto
93f51d80d4
Update gpt-vision.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-10 16:29:46 +02:00
Koen Farell
36da11a0ee
deps: Update version of vLLM to add support of Cohere Command_R model in vLLM inference ( #1975 )
...
* Update vLLM version to add support of Command_R
Signed-off-by: Koen Farell <hellios.dt@gmail.com >
* fix: Fixed vllm version from requirements
Signed-off-by: Koen Farell <hellios.dt@gmail.com >
* chore: Update transformers-rocm.yml
Signed-off-by: Koen Farell <hellios.dt@gmail.com >
* chore: Update transformers.yml version of vllm
Signed-off-by: Koen Farell <hellios.dt@gmail.com >
---------
Signed-off-by: Koen Farell <hellios.dt@gmail.com >
2024-04-10 11:25:26 +00:00
Sebastian.W
d23e73b118
fix(autogptq): do not use_triton with qwen-vl ( #1985 )
...
* Enhance autogptq backend to support VL models
* update dependencies for autogptq
* remove redundant auto-gptq dependency
* Convert base64 to image_url for Qwen-VL model
* implemented model inference for qwen-vl
* remove user prompt from generated answer
* fixed write image error
* fixed use_triton issue when loading Qwen-VL model
---------
Co-authored-by: Binghua Wu <bingwu@estee.com >
2024-04-10 10:36:10 +00:00
Ettore Di Giacinto
d692b2c32a
ci: push latest images for dockerhub ( #1984 )
...
Fixes : #1983
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-10 10:31:59 +02:00
LocalAI [bot]
7e2f8bb408
⬆️ Update ggerganov/whisper.cpp ( #1980 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-10 09:08:00 +02:00
LocalAI [bot]
951e39d36c
⬆️ Update ggerganov/llama.cpp ( #1979 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-10 09:07:41 +02:00
LocalAI [bot]
aeb3f835ae
⬆️ Update docs version mudler/LocalAI ( #1978 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-10 09:07:21 +02:00
Ettore Di Giacinto
cc3d601836
ci: fixup latest image push
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-09 09:49:11 +02:00
Ettore Di Giacinto
2bbb221fb1
tests(petals): temp disable
2024-04-08 21:28:59 +00:00
LocalAI [bot]
195be10050
⬆️ Update ggerganov/llama.cpp ( #1973 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-08 23:26:52 +02:00
fakezeta
a38618db02
fix regression #1971 ( #1972 )
...
fixes regression #1971 introduced by intel_extension_for_transformers==1.4
2024-04-08 22:33:51 +02:00
LocalAI [bot]
efcca15d3f
⬆️ Update ggerganov/llama.cpp ( #1970 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-08 08:38:47 +02:00
LocalAI [bot]
a153b628c2
⬆️ Update ggerganov/whisper.cpp ( #1969 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-08 08:38:17 +02:00
Ettore Di Giacinto
f36d86ba6d
fix(hermes-2-pro-mistral): correct dashes in template to suppress newlines ( #1966 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-07 18:23:47 +02:00
Ettore Di Giacinto
74492a81c7
doc(quickstart): fix typo
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-04-07 11:06:35 +02:00
LocalAI [bot]
ed13782986
⬆️ Update ggerganov/llama.cpp ( #1964 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-07 10:32:10 +02:00
Ettore Di Giacinto
8342553214
fix(llama.cpp): set better defaults for llama.cpp ( #1961 )
...
fix(defaults): set better defaults for llama.cpp
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-06 22:56:45 +02:00
LocalAI [bot]
8aa5f5a660
⬆️ Update ggerganov/llama.cpp ( #1960 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-06 19:15:25 +00:00
LocalAI [bot]
b2d9e3f704
⬆️ Update ggerganov/llama.cpp ( #1959 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-05 08:41:55 +02:00
LocalAI [bot]
f744e1f931
⬆️ Update ggerganov/whisper.cpp ( #1958 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-05 08:41:35 +02:00
cryptk
b85dad0286
feat: first pass at improving logging ( #1956 )
...
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com >
2024-04-04 09:24:22 +02:00
LocalAI [bot]
3851b51d98
⬆️ Update ggerganov/llama.cpp ( #1953 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-04 00:27:57 +02:00
Ettore Di Giacinto
ff77d3bc22
fix(seed): generate random seed per-request if -1 is set ( #1952 )
...
* fix(seed): generate random seed per-request if -1 is set
Also update ci with new workflows and allow the aio tests to run with an
api key
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* docs(openvino): Add OpenVINO example
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-03 22:25:47 +02:00
Ettore Di Giacinto
93cfec3c32
ci: correctly tag latest and aio images
2024-04-03 11:30:23 +02:00
Ettore Di Giacinto
89560ef87f
fix(ci): manually tag latest images ( #1948 )
...
fix(ci): manually tag images
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-02 19:25:46 +02:00
Ettore Di Giacinto
9bc209ba73
fix(welcome): stable model list ( #1949 )
2024-04-02 19:25:32 +02:00
Ettore Di Giacinto
84e0dc3246
fix(hermes-2-pro-mistral): correct stopwords ( #1947 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-02 15:38:00 +02:00
LocalAI [bot]
4d4d76114d
⬆️ Update ggerganov/llama.cpp ( #1941 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-02 09:16:04 +02:00
cryptk
86bc5f1350
fix: use exec in entrypoint scripts to fix signal handling ( #1943 )
2024-04-02 09:15:44 +02:00
Ettore Di Giacinto
e8f02c083f
fix(functions): respect when selected from string ( #1940 )
...
* fix(functions): respect when selected from string
* fix(toolschoice): decode both string and objects
2024-04-01 19:39:54 +02:00
Ettore Di Giacinto
ebb1fcedea
fix(hermes-2-pro-mistral): add stopword for toolcall ( #1939 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-04-01 11:48:35 +02:00
LocalAI [bot]
66f90f8dc1
⬆️ Update ggerganov/llama.cpp ( #1937 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-04-01 08:59:23 +02:00
Ettore Di Giacinto
3c778b538a
Update phi-2-orange.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-31 13:06:41 +02:00
Ettore Di Giacinto
35290e146b
fix(grammar): respect JSONmode and grammar from user input ( #1935 )
...
* fix(grammar): Fix JSON mode and custom grammar
* tests(aio): add jsonmode test
* tests(aio): add functioncall test
* fix(aio): use hermes-2-pro-mistral as llm for CPU profile
* add phi-2-orange
2024-03-31 13:04:09 +02:00
LocalAI [bot]
784657a652
⬆️ Update ggerganov/llama.cpp ( #1934 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-31 00:27:38 +01:00
LocalAI [bot]
831efa8893
⬆️ Update ggerganov/whisper.cpp ( #1933 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-31 00:27:16 +01:00
Ettore Di Giacinto
957f428fd5
fix(tools): correctly render tools response in templates ( #1932 )
...
* fix(tools): allow to correctly display both Functions and Tools
* models(hermes-2-pro): correctly display function results
2024-03-30 19:02:07 +01:00
Ettore Di Giacinto
61e5e6bc36
fix(swagger): do not specify a host ( #1930 )
...
In this way the requests are redirected to the host used by the client
to perform the request.
2024-03-30 12:04:41 +01:00
Ettore Di Giacinto
eab4a91a9b
fix(aio): correctly detect intel systems ( #1931 )
...
Also rename SIZE to PROFILE
2024-03-30 12:04:32 +01:00
LocalAI [bot]
2bba62ca4d
⬆️ Update ggerganov/llama.cpp ( #1928 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-29 22:52:01 +00:00
Ettore Di Giacinto
bcdc83b46d
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-29 23:00:06 +01:00
Ettore Di Giacinto
92fbdfd06f
feat(swagger): update ( #1929 )
2024-03-29 22:48:58 +01:00
cryptk
93702e39d4
feat(build): adjust number of parallel make jobs ( #1915 )
...
* feat(build): adjust number of parallel make jobs
* fix: update make on MacOS from brew to support --output-sync argument
* fix: cache grpc with version as part of key to improve validity of cache hits
* fix: use gmake for tests-apple to use the updated GNU make version
* fix: actually use the new make version for tests-apple
* feat: parallelize tests-extra
* feat: attempt to cache grpc build for docker images
* fix: don't quote GRPC version
* fix: don't cache go modules, we have limited cache space, better used elsewhere
* fix: release with the same version of go that we test with
* fix: don't fail on exporting cache layers
* fix: remove deprecated BUILD_GRPC docker arg from Makefile
2024-03-29 22:32:40 +01:00
LocalAI [bot]
a7fc89c207
⬆️ Update ggerganov/whisper.cpp ( #1927 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-29 22:29:50 +01:00
Ettore Di Giacinto
123a5a2e16
feat(swagger): Add swagger API doc ( #1926 )
...
* makefile(build): add minimal and api build target
* feat(swagger): Add swagger
2024-03-29 22:29:33 +01:00
LocalAI [bot]
ab2f403dd0
⬆️ Update ggerganov/whisper.cpp ( #1924 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-29 00:13:59 +01:00
LocalAI [bot]
b9c5e14e2c
⬆️ Update ggerganov/llama.cpp ( #1923 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-29 00:13:38 +01:00
Ettore Di Giacinto
bf65ed6eb8
feat(webui): add partials, show backends associated to models ( #1922 )
...
* feat(webui): add partials, show backends associated to models
* fix(auth): put assistant and backend under auth
2024-03-28 21:52:52 +01:00
Ettore Di Giacinto
4e79294f97
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-28 19:52:40 +01:00
Ettore Di Giacinto
8477e8fac3
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-28 18:28:30 +01:00
Ettore Di Giacinto
13ccd2afef
docs(aio-usage): update docs to show examples ( #1921 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-03-28 18:16:58 +01:00
Ettore Di Giacinto
23b833d171
Update run-other-models.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-28 12:42:37 +01:00
LocalAI [bot]
07c49ee4b8
⬆️ Update ggerganov/whisper.cpp ( #1914 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-27 22:53:13 +00:00
LocalAI [bot]
07c4bdda7c
⬆️ Update ggerganov/llama.cpp ( #1913 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-27 21:57:59 +00:00
Ettore Di Giacinto
2266d8263c
Update README.md
2024-03-27 22:48:46 +01:00
Ettore Di Giacinto
160eb48b2b
Update quickstart.md
2024-03-27 22:47:59 +01:00
cryptk
0c0efc871c
fix(build): better CI logging and correct some build failure modes in Makefile ( #1899 )
...
* feat: group make output by target when running parallelized builds in CI
* fix: quote GO_TAGS in makefile to fix handling of whitespace in value
* fix: set CPATH to find opencv2 in it's commonly installed location
* fix: add missing go mod dropreplace for go-llama.cpp
* chore: remove opencv symlink from github workflows
2024-03-27 21:12:19 +01:00
Gianluca Boiano
7ef5f3b473
⬆️ Update M0Rf30/go-tiny-dream ( #1911 )
2024-03-27 21:12:04 +01:00
Ettore Di Giacinto
66ee4afb95
feat(welcome): add simple welcome page ( #1912 )
...
* feat(welcome): add simple welcome page
* feat(api): add 404 handling
2024-03-27 21:10:58 +01:00
Ettore Di Giacinto
93f0b7ae03
update hot topics
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-27 18:17:12 +01:00
fakezeta
8210ffcb6c
feat: Token Stream support for Transformer, fix: missing package for OpenVINO ( #1908 )
...
* Streaming working
* Small fix for regression on CUDA and XPU
* use pip version of optimum[openvino]
* Update backend/python/transformers/transformers_server.py
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* Token streaming support
fix optimum[openvino] package in install.sh
* Token Streaming support
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-27 17:50:35 +01:00
fakezeta
e7cbe32601
feat: Openvino runtime for transformer backend and streaming support for Openvino and CUDA ( #1892 )
...
* fixes #1775 and #1774
Add BitsAndBytes Quantization and fixes embedding on CUDA devices
* Manage 4bit and 8 bit quantization
Manage different BitsAndBytes options with the quantization: parameter in yaml
* fix compilation errors on non CUDA environment
* OpenVINO draft
First draft of OpenVINO integration in transformer backend
* first working implementation
* Streaming working
* Small fix for regression on CUDA and XPU
* use pip version of optimum[openvino]
* Update backend/python/transformers/transformers_server.py
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-26 23:31:43 +00:00
LocalAI [bot]
b500ceaf73
⬆️ Update ggerganov/llama.cpp ( #1904 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-26 23:21:54 +00:00
LocalAI [bot]
d3c283ac19
⬆️ Update docs version mudler/LocalAI ( #1903 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-26 22:56:42 +01:00
Ettore Di Giacinto
607586e0b7
fix: downgrade torch ( #1902 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-26 22:56:02 +01:00
Steven Christou
2d7913b3be
feat(assistant): Assistant and AssistantFiles api ( #1803 )
...
* Initial implementation of assistants api
* Move load/save configs to utils
* Save assistant and assistantfiles config to disk.
* Add tsets for assistant api
* Fix models path spelling mistake.
* Remove personal go.mod information
---------
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-26 18:54:35 +01:00
Sebastian.W
b7ffe66219
Enhance autogptq backend to support VL models ( #1860 )
...
* Enhance autogptq backend to support VL models
* update dependencies for autogptq
* remove redundant auto-gptq dependency
* Convert base64 to image_url for Qwen-VL model
* implemented model inference for qwen-vl
* remove user prompt from generated answer
* fixed write image error
---------
Co-authored-by: Binghua Wu <bingwu@estee.com >
2024-03-26 18:48:14 +01:00
Ettore Di Giacinto
e58410fa99
feat(aio): add intel profile ( #1901 )
...
* feat(aio): add intel profile
* docs: clarify AIO images features
2024-03-26 18:45:25 +01:00
LocalAI [bot]
1395e505cd
⬆️ Update ggerganov/llama.cpp ( #1897 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-26 00:34:10 +01:00
LocalAI [bot]
42a4c86dca
⬆️ Update ggerganov/whisper.cpp ( #1896 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-26 00:33:46 +01:00
Ettore Di Giacinto
c9adc5680c
fix(aio): make image-gen for GPU functional, update docs ( #1895 )
...
* readme: update quickstart
* aio(gpu): fix dreamshaper
* tests(aio): allow to run tests also against an endpoint
* docs: split content
* tests: less verbosity
---------
Co-authored-by: Dave <dave@gray101.com >
2024-03-25 21:04:32 +00:00
Enrico Ros
08c7b17298
Fix NVIDIA VRAM detection on WSL2 environments ( #1894 )
...
* NVIDIA VRAM detection on WSL2 environments
More robust single NVIDIA GPU memory detection, following the
improved NVIDIA WSL2 detection patch yesterday #1891 .
Tested and working on WSL2, Linux.
Signed-off-by: Enrico Ros <enrico.ros@gmail.com >
* Update aio/entrypoint.sh
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: Enrico Ros <enrico.ros@gmail.com >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-25 18:36:18 +01:00
Enrico Ros
5e12382524
NVIDIA GPU detection support for WSL2 environments ( #1891 )
...
This change makes the assumption that "Microsoft Corporation Device 008e"
is an NVIDIA CUDA device. If this is not the case, please update the
hardware detection script here.
Signed-off-by: Enrico Ros <enrico.ros@gmail.com >
Co-authored-by: Dave <dave@gray101.com >
2024-03-25 08:32:40 +01:00
Ettore Di Giacinto
6cf99527f8
docs(aio): Add All-in-One images docs ( #1887 )
...
* docs(aio): Add AIO images docs
* add image generation link to quickstart
* while reviewing I noticed this one link was missing, so quickly adding it.
Signed-off-by: Dave <dave@gray101.com >
Co-authored-by: Dave <dave@gray101.com >
2024-03-25 02:01:30 +00:00
LocalAI [bot]
3e293f1465
⬆️ Update ggerganov/llama.cpp ( #1889 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-24 21:12:18 +00:00
LocalAI [bot]
0106c58181
⬆️ Update ggerganov/llama.cpp ( #1885 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-24 14:54:01 +01:00
Ettore Di Giacinto
bd25d8049c
fix(watchdog): use ShutdownModel instead of StopModel ( #1882 )
...
Fixes #1760
2024-03-23 16:19:57 +01:00
Ettore Di Giacinto
49cec7fd61
ci(aio): add latest tag images ( #1884 )
...
Tangentially also fixes #1868
2024-03-23 16:08:32 +01:00
Ettore Di Giacinto
d9456f2a23
ci(aio): publish hipblas and Intel GPU images ( #1883 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-03-23 15:54:14 +01:00
Ettore Di Giacinto
8495750cb8
Update release.yml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-23 15:22:26 +01:00
Ettore Di Giacinto
1f501cc1ef
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-23 10:42:14 +01:00
LocalAI [bot]
a922119c41
⬆️ Update ggerganov/llama.cpp ( #1881 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-23 09:23:28 +01:00
Richard Palethorpe
643d85d2cc
feat(stores): Vector store backend ( #1795 )
...
Add simple vector store backend
Signed-off-by: Richard Palethorpe <io@richiejp.com >
2024-03-22 21:14:04 +01:00
Ettore Di Giacinto
4b1ee0c170
feat(aio): add tests, update model definitions ( #1880 )
2024-03-22 21:13:11 +01:00
Ettore Di Giacinto
3bec467a91
feat(models): add phi-2-chat, llava-1.6, bakllava, cerbero ( #1879 )
2024-03-22 21:12:48 +01:00
Ettore Di Giacinto
600152df23
fix(config): pass by config options, respect defaults ( #1878 )
...
This bug had the unpleasant effect that it ignored defaults passed by
the CLI. For instance threads could be changed only via model config
file.
2024-03-22 20:55:11 +01:00
LocalAI [bot]
dd84c29a3d
⬆️ Update ggerganov/whisper.cpp ( #1875 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-22 09:14:56 +01:00
LocalAI [bot]
07468c8786
⬆️ Update ggerganov/llama.cpp ( #1874 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-22 09:14:42 +01:00
Ettore Di Giacinto
418ba02025
ci: fix typo
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-22 09:14:17 +01:00
Ettore Di Giacinto
abc9360dc6
feat(aio): entrypoint, update workflows ( #1872 )
2024-03-21 22:09:04 +01:00
Sebastian
743095b7d8
docs(mac): improve documentation for mac build ( #1873 )
...
* docs(mac): Improve documentation for mac build
- added documentation to build from current master
- added troubleshooting information
Signed-off-by: Sebastian <tauven@gmail.com >
* docs(max): fix typo
Signed-off-by: Sebastian <tauven@gmail.com >
---------
Signed-off-by: Sebastian <tauven@gmail.com >
2024-03-21 22:08:33 +01:00
Ettore Di Giacinto
3cf64d1e7e
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-21 08:57:41 +01:00
Ettore Di Giacinto
e533dcf506
feat(functions/aio): all-in-one images, function template enhancements ( #1862 )
...
* feat(startup): allow to specify models from local files
* feat(aio): add Dockerfile, make targets, aio profiles
* feat(template): add Function and LastMessage
* add hermes2-pro-mistral
* update hermes2 definition
* feat(template): add sprig
* feat(template): expose FunctionCall
* feat(aio): switch llm for text
2024-03-21 01:12:20 +01:00
LocalAI [bot]
eeaf8c7ccd
⬆️ Update ggerganov/whisper.cpp ( #1867 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-20 22:26:29 +00:00
LocalAI [bot]
7e34dfdae7
⬆️ Update ggerganov/llama.cpp ( #1866 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-20 22:13:29 +00:00
LocalAI [bot]
e4bf51d5bd
⬆️ Update ggerganov/llama.cpp ( #1864 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-20 09:05:53 +01:00
LocalAI [bot]
ead61bf9d5
⬆️ Update ggerganov/llama.cpp ( #1857 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-19 00:03:17 +00:00
LocalAI [bot]
b12a205320
⬆️ Update docs version mudler/LocalAI ( #1856 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-19 00:44:45 +01:00
LocalAI [bot]
621541a92f
⬆️ Update ggerganov/whisper.cpp ( #1508 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-19 00:44:23 +01:00
Dave
ed5734ae25
test/fix: OSX Test Repair ( #1843 )
...
* test with gguf instead of ggml. Updates testPrompt to match? Adds debugging line to Dockerfile that I've found helpful recently.
* fix testPrompt slightly
* Sad Experiment: Test GH runner without metal?
* break apart CGO_LDFLAGS
* switch runner
* upstream llama.cpp disables Metal on Github CI!
* missed a dir from clean-tests
* CGO_LDFLAGS
* tmate failure + NO_ACCELERATE
* whisper.cpp has a metal fix
* do the exact opposite of the name of this branch, but keep it around for unrelated fixes?
* add back newlines
* add tmate to linux for testing
* update fixtures
* timeout for tmate
2024-03-18 19:19:43 +01:00
Ettore Di Giacinto
a046dcac5e
fix(config-watcher): start only if config-directory exists ( #1854 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-03-18 19:14:48 +01:00
Ettore Di Giacinto
843f93e1ab
fix(config): default to debug=false if not set ( #1853 )
2024-03-18 18:59:39 +01:00
Ettore Di Giacinto
fa9e330fc6
fix(llama.cpp): fix eos without cache ( #1852 )
2024-03-18 18:59:24 +01:00
Ettore Di Giacinto
b202bfaaa0
deps(whisper.cpp): update, fix cublas build ( #1846 )
...
fix(whisper.cpp): Add stubs and -lcuda
2024-03-18 15:56:53 +01:00
LocalAI [bot]
0eb0ac7dd0
⬆️ Update ggerganov/llama.cpp ( #1848 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-18 08:57:58 +01:00
LocalAI [bot]
d2b83d8357
⬆️ Update docs version mudler/LocalAI ( #1847 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-17 23:08:32 +01:00
Ettore Di Giacinto
88b65f63d0
fix(go-llama): use llama-cpp as default ( #1849 )
...
* fix(go-llama): use llama-cpp as default
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* fix(backends): drop obsoleted lines
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-17 23:08:22 +01:00
cryptk
020ce29cd8
fix(make): allow to parallelize jobs ( #1845 )
...
* fix: clean up Makefile dependencies to allow for parallel builds
* refactor: remove old unused backend from Makefile
* fix: finish removing legacy backend, update piper
* fix: I broke llama... I fixed llama
* feat: give the tests and builds a few threads
* fix: ensure libraries are replaced before build, add dropreplace target
* Fix image build workflows
2024-03-17 15:39:20 +01:00
Chakib Benziane
801b481beb
fixes #1051 : handle openai presence and request penalty parameters ( #1817 )
...
* fix request debugging, disable marshalling of context fields
Signed-off-by: blob42 <contact@blob42.xyz >
* merge frequency_penalty request parm with config
Signed-off-by: blob42 <contact@blob42.xyz >
* openai: add presence_penalty parameter
Signed-off-by: blob42 <contact@blob42.xyz >
---------
Signed-off-by: blob42 <contact@blob42.xyz >
2024-03-17 09:43:20 +01:00
LocalAI [bot]
8967ed1601
⬆️ Update ggerganov/llama.cpp ( #1840 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-16 11:25:41 +00:00
LocalAI [bot]
5826fb8e6d
⬆️ Update mudler/go-piper ( #1844 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-15 23:51:03 +00:00
Ettore Di Giacinto
89351f1a7d
feat(embeddings): do not require to be configured ( #1842 )
...
Certain engines requires to know during model loading
if the embedding feature has to be enabled, however, it is impractical
to have to set it to ALL the backends that supports embeddings.
There are transformers and sentencentransformers that seamelessly handle
both cases, without having this settings to be explicitly enabled.
The case sussist only for ggml-based models that needs to enable
featuresets during model loading (and thus settings `embedding` is
required), however most of the other engines does not require this.
This change disables the check done at code side, making easier to use
embeddings by not having to specify explicitly `embeddings: true`.
Part of: https://github.com/mudler/LocalAI/issues/1373
2024-03-15 18:14:23 +01:00
Ettore Di Giacinto
ae2e4fc2fe
docs(transformers): add docs section about transformers ( #1841 )
2024-03-15 18:13:30 +01:00
Dave
db199f61da
fix: osx build default.metallib ( #1837 )
...
fix: osx build default.metallib (#1837 )
* port osx fix from refactor pr to slim pr
* manually bump llama.cpp version to unstick CI?
2024-03-15 08:18:58 +00:00
LocalAI [bot]
44adbd2c75
⬆️ Update go-skynet/go-llama.cpp ( #1835 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-14 23:06:42 +00:00
Ettore Di Giacinto
20136ca8b7
feat(tts): add Elevenlabs and OpenAI TTS compatibility layer ( #1834 )
...
* feat(elevenlabs): map elevenlabs API support to TTS
This allows elevenlabs Clients to work automatically with LocalAI by
supporting the elevenlabs API.
The elevenlabs server endpoint is implemented such as it is wired to the
TTS endpoints.
Fixes: https://github.com/mudler/LocalAI/issues/1809
* feat(openai/tts): compat layer with openai tts
Fixes : #1276
* fix: adapt tts CLI
2024-03-14 23:08:34 +01:00
Dave
45d520f913
fix: OSX Build Files for llama.cpp ( #1836 )
...
bot ate my changes, seperate branch
2024-03-14 23:07:47 +01:00
fakezeta
3882130911
feat: Add Bitsandbytes quantization for transformer backend enhancement #1775 and fix: Transformer backend error on CUDA #1774 ( #1823 )
...
* fixes #1775 and #1774
Add BitsAndBytes Quantization and fixes embedding on CUDA devices
* Manage 4bit and 8 bit quantization
Manage different BitsAndBytes options with the quantization: parameter in yaml
* fix compilation errors on non CUDA environment
2024-03-14 23:06:30 +01:00
cryptk
a6b540737f
fix: missing OpenCL libraries from docker containers during clblas docker build ( #1830 )
2024-03-14 08:40:37 +01:00
LocalAI [bot]
f82065703d
⬆️ Update ggerganov/llama.cpp ( #1827 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-14 08:39:39 +01:00
cryptk
b423af001d
fix: the correct BUILD_TYPE for OpenCL is clblas (with no t) ( #1828 )
2024-03-14 08:39:21 +01:00
Ettore Di Giacinto
b9e77d394b
feat(model-help): display help text in markdown ( #1825 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-03-13 21:50:46 +01:00
Ettore Di Giacinto
57222497ec
fix(docker-compose): update docker compose file ( #1824 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-03-13 17:57:45 +01:00
LocalAI [bot]
5c5f07c1e7
⬆️ Update ggerganov/llama.cpp ( #1821 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-13 10:05:46 +01:00
Ettore Di Giacinto
f895d06605
fix(config): set better defaults for inferencing ( #1822 )
...
* fix(defaults): set better defaults for inferencing
This changeset aim to have better defaults and to properly detect when
no inference settings are provided with the model.
If not specified, we defaults to mirostat sampling, and offload all the
GPU layers (if a GPU is detected).
Related to https://github.com/mudler/LocalAI/issues/1373 and https://github.com/mudler/LocalAI/issues/1723
* Adapt tests
* Also pre-initialize default seed
2024-03-13 10:05:30 +01:00
Ettore Di Giacinto
bc8f648a91
fix(doc/examples): set defaults to mirostat ( #1820 )
...
The default sampler on some models don't return enough candidates which
leads to a false sense of randomness. Tracing back the code it looks
that with the temperature sampler there might not be enough
candidates to pick from, and since the seed and "randomness" take effect
while picking a good candidate this yields to the same results over and
over.
Fixes https://github.com/mudler/LocalAI/issues/1723 by updating the
examples and documentation to use mirostat instead.
2024-03-11 19:49:03 +01:00
LocalAI [bot]
8e57f4df31
⬆️ Update ggerganov/llama.cpp ( #1818 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-11 00:02:37 +01:00
LocalAI [bot]
a08cc5adbb
⬆️ Update ggerganov/llama.cpp ( #1816 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-10 09:32:09 +01:00
LocalAI [bot]
595a73fce4
⬆️ Update ggerganov/llama.cpp ( #1813 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-09 09:27:06 +01:00
LocalAI [bot]
dc919e08e8
⬆️ Update ggerganov/llama.cpp ( #1811 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-08 08:21:25 +01:00
Ettore Di Giacinto
5d1018495f
feat(intel): add diffusers/transformers support ( #1746 )
...
* feat(intel): add diffusers support
* try to consume upstream container image
* Debug
* Manually install deps
* Map transformers/hf cache dir to modelpath if not specified
* fix(compel): update initialization, pass by all gRPC options
* fix: add dependencies, implement transformers for xpu
* base it from the oneapi image
* Add pillow
* set threads if specified when launching the API
* Skip conda install if intel
* defaults to non-intel
* ci: add to pipelines
* prepare compel only if enabled
* Skip conda install if intel
* fix cleanup
* Disable compel by default
* Install torch 2.1.0 with Intel
* Skip conda on some setups
* Detect python
* Quiet output
* Do not override system python with conda
* Prefer python3
* Fixups
* exllama2: do not install without conda (overrides pytorch version)
* exllama/exllama2: do not install if not using cuda
* Add missing dataset dependency
* Small fixups, symlink to python, add requirements
* Add neural_speed to the deps
* correctly handle model offloading
* fix: device_map == xpu
* go back at calling python, fixed at dockerfile level
* Exllama2 restricted to only nvidia gpus
* Tokenizer to xpu
2024-03-07 14:37:45 +01:00
LocalAI [bot]
ad6fd7a991
⬆️ Update ggerganov/llama.cpp ( #1805 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-06 23:28:31 +01:00
LocalAI [bot]
e022b5959e
⬆️ Update mudler/go-stable-diffusion ( #1802 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-05 23:39:57 +00:00
LocalAI [bot]
db7f4955a1
⬆️ Update ggerganov/llama.cpp ( #1801 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-05 21:50:27 +00:00
Dave
5c69dd155f
feat(autogpt/transformers): consume trust_remote_code ( #1799 )
...
trusting remote code by default is a danger to our users
2024-03-05 19:47:15 +01:00
TwinFin
504f2e8bf4
Update Backend Dependancies ( #1797 )
...
* Update transformers.yml
Signed-off-by: TwinFin <57421631+TwinFinz@users.noreply.github.com >
* Update transformers-rocm.yml
Signed-off-by: TwinFin <57421631+TwinFinz@users.noreply.github.com >
* Update transformers-nvidia.yml
Signed-off-by: TwinFin <57421631+TwinFinz@users.noreply.github.com >
---------
Signed-off-by: TwinFin <57421631+TwinFinz@users.noreply.github.com >
2024-03-05 10:10:00 +00:00
Luna Midori
e586dc2924
Edit links in readme and integrations page ( #1796 )
...
* Update integrations.md
Signed-off-by: Luna Midori <118759930+lunamidori5@users.noreply.github.com >
* Update README.md
Signed-off-by: Luna Midori <118759930+lunamidori5@users.noreply.github.com >
* Update README.md
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Signed-off-by: Luna Midori <118759930+lunamidori5@users.noreply.github.com >
* Update README.md
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Signed-off-by: Luna Midori <118759930+lunamidori5@users.noreply.github.com >
---------
Signed-off-by: Luna Midori <118759930+lunamidori5@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-05 10:14:30 +01:00
Ettore Di Giacinto
333f918005
Update integrations.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-05 09:45:54 +01:00
LocalAI [bot]
c8e29033c2
⬆️ Update ggerganov/llama.cpp ( #1794 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-05 08:59:09 +01:00
LocalAI [bot]
d0bd961bde
⬆️ Update ggerganov/llama.cpp ( #1791 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-04 09:44:21 +01:00
Ettore Di Giacinto
006511ee25
Revert "feat(assistant): Initial implementation of assistants api" ( #1790 )
...
Revert "feat(assistant): Initial implementation of assistants api (#1761 )"
This reverts commit 4ab72146cd .
2024-03-03 10:31:06 +01:00
Steven Christou
4ab72146cd
feat(assistant): Initial implementation of assistants api ( #1761 )
...
Initial implementation of assistants api
2024-03-03 08:50:43 +01:00
LocalAI [bot]
b60a3fc879
⬆️ Update ggerganov/llama.cpp ( #1789 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-03 08:49:23 +01:00
Ettore Di Giacinto
a0eeb74957
Update hot topics/roadmap
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-03-02 09:35:40 +01:00
LocalAI [bot]
daa0b8741c
⬆️ Update ggerganov/llama.cpp ( #1785 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-03-01 22:38:24 +00:00
Ludovic Leroux
939411300a
Bump vLLM version + more options when loading models in vLLM ( #1782 )
...
* Bump vLLM version to 0.3.2
* Add vLLM model loading options
* Remove transformers-exllama
* Fix install exllama
2024-03-01 22:48:53 +01:00
Dave
1c312685aa
refactor: move remaining api packages to core ( #1731 )
...
* core 1
* api/openai/files fix
* core 2 - core/config
* move over core api.go and tests to the start of core/http
* move over localai specific endpoints to core/http, begin the service/endpoint split there
* refactor big chunk on the plane
* refactor chunk 2 on plane, next step: port and modify changes to request.go
* easy fixes for request.go, major changes not done yet
* lintfix
* json tag lintfix?
* gitignore and .keep files
* strange fix attempt: rename the config dir?
2024-03-01 16:19:53 +01:00
LocalAI [bot]
316de82f51
⬆️ Update ggerganov/llama.cpp ( #1779 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-29 22:33:30 +00:00
Ettore Di Giacinto
9068bc5271
Create SECURITY.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-29 19:53:04 +01:00
Oussama
31a4c9c9d3
Fix Command Injection Vulnerability ( #1778 )
...
* Added fix for command injection
* changed function name from sh to runCommand
2024-02-29 18:32:29 +00:00
Ettore Di Giacinto
c1966af2cf
ci: reduce stress on self-hosted runners ( #1776 )
...
Split jobs by self-hosted and free public runner provided by Github
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-29 11:40:08 +01:00
LocalAI [bot]
c665898652
⬆️ Update donomii/go-rwkv.cpp ( #1771 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-28 23:50:27 +00:00
LocalAI [bot]
f651a660aa
⬆️ Update ggerganov/llama.cpp ( #1772 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-28 23:02:30 +01:00
Ettore Di Giacinto
ba672b51da
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-28 16:03:38 +01:00
Ettore Di Giacinto
be498c5dd9
Update openai-functions.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-28 15:58:31 +01:00
Ettore Di Giacinto
6e95beccb9
Update overview.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-28 15:24:08 +01:00
Ettore Di Giacinto
c8be839481
Update openai-functions.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-27 23:24:46 +01:00
LocalAI [bot]
c7e08813a5
⬆️ Update ggerganov/llama.cpp ( #1767 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-27 23:12:51 +01:00
LocalAI [bot]
d21a6b33ab
⬆️ Update ggerganov/llama.cpp ( #1756 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-27 18:07:51 +00:00
Joshua Waring
9112cf153e
Update integrations.md ( #1765 )
...
Added Jetbrains compatible plugin for LocalAI
Signed-off-by: Joshua Waring <Joshhua5@users.noreply.github.com >
2024-02-27 17:35:59 +01:00
Ettore Di Giacinto
3868ac8402
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-27 15:44:15 +01:00
Ettore Di Giacinto
3f09010227
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-27 15:43:15 +01:00
Ettore Di Giacinto
d6cf82aba3
fix(tests): re-enable tests after code move ( #1764 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-27 15:04:19 +01:00
Ettore Di Giacinto
dfe54639b1
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-27 10:37:56 +01:00
Ettore Di Giacinto
bc5f5aa538
deps(llama.cpp): update ( #1759 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-26 13:18:44 +01:00
Ettore Di Giacinto
05818e0425
fix(functions): handle correctly when there are no results ( #1758 )
2024-02-26 08:38:23 +01:00
Sertaç Özercan
7f72a61104
ci: add stablediffusion to release ( #1757 )
...
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
2024-02-25 23:06:18 +00:00
LocalAI [bot]
8e45d47740
⬆️ Update ggerganov/llama.cpp ( #1753 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-25 10:03:19 +01:00
LocalAI [bot]
71771d1e9b
⬆️ Update docs version mudler/LocalAI ( #1752 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-25 10:02:52 +01:00
Ettore Di Giacinto
aa098e4d0b
fix(sse): do not omit empty finish_reason ( #1745 )
...
Fixes https://github.com/mudler/LocalAI/issues/1744
2024-02-24 11:51:59 +01:00
Ludovic Leroux
0135e1e3b9
fix: vllm - use AsyncLLMEngine to allow true streaming mode ( #1749 )
...
* fix: use vllm AsyncLLMEngine to bring true stream
Current vLLM implementation uses the LLMEngine, which was designed for offline batch inference, which results in the streaming mode outputing all blobs at once at the end of the inference.
This PR reworks the gRPC server to use asyncio and gRPC.aio, in combination with vLLM's AsyncLLMEngine to bring true stream mode.
This PR also passes more parameters to vLLM during inference (presence_penalty, frequency_penalty, stop, ignore_eos, seed, ...).
* Remove unused import
2024-02-24 11:48:45 +01:00
LocalAI [bot]
ff88c390bb
⬆️ Update ggerganov/llama.cpp ( #1750 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-24 00:06:46 +01:00
LocalAI [bot]
d825821a22
⬆️ Update ggerganov/llama.cpp ( #1740 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-23 00:07:15 +01:00
Luna Midori
cbed6ab1bb
Update README.md ( #1739 )
...
* Update README.md
Signed-off-by: Luna Midori <118759930+lunamidori5@users.noreply.github.com >
* Update README.md
Signed-off-by: Luna Midori <118759930+lunamidori5@users.noreply.github.com >
---------
Signed-off-by: Luna Midori <118759930+lunamidori5@users.noreply.github.com >
2024-02-22 16:35:06 +01:00
LocalAI [bot]
6fc122fa1a
⬆️ Update ggerganov/llama.cpp ( #1705 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-22 09:33:23 +00:00
Ettore Di Giacinto
feba38be36
examples(mistral-openorca): add stopword
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-22 00:15:08 +01:00
Ettore Di Giacinto
ba85d0bcad
feat(upload-api): do not display error if uploadedFiles.json is not present
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-22 00:15:08 +01:00
Ettore Di Giacinto
ad3623dd8d
examples(phi-2): strip newline at the end of the prompt template
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-21 23:17:51 +01:00
Ettore Di Giacinto
8292781045
deps(llama.cpp): update, support Gemma models ( #1734 )
...
deps(llama.cpp): update
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-21 17:23:38 +01:00
Ettore Di Giacinto
54ec6348fa
deps(llama.cpp): update ( #1714 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-21 11:35:44 +01:00
Dave
255748bcba
MQTT Startup Refactoring Part 1: core/ packages part 1 ( #1728 )
...
This PR specifically introduces a `core` folder and moves the following packages over, without any other changes:
- `api/backend`
- `api/config`
- `api/options`
- `api/schema`
Once this is merged and we confirm there's no regressions, I can migrate over the remaining changes piece by piece to split up application startup, backend services, http, and mqtt as was the goal of the earlier PRs!
2024-02-21 01:21:19 +00:00
Chakib Benziane
594eb468df
Add TTS dependency for cuda based builds fixes #1727 ( #1730 )
...
Signed-off-by: Chakib Benziane <contact@blob42.xyz >
2024-02-20 21:59:43 +01:00
Ettore Di Giacinto
960d314e4f
feat(tools): Parallel function calling ( #1726 )
...
feat(tools): support returning multiple tools choices
Fixes: https://github.com/mudler/LocalAI/issues/1275
2024-02-20 21:58:45 +01:00
Ettore Di Giacinto
ed3b50622b
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-20 19:55:36 +01:00
Ettore Di Giacinto
9f2235c208
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-19 19:49:00 +01:00
Ettore Di Giacinto
4ec50bfc41
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-19 19:03:09 +01:00
Ettore Di Giacinto
51b67a247a
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-18 13:37:16 +01:00
Steven Christou
01205fd4c0
Initial implementation of upload files api. ( #1703 )
...
* Initial implementation of upload files api.
* Move sanitize method to utils.
* Save uploaded data to uploads folder.
* Avoid loop if we do not have a purpose.
* Minor cleanup of api and fix bug where deleting duplicate filename cause error.
* Revert defer of saving config
* Moved creation of directory to startup.
* Make file names unique when storing on disk.
* Add test for files api.
* Update dependencies.
2024-02-18 10:12:02 +00:00
Ettore Di Giacinto
c72808f18b
feat(tools): support Tool calls in the API ( #1715 )
...
* feat(tools): support Tools in the API
Co-authored-by: =?UTF-8?q?Stephan=20A=C3=9Fmus?= <stephan.assmus@sap.com >
* feat(tools): support function streaming
* Adhere to new return types when using tools instead of functions
* Keep backward compatibility with function calling
* Evaluate function names in chat templates
* Disable recovery with --debug
* Correctly stream out the entire result
* Detect when llm chooses to reply and to not perform any action in SSE
* Feedback from code review
---------
Co-authored-by: =?UTF-8?q?Stephan=20A=C3=9Fmus?= <stephan.assmus@sap.com >
2024-02-17 10:00:34 +01:00
Ettore Di Giacinto
6b539a2972
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-16 15:22:35 +01:00
LocalAI [bot]
2151d21862
⬆️ Update docs version mudler/LocalAI ( #1718 )
...
* ⬆️ Update docs version mudler/LocalAI
Signed-off-by: GitHub <noreply@github.com >
* Update docs/data/version.json
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: GitHub <noreply@github.com >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-16 15:11:53 +01:00
fenfir
fb0a4c5d9a
Build docker container for ROCm ( #1595 )
...
* Dockerfile changes to build for ROCm
* Adjust linker flags for ROCm
* Update conda env for diffusers and transformers to use ROCm pytorch
* Update transformers conda env for ROCm
* ci: build hipblas images
* fixup rebase
* use self-hosted
Signed-off-by: mudler <mudler@localai.io >
* specify LD_LIBRARY_PATH only when BUILD_TYPE=hipblas
---------
Signed-off-by: mudler <mudler@localai.io >
Co-authored-by: mudler <mudler@localai.io >
2024-02-16 15:08:50 +01:00
Ettore Di Giacinto
e690bf387a
fix(tts): fix regression when supplying backend from requests ( #1713 )
...
fixes #1707
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-15 17:33:06 +01:00
Ettore Di Giacinto
5e155fb081
fix(python): pin exllama2 ( #1711 )
...
fix(python): pin python deps
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-14 21:44:12 +01:00
Ettore Di Giacinto
39a6b562cf
fix(llama.cpp): downgrade to a known working version ( #1706 )
...
sycl support is broken otherwise.
See upstream issue: https://github.com/ggerganov/llama.cpp/issues/5469
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-14 10:28:06 +01:00
Ettore Di Giacinto
c56b6ddb1c
fix(llama.cpp): disable infinite context shifting ( #1704 )
...
Infinite context loop might as well trigger an infinite loop of context
shifting if the model hallucinates and does not stop answering.
This has the unpleasant effect that the predicion never terminates,
which is the case especially on small models which tends to hallucinate.
Workarounds https://github.com/mudler/LocalAI/issues/1333 by removing
context-shifting.
See also upstream issue: https://github.com/ggerganov/llama.cpp/issues/3969
2024-02-13 21:17:21 +01:00
Sertaç Özercan
2e61ff32ad
ci: add cuda builds to release ( #1702 )
...
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
2024-02-13 08:35:39 +00:00
LocalAI [bot]
02f6e18adc
⬆️ Update ggerganov/llama.cpp ( #1700 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-12 21:43:33 +00:00
LocalAI [bot]
4436e62cf1
⬆️ Update ggerganov/llama.cpp ( #1698 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-12 09:56:04 +01:00
Ettore Di Giacinto
6e0eb96c61
fix: drop unused code ( #1697 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-11 11:28:59 +01:00
Ettore Di Giacinto
fd68bf7084
fix(vall-e-x): Fix voice cloning ( #1696 )
2024-02-11 11:20:00 +01:00
LocalAI [bot]
58cdf97361
⬆️ Update ggerganov/llama.cpp ( #1694 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-11 10:01:11 +01:00
Ettore Di Giacinto
53dbe36f32
feat(tts): respect YAMLs config file, add sycl docs/examples ( #1692 )
...
* feat(refactor): refactor config and input reading
* feat(tts): read config file for TTS
* examples(kubernetes): Add simple deployment example
* examples(kubernetes): Add simple deployment for intel arc
* docs(sycl): add sycl example
* feat(tts): do not always pick a first model
* fixups to run vall-e-x on container
* Correctly resolve backend
2024-02-10 21:37:03 +01:00
LocalAI [bot]
081bd07fd1
⬆️ Update docs version mudler/LocalAI ( #1693 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-10 21:33:14 +01:00
LocalAI [bot]
ef1306f703
⬆️ Update mudler/go-stable-diffusion ( #1674 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-09 21:59:15 +00:00
LocalAI [bot]
3196967995
⬆️ Update ggerganov/llama.cpp ( #1691 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-09 21:50:34 +00:00
Ettore Di Giacinto
3875e5e0e5
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-09 00:03:07 +01:00
LocalAI [bot]
fc8423392f
⬆️ Update ggerganov/llama.cpp ( #1688 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-09 00:02:23 +01:00
Ettore Di Giacinto
f1f6035967
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-08 20:39:00 +01:00
Ettore Di Giacinto
ddd21f1644
feat: Use ubuntu as base for container images, drop deprecated ggml-transformers backends ( #1689 )
...
* cleanup backends
* switch image to ubuntu 22.04
* adapt commands for ubuntu
* transformers cleanup
* no contrib on ubuntu
* Change test model to gguf
* ci: disable bark tests (too cpu-intensive)
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* cleanup
* refinements
* use intel base image
* Makefile: Add docker targets
* Change test model
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-08 20:12:51 +01:00
Ettore Di Giacinto
d0a6a35b55
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-07 09:40:31 +01:00
Ettore Di Giacinto
e0632f2ce2
fix(llama.cpp): downgrade to fix sycl build
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-07 00:16:52 +01:00
Ettore Di Giacinto
37e6974afe
ci: fix extra(bark) tests
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-06 20:49:28 +01:00
Ettore Di Giacinto
e23e490455
Revert "fix(Dockerfile): sycl dependencies" ( #1687 )
...
Revert "fix(Dockerfile): sycl dependencies (#1686 )"
This reverts commit f76bb8954b .
2024-02-06 20:48:29 +01:00
Ettore Di Giacinto
f76bb8954b
fix(Dockerfile): sycl dependencies ( #1686 )
...
* fix(Dockerfile): sycl dependencies
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix(ci): cleanup before running bark test
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-06 19:42:52 +01:00
Ettore Di Giacinto
d168c7c9dc
ci: cleanup worker before run ( #1685 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-06 19:42:27 +01:00
Ettore Di Giacinto
fd9d060c94
ci: fix sycl image suffix
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-06 15:52:21 +01:00
LocalAI [bot]
d8b17795d7
⬆️ Update ggerganov/llama.cpp ( #1683 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-06 09:26:01 +01:00
Ettore Di Giacinto
ea7b33b0d2
Update integrations.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-02-05 15:59:31 +01:00
LocalAI [bot]
8ace0a9ba7
⬆️ Update ggerganov/llama.cpp ( #1681 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-04 21:59:14 +00:00
Ettore Di Giacinto
98ad93d53e
Drop ggml-based gpt2 and starcoder (supported by llama.cpp) ( #1679 )
...
* Drop ggml-based gpt2 and starcoder (supported by llama.cpp)
* Update compatibility table
2024-02-04 13:15:51 +01:00
LocalAI [bot]
38e4ec0b2a
⬆️ Update ggerganov/llama.cpp ( #1678 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-04 00:55:12 +01:00
Nicolas Vermande
f083a901fe
Fix HTTP links in README.md ( #1677 )
...
Signed-off-by: Nicolas Vermande <vfiftyfive@gmail.com >
2024-02-04 00:54:49 +01:00
Ettore Di Giacinto
df13ba655c
Drop old falcon backend (deprecated) ( #1675 )
...
Drop old falcon backend
2024-02-03 13:01:13 +01:00
LocalAI [bot]
7678b25755
⬆️ Update ggerganov/llama.cpp ( #1673 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-02 21:46:26 +00:00
LocalAI [bot]
c87ca4f320
⬆️ Update ggerganov/llama.cpp ( #1669 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-02-02 19:14:03 +01:00
Ivan Smirnov
3c24a70a1b
fix (docs): fixed broken links github/ -> github.com/ ( #1672 )
...
fix broken links
2024-02-02 18:18:03 +01:00
Richard Palethorpe
e46db63e06
feat(mamba): Add bagel-dpo-2.8b ( #1671 )
...
Adds the Mamba-slimpj model fine-tuned with bagel.
https://huggingface.co/jondurbin/bagel-dpo-2.8b-v0.2
Signed-off-by: Richard Palethorpe <io@richiejp.com >
2024-02-02 18:17:44 +01:00
Ettore Di Giacinto
1c57f8d077
feat(sycl): Add support for Intel GPUs with sycl ( #1647 ) ( #1660 )
...
* feat(sycl): Add sycl support (#1647 )
* onekit: install without prompts
* set cmake args only in grpc-server
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* cleanup
* fixup sycl source env
* Cleanup docs
* ci: runs on self-hosted
* fix typo
* bump llama.cpp
* llama.cpp: update server
* adapt to upstream changes
* adapt to upstream changes
* docs: add sycl
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-02-01 19:21:52 +01:00
LocalAI [bot]
16cebf0390
⬆️ Update ggerganov/llama.cpp ( #1665 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-30 23:38:05 +00:00
Ettore Di Giacinto
555bc02665
Update codellama-7b.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-30 11:36:20 +01:00
LocalAI [bot]
c1bae1ee81
⬆️ Update ggerganov/llama.cpp ( #1656 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-30 00:43:36 +01:00
LocalAI [bot]
f2ed3df3da
⬆️ Update docs version mudler/LocalAI ( #1661 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-30 00:43:18 +01:00
LocalAI [bot]
abd678e147
⬆️ Update ggerganov/llama.cpp ( #1655 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-28 09:24:44 +01:00
Ettore Di Giacinto
6ac5d814fb
feat(startup): fetch model definition remotely ( #1654 )
2024-01-28 00:14:16 +01:00
LocalAI [bot]
f928899338
⬆️ Update ggerganov/llama.cpp ( #1652 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-27 00:13:38 +01:00
Ettore Di Giacinto
5a6fd98839
fix(paths): automatically create paths ( #1650 )
...
Especially useful when running inside a container.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-01-27 00:13:19 +01:00
Ettore Di Giacinto
072f71dfb7
Update codellama-7b.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-26 18:35:33 +01:00
Ettore Di Giacinto
670cee8274
Update transformers-tinyllama.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-26 18:29:38 +01:00
Ettore Di Giacinto
9f1be45552
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-26 17:55:20 +01:00
Ettore Di Giacinto
f1846ae5ac
Update phi-2.yaml
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-26 16:22:54 +01:00
LocalAI [bot]
ac19998e5e
⬆️ Update ggerganov/llama.cpp ( #1644 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-26 00:13:39 +01:00
Ettore Di Giacinto
cb7512734d
transformers: correctly load automodels ( #1643 )
...
* backends(transformers): use AutoModel with LLM types
* examples: animagine-xl
* Add codellama examples
2024-01-26 00:13:21 +01:00
LocalAI [bot]
3733250b3c
⬆️ Update ggerganov/llama.cpp ( #1642 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-24 22:51:59 +01:00
LocalAI [bot]
da3cd8993d
⬆️ Update docs version mudler/LocalAI ( #1631 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-24 19:50:33 +01:00
LocalAI [bot]
7690caf020
⬆️ Update ggerganov/llama.cpp ( #1632 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-23 23:07:51 +01:00
Ettore Di Giacinto
5e335eaead
feat(transformers): support also text generation ( #1630 )
...
* feat(transformers): support also text generation
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* embedded: set seed -1
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-01-23 23:07:31 +01:00
coyzeng
d5d82ba344
feat(grpc): backend SPI pluggable in embedding mode ( #1621 )
...
* run server
* grpc backend embedded support
* backend providable
2024-01-23 08:56:36 +01:00
LocalAI [bot]
efe2883c5d
⬆️ Update ggerganov/llama.cpp ( #1626 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-22 23:22:01 +01:00
LocalAI [bot]
47237c7c3c
⬆️ Update ggerganov/llama.cpp ( #1623 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-22 08:54:06 +01:00
Ettore Di Giacinto
697c769b64
fix(llama.cpp): enable cont batching when parallel is set ( #1622 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-01-21 14:59:48 +01:00
Ettore Di Giacinto
94261b1717
Update gpt-vision.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-21 10:07:30 +01:00
Sebastian
eaf85a30f9
fix(llama.cpp): Enable parallel requests ( #1616 )
...
integrate changes from llama.cpp
Signed-off-by: Sebastian <tauven@gmail.com >
2024-01-21 09:56:14 +01:00
LocalAI [bot]
6a88b030ea
⬆️ Update ggerganov/llama.cpp ( #1620 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-20 23:34:46 +01:00
LocalAI [bot]
f538416fb3
⬆️ Update docs version mudler/LocalAI ( #1619 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-20 21:37:02 +00:00
Ettore Di Giacinto
06cd9ef98d
feat(extra-backends): Improvements, adding mamba example ( #1618 )
...
* feat(extra-backends): Improvements
vllm: add max_tokens, wire up stream event
mamba: fixups, adding examples for mamba-chat
* examples(mamba-chat): add
* docs: update
2024-01-20 17:56:08 +01:00
James Braza
f3d71f8819
Modernized LlamaIndex integration ( #1613 )
...
Updated LlamaIndex example
2024-01-20 10:06:32 +01:00
James Braza
b7127c2dc9
Expanded and interlinked Docker documentation ( #1614 )
...
* Corrected dockerhub to Docker Hub
* Consolidated two Docker examples
* Linked Container Images in Manual Images
2024-01-20 10:05:14 +01:00
LocalAI [bot]
b2dc5fbd7e
⬆️ Update ggerganov/llama.cpp ( #1612 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-20 00:38:14 +01:00
Ettore Di Giacinto
9e653d6abe
feat: 🐍 add mamba support ( #1589 )
...
feat(mamba): Initial import
This is a first iteration of the mamba backend, loosely based on
mamba-chat(https://github.com/havenhq/mamba-chat ).
2024-01-19 23:42:50 +01:00
Ettore Di Giacinto
52c9a7f45d
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-19 19:30:29 +01:00
Ettore Di Giacinto
ee42c9bfe6
docs: re-use original permalinks ( #1610 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-01-19 19:23:58 +01:00
Ettore Di Giacinto
e6c3e483a1
Update build.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-19 19:09:35 +01:00
Ettore Di Giacinto
3a253c6cd7
Makefile: allow to build without GRPC_BACKENDS ( #1607 )
2024-01-19 15:38:43 +01:00
Luna Midori
e9c3bbc6d7
Update README.md ( #1601 )
...
Signed-off-by: Luna Midori <118759930+lunamidori5@users.noreply.github.com >
2024-01-19 08:55:37 +01:00
LocalAI [bot]
23d64ac53a
⬆️ Update ggerganov/llama.cpp ( #1604 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-18 21:20:50 +00:00
Ettore Di Giacinto
34f9f20ff4
Update quickstart.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-18 20:49:04 +01:00
Ettore Di Giacinto
a4a72a79ae
Update integrations.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-18 19:53:41 +01:00
Ettore Di Giacinto
6ca4d38a01
docs/examples: enhancements ( #1572 )
...
* docs: re-order sections
* fix references
* Add mixtral-instruct, tinyllama-chat, dolphin-2.5-mixtral-8x7b
* Fix link
* Minor corrections
* fix: models is a StringSlice, not a String
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* WIP: switch docs theme
* content
* Fix GH link
* enhancements
* enhancements
* Fixed how to link
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* fixups
* logo fix
* more fixups
* final touches
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
Co-authored-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2024-01-18 19:41:08 +01:00
LocalAI [bot]
b5c93f176a
⬆️ Update ggerganov/llama.cpp ( #1599 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-18 14:39:30 +01:00
LocalAI [bot]
1aaf88098d
⬆️ Update ggerganov/llama.cpp ( #1597 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-17 09:27:02 +01:00
Dionysius
6f447e613d
docs: missing golang requirement for local build for debian ( #1596 )
...
docs: fix missing golang requirement for local build for debian
2024-01-17 09:26:43 +01:00
LocalAI [bot]
dfb7c3b1aa
⬆️ Update ggerganov/llama.cpp ( #1594 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-16 14:47:57 +01:00
Dionysius
b41eb5e1f3
prepend built binaries in PATH for BUILD_GRPC_FOR_BACKEND_LLAMA ( #1593 )
...
prepend built binaries in PATH
2024-01-16 14:47:47 +01:00
LocalAI [bot]
9c2d264979
⬆️ Update ggerganov/llama.cpp ( #1590 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-15 09:01:07 +01:00
LocalAI [bot]
b996c3198c
⬆️ Update ggerganov/llama.cpp ( #1587 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-14 09:46:47 +00:00
Ettore Di Giacinto
f879c07c86
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-14 10:00:46 +01:00
Dionysius
441e2965ff
move BUILD_GRPC_FOR_BACKEND_LLAMA logic to makefile: errors in this section now immediately fail the build ( #1576 )
...
* move BUILD_GRPC_FOR_BACKEND_LLAMA option to makefile
* review: oversight, fixup cmake_args
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Signed-off-by: Dionysius <1341084+dionysius@users.noreply.github.com >
---------
Signed-off-by: Dionysius <1341084+dionysius@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-13 10:08:26 +01:00
LocalAI [bot]
cbe9a03e3c
⬆️ Update ggerganov/llama.cpp ( #1583 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-12 23:04:04 +01:00
LocalAI [bot]
4ee7e73d00
⬆️ Update ggerganov/llama.cpp ( #1578 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-12 16:04:33 +01:00
lunamidori5
1cca449726
Moving the how tos to self hosted ( #1574 )
...
* Update _index.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-setup-sd.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-setup-full.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-setup-embeddings.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-setup-docker.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-request.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-model.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.en.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update README.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos directory
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
---------
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2024-01-11 09:25:18 +01:00
LocalAI [bot]
faf7c1c325
⬆️ Update ggerganov/llama.cpp ( #1573 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-11 08:41:32 +01:00
LocalAI [bot]
58288494d6
⬆️ Update ggerganov/llama.cpp ( #1568 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-10 10:18:57 +01:00
Dionysius
72283dc744
minor: replace shell pwd in Makefile with CURDIR for better windows compatibility ( #1571 )
...
replace shell pwd in Makefile with CURDIR
2024-01-10 08:39:50 +00:00
LocalAI [bot]
b8240b4c18
⬆️ Update docs version mudler/LocalAI ( #1567 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-09 21:56:12 +01:00
Ettore Di Giacinto
5309da40b7
Update Dockerfile
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-09 08:55:43 +01:00
Ettore Di Giacinto
08b90b4720
Update _index.en.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-09 08:50:19 +01:00
LocalAI [bot]
2e890b3838
⬆️ Update ggerganov/llama.cpp ( #1563 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-09 08:48:40 +01:00
LocalAI [bot]
06656fc057
⬆️ Update docs version mudler/LocalAI ( #1562 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-09 08:48:24 +01:00
LocalAI [bot]
574fa67bdc
⬆️ Update ggerganov/llama.cpp ( #1558 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-08 00:38:03 +01:00
Ettore Di Giacinto
e19d7226f8
feat: more embedded models, coqui fixes, add model usage and description ( #1556 )
...
* feat: add model descriptions and usage
* remove default model gallery
* models: add embeddings and tts
* docs: update table
* docs: updates
* images: cleanup pip cache after install
* images: always run apt-get clean
* ux: improve gRPC connection errors
* ux: improve some messages
* fix: fix coqui when no AudioPath is passed by
* embedded: add more models
* Add usage
* Reorder table
2024-01-08 00:37:02 +01:00
LocalAI [bot]
0843fe6c65
⬆️ Update docs version mudler/LocalAI ( #1557 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-07 09:36:21 +01:00
Ettore Di Giacinto
62a02cd1fe
deps(conda): use transformers environment with autogptq ( #1555 )
2024-01-06 15:30:53 +01:00
Ettore Di Giacinto
949da7792d
deps(conda): use transformers-env with vllm,exllama(2) ( #1554 )
...
* deps(conda): use transformers with vllm
* join vllm, exllama, exllama2, split petals
2024-01-06 13:32:28 +01:00
Ettore Di Giacinto
ce724a7e55
docs: improve getting started ( #1553 )
...
* docs: improve getting started
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* cleanups
* Use dockerhub links
* Shrink command to minimum
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-06 01:04:14 +01:00
LocalAI [bot]
0a06c80801
⬆️ Update ggerganov/llama.cpp ( #1547 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-05 23:27:51 +01:00
LocalAI [bot]
edc55ade61
⬆️ Update docs version mudler/LocalAI ( #1546 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
Co-authored-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2024-01-05 23:27:30 +01:00
Ettore Di Giacinto
09e5d9007b
feat: embedded model configurations, add popular model examples, refactoring ( #1532 )
...
* move downloader out
* separate startup functions for preloading configuration files
* docs: add popular model examples
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* shorteners
* Add llava
* Add mistral-openorca
* Better link to build section
* docs: update
* fixup
* Drop code dups
* Minor fixups
* Apply suggestions from code review
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* ci: try to cache gRPC build during tests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci: do not build all images for tests, just necessary
* ci: cache gRPC also in release pipeline
* fixes
* Update model_preload_test.go
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-05 23:16:33 +01:00
Ettore Di Giacinto
db926896bd
Revert "[Refactor]: Core/API Split" ( #1550 )
...
Revert "[Refactor]: Core/API Split (#1506 )"
This reverts commit ab7b4d5ee9 .
2024-01-05 18:04:46 +01:00
Dave
ab7b4d5ee9
[Refactor]: Core/API Split ( #1506 )
...
Refactors api folder to core, creates firm split between backend code and api frontend.
2024-01-05 15:34:56 +01:00
Ettore Di Giacinto
bcf02449b3
ci(dockerhub): push images also to dockerhub ( #1542 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-04 08:32:29 +01:00
LocalAI [bot]
d48faf35ab
⬆️ Update ggerganov/llama.cpp ( #1544 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-04 00:08:03 +01:00
Ettore Di Giacinto
583bd28a5c
fix(diffusers): add omegaconf dependency ( #1540 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-04 00:06:41 +01:00
LocalAI [bot]
7e1d8c489b
⬆️ Update ggerganov/llama.cpp ( #1533 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-03 08:43:35 +01:00
LocalAI [bot]
de28867374
⬆️ Update ggerganov/llama.cpp ( #1531 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2024-01-02 00:28:22 +00:00
Ettore Di Giacinto
a1aa6cb7c2
fix(entrypoint): cd to backend dir before start ( #1530 )
...
Certain backends as vall-e-x are not meant to be used as a library, so
we want to start the process in the same folder where the backend and
all the assets are fixes #1394
2024-01-01 22:02:48 +01:00
Ettore Di Giacinto
85e2767dca
feat: add trimsuffix ( #1528 )
2024-01-01 14:39:42 +01:00
Ettore Di Giacinto
fd48cb6506
deps(llama.cpp): update and sync grpc server ( #1527 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2024-01-01 14:39:31 +01:00
Ettore Di Giacinto
522659eb59
feat(prepare): allow to specify additional files to download ( #1526 )
2024-01-01 14:39:13 +01:00
Ettore Di Giacinto
f068efe509
docs(phi-2): add example ( #1525 )
2024-01-01 10:51:47 +01:00
Ettore Di Giacinto
726fe416bb
docs: update hot topics
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2024-01-01 10:41:39 +01:00
Ettore Di Giacinto
66fa4f1767
feat: share models by url ( #1522 )
...
* feat: allow to pass by models via args
* expose it also as an env/arg
* docs: enhancements to build/requirements
* do not display status always
* print download status
* not all mesages are debug
2024-01-01 10:31:03 +01:00
Ettore Di Giacinto
d6565f3b99
Update _index.en.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-31 10:58:22 +01:00
LocalAI [bot]
27686ff20b
⬆️ Update ggerganov/llama.cpp ( #1518 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-31 00:19:08 +00:00
LocalAI [bot]
a8b865022f
⬆️ Update docs version mudler/LocalAI ( #1517 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-30 23:50:24 +00:00
Ettore Di Giacinto
c1888a8062
feat(preload): prepare models in galleries ( #1515 )
...
Previously if applying models from the gallery API, we didn't actually
allowed remote URLs as models as nothing was actually downloading the
models referenced in the configuration file. Now we call Preload after
we have all the models loaded in memory.
2023-12-30 18:55:18 +01:00
Ettore Di Giacinto
a95bb0521d
fix(download): correctly check for not found error ( #1514 )
2023-12-30 15:36:46 +01:00
Chris Natale
e2311a145c
Fix: Set proper Homebrew install location for x86 Macs ( #1510 )
...
* set proper Homebrew install location for x86 Macs
* fix: remove prior conditional that my logic replaces
2023-12-30 12:37:26 +01:00
lunamidori5
d4e0bab6be
Update version.json (2.3.0) ( #1511 )
...
Update version.json
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2023-12-30 10:19:46 +01:00
LocalAI [bot]
5b0dc20e4c
⬆️ Update ggerganov/llama.cpp ( #1509 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-30 09:19:07 +00:00
Ettore Di Giacinto
9723c3c21d
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-28 23:06:40 +01:00
Ettore Di Giacinto
9dc32275ad
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-28 23:03:44 +01:00
Ettore Di Giacinto
611c11f57b
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-28 23:03:10 +01:00
Ettore Di Giacinto
763d1f524a
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-28 23:01:52 +01:00
LocalAI [bot]
6428003c3b
⬆️ Update ggerganov/llama.cpp ( #1503 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-28 22:44:50 +01:00
LocalAI [bot]
2eac4f93bb
⬆️ Update ggerganov/llama.cpp ( #1501 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-28 00:51:29 +00:00
JZacharie
24adf9cbcb
remove default to stablediffusion ( #1500 )
2023-12-27 23:16:49 +00:00
LocalAI [bot]
c45f581c47
⬆️ Update ggerganov/llama.cpp ( #1496 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-26 19:15:58 -05:00
Ettore Di Giacinto
ae0c48e6bd
ci(apple): speedups ( #1471 )
...
* ci(apple): install grpc from brew
* ci(apple): use brew deps also on release
* ci(linux): install grpc from package manager
* ci: set concurrency
* Revert "ci(linux): install grpc from package manager"
This reverts commit 004e3e308e .
2023-12-26 19:19:37 +01:00
LocalAI [bot]
4ca649154d
⬆️ Update ggerganov/llama.cpp ( #1495 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-26 17:53:59 +00:00
Ettore Di Giacinto
66dd387858
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-25 09:04:35 +01:00
LocalAI [bot]
9789f5a96a
⬆️ Update ggerganov/llama.cpp ( #1492 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-25 02:43:35 -05:00
Gianluca Boiano
cae7b197ec
feat: add tiny dream stable diffusion support ( #1283 )
...
Signed-off-by: Gianluca Boiano <morf3089@gmail.com >
2023-12-24 19:27:24 +00:00
l
f7621b2c6c
feat: partial download ( #1486 )
...
* add .partial download
* fix Stat check
* review partial download
2023-12-24 19:39:33 +01:00
Ettore Di Giacinto
95eb72bfd3
feat: add 🐸 coqui ( #1489 )
...
* feat: add coqui
* docs: update news
2023-12-24 19:38:54 +01:00
BobMaster
7e2d101a46
fix: guidance_scale not work in sd ( #1488 )
...
Signed-off-by: hibobmaster <32976627+hibobmaster@users.noreply.github.com >
2023-12-24 19:24:52 +01:00
Sertaç Özercan
6597881854
fix: exllama2 backend ( #1484 )
...
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
2023-12-24 08:32:12 +00:00
LocalAI [bot]
eaa899df63
⬆️ Update ggerganov/whisper.cpp ( #1483 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-24 02:53:29 -05:00
LocalAI [bot]
16ed0bd0c5
⬆️ Update ggerganov/llama.cpp ( #1482 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-24 02:53:12 -05:00
Ettore Di Giacinto
939187a129
env(conda): use transformers for vall-e-x ( #1481 )
2023-12-23 14:31:34 -05:00
Ettore Di Giacinto
4b520c3343
docs: add langchain4j integration ( #1476 )
...
* docs: add langchain4j integration
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* Update docs/content/integrations/langchain4j.md
Co-authored-by: LangChain4j <langchain4j@gmail.com >
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update langchain4j.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
Co-authored-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
Co-authored-by: LangChain4j <langchain4j@gmail.com >
2023-12-23 09:13:56 +00:00
LocalAI [bot]
51215d480a
⬆️ Update ggerganov/whisper.cpp ( #1480 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-23 09:11:40 +00:00
LocalAI [bot]
987f0041d3
⬆️ Update ggerganov/llama.cpp ( #1469 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-23 00:05:56 +00:00
LocalAI [bot]
a29de9bf50
⬆️ Update donomii/go-rwkv.cpp ( #1478 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-22 15:02:32 +01:00
LocalAI [bot]
9bd5831fda
⬆️ Update ggerganov/whisper.cpp ( #1479 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-22 08:26:39 +01:00
LocalAI [bot]
59f0f2f0fd
⬆️ Update docs version mudler/LocalAI ( #1477 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-22 00:28:42 +00:00
Ettore Di Giacinto
9ae47d37e9
pin go-rwkv
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-21 08:42:40 +01:00
Ettore Di Giacinto
2b3ad7f41c
Revert " ⬆️ Update donomii/go-rwkv.cpp" ( #1474 )
...
Revert "⬆️ Update donomii/go-rwkv.cpp (#1470 )"
This reverts commit 51db10b18f .
2023-12-21 08:38:50 +01:00
LocalAI [bot]
51db10b18f
⬆️ Update donomii/go-rwkv.cpp ( #1470 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-21 08:35:31 +01:00
Ettore Di Giacinto
b4b21a446b
feat(conda): share envs with transformer-based backends ( #1465 )
...
* feat(conda): share env between diffusers and bark
* Detect if env already exists
* share diffusers and petals
* tests: add petals
* Use smaller model for tests with petals
* test only model load on petals
* tests(petals): run only load model tests
* Revert "test only model load on petals"
This reverts commit 111cfa97f1 .
* move transformers and sentencetransformers to common env
* Share also transformers-musicgen
2023-12-21 08:35:15 +01:00
LocalAI [bot]
23eced1644
⬆️ Update ggerganov/llama.cpp ( #1461 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-20 18:02:52 +01:00
LocalAI [bot]
7741a6e75d
⬆️ Update ggerganov/whisper.cpp ( #1462 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-20 00:21:49 +00:00
LocalAI [bot]
d4210db0c9
⬆️ Update ggerganov/llama.cpp ( #1457 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-19 00:42:19 +01:00
lunamidori5
17dde75107
How To (Updates and Fixes) ( #1456 )
...
* Update easy-setup-embeddings.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update and rename easy-setup-docker-cpu.md to easy-setup-docker.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-sd.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-sd.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
---------
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2023-12-18 18:59:08 +01:00
Ettore Di Giacinto
1fc3a375df
feat: inline templates and accept URLs in models ( #1452 )
...
* feat: Allow inline templates
* feat: Allow to specify url in model config files
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* feat: support 'huggingface://' format
* style: reuse-code from gallery
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-18 18:58:44 +01:00
LocalAI [bot]
64a8471dd5
⬆️ Update ggerganov/llama.cpp ( #1455 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-18 08:55:29 +01:00
LocalAI [bot]
86a8df1c8b
⬆️ Update ggerganov/llama.cpp ( #1450 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-17 19:02:28 +01:00
Ettore Di Giacinto
2eeed2287b
docs: automatically track latest versions ( #1451 )
2023-12-17 19:02:13 +01:00
Ettore Di Giacinto
3d83128f16
feat(alias): alias llama to llama-cpp, update docs ( #1448 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-16 18:22:45 +01:00
Ettore Di Giacinto
1c286c3c2f
docs(mixtral): add mixtral example ( #1449 )
2023-12-16 17:44:43 +01:00
LocalAI [bot]
2f7beb6744
⬆️ Update ggerganov/whisper.cpp ( #1434 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-16 09:22:28 +01:00
LocalAI [bot]
ab0370a0b9
⬆️ Update ggerganov/llama.cpp ( #1429 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-16 09:22:13 +01:00
LocalAI [bot]
3f9a41684a
⬆️ Update mudler/go-piper ( #1441 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-16 09:21:56 +01:00
Ettore Di Giacinto
dd982acf2c
feat(img2vid,txt2vid): Initial support for img2vid,txt2vid ( #1442 )
...
* feat(img2vid): Initial support for img2vid
* doc(SD): fix SDXL Example
* Minor fixups for img2vid
* docs(img2img): fix example curl call
* feat(txt2vid): initial support
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* diffusers: be retro-compatible with CUDA settings
* docs(img2vid, txt2vid): examples
* Add notice on docs
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-15 18:06:20 -05:00
Ettore Di Giacinto
fb6a5bc620
update(llama.cpp): update server, correctly propagate LLAMA_VERSION ( #1440 )
...
* fix(Makefile): correctly propagate LLAMA_VERSION
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* update grpc-server.cpp
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-15 08:26:48 +01:00
Ettore Di Giacinto
7641f92cde
feat(diffusers): update, add autopipeline, controlnet ( #1432 )
...
* feat(diffusers): update, add autopipeline, controlenet
* tests with AutoPipeline
* simplify logic
2023-12-13 19:20:22 +01:00
LocalAI [bot]
72325fd0a3
⬆️ Update ggerganov/whisper.cpp ( #1430 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-13 08:37:02 +01:00
Sertaç Özercan
1b7ed5e2e6
docs: add aikit to integrations ( #1412 )
...
* docs: add aikit to integrations
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
* docs: add to readme
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
---------
Signed-off-by: Sertac Ozercan <sozercan@gmail.com >
Co-authored-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2023-12-12 18:58:57 +01:00
LocalAI [bot]
86fac272d8
⬆️ Update ggerganov/llama.cpp ( #1391 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-12 18:22:48 +01:00
Samuel Walker
865e523ff1
Documentation for Hipblas ( #1425 )
...
hiplas arch
2023-12-12 15:05:01 +01:00
Ettore Di Giacinto
9aa2a7ca13
extras: add vllm,bark,vall-e-x tests, bump diffusers ( #1422 )
...
* tests: add vllm
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* tests: Add vall-e-x tests
* Add bark tests
* bump diffusers
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-12 00:39:26 +01:00
Ettore Di Giacinto
e80cbca6b0
Update _index.en.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-12 00:37:01 +01:00
Ettore Di Giacinto
718a5d4a9e
fix(transformers*): add sentence-transformers and transformers-musicgen tests, fix musicgen wrapper ( #1420 )
...
* tests: add sentence-transformers and transformers-musicgen
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* fix: tranformers-musicgen conda env
Initialize correctly the environment for the transformers-musicgen backend.
* fix(tests): transformer-musicgen tests fixups
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-11 19:26:02 +01:00
lunamidori5
9222bec8b1
How To Updates / Model Used Switched / Removed "docker-compose" (RIP) ( #1417 )
...
* Update _index.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-model.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.en.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-model.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.en.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
---------
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2023-12-11 14:27:29 +00:00
LocalAI [bot]
4a965e1b0e
⬆️ Update ggerganov/whisper.cpp ( #1418 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-11 08:24:48 +01:00
Ettore Di Giacinto
48e5380e45
tests: add diffusers tests ( #1419 )
2023-12-11 08:20:34 +01:00
LocalAI [bot]
831418612b
⬆️ Update mudler/go-piper ( #1400 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-10 08:50:26 +01:00
LocalAI [bot]
89ff12309d
⬆️ Update ggerganov/whisper.cpp ( #1390 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-09 09:23:40 +01:00
Ettore Di Giacinto
3a4fb6fa4b
feat(entrypoint): optionally prepare extra endpoints ( #1405 )
...
entrypoint: optionally prepare extra endpoints
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-12-08 20:04:13 +01:00
Ettore Di Giacinto
b181503c30
docs: update v2.0.0 notes
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-08 16:06:24 +01:00
Ettore Di Giacinto
887b3dff04
feat: cuda transformers ( #1401 )
...
* Use cuda in transformers if available
tensorflow probably needs a different check.
Signed-off-by: Erich Schubert <kno10@users.noreply.github.com >
* feat: expose CUDA at top level
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* tests: add to tests and create workflow for py extra backends
* doc: update note on how to use core images
---------
Signed-off-by: Erich Schubert <kno10@users.noreply.github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: Erich Schubert <kno10@users.noreply.github.com >
2023-12-08 15:45:04 +01:00
Ettore Di Giacinto
3822bd2369
docs: updates
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-08 15:11:44 +01:00
Ettore Di Giacinto
4de2c6a421
docs: update news
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-08 14:59:25 +01:00
Ettore Di Giacinto
6c4231fd35
docs: 2.0 updates
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-08 14:58:53 +01:00
lunamidori5
adfa7aa1fa
docs: site update fixing old image text / How To update updating GPU and CPU docker pages ( #1399 )
...
* Update _index.en.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
---------
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2023-12-08 10:27:21 +01:00
Dave
8b6e601405
Feat: new backend: transformers-musicgen ( #1387 )
...
Transformers-MusicGen
---------
Signed-off-by: Dave <dave@gray101.com >
2023-12-08 10:01:02 +01:00
Ettore Di Giacinto
6011911746
fix(piper): pin petals, phonemize and espeak ( #1393 )
...
* fix: pin phonemize and espeak
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: pin petals deps
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-12-07 22:58:41 +01:00
LocalAI [bot]
997119c27a
⬆️ Update ggerganov/llama.cpp ( #1385 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-05 15:44:24 +01:00
Dave
2eb6865a27
Fix: API Key / JSON Fast Follow #1 ( #1388 )
...
fast follow fix #1 - imports, final loop, one last chance to skip
Co-authored-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2023-12-05 10:35:27 +00:00
Ettore Di Giacinto
2b2d6673ff
exllama(v2): fix exllamav1, add exllamav2 ( #1384 )
...
* fix(exllama): fix exllama deps with anaconda
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(exllamav2): add exllamav2 backend
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-12-05 08:15:37 +01:00
lunamidori5
563c5b7ea0
Added Check API KEYs file to API.go ( #1381 )
...
Added API KEYs file
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2023-12-04 22:06:45 -05:00
LocalAI [bot]
67966b623c
⬆️ Update ggerganov/llama.cpp ( #1379 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-04 18:36:34 +01:00
LocalAI [bot]
9fc3fd04be
⬆️ Update ggerganov/whisper.cpp ( #1378 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-04 18:36:22 +01:00
Ettore Di Giacinto
238fec244a
fix(vall-e-x): correctly install reqs in environment ( #1377 )
2023-12-03 21:16:36 +01:00
LocalAI [bot]
3d71bc9b64
⬆️ Update ggerganov/whisper.cpp ( #1227 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-03 01:16:07 +01:00
Felix Erkinger
3923024d84
update whisper_cpp with CUBLAS, HIPBLAS, METAL, OPENBLAS, CLBLAST support ( #1302 )
...
update whisper_cpp to 1.5.1 with OPENBLAS, METAL, HIPBLAS, CUBLAS, CLBLAST support
2023-12-02 10:10:18 +00:00
Ettore Di Giacinto
710b195be1
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-02 08:55:26 +01:00
Ettore Di Giacinto
6e408137ee
Update fine-tuning.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-02 08:54:21 +01:00
Ettore Di Giacinto
9b205cfcfc
Update fine-tuning.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-02 08:52:00 +01:00
LocalAI [bot]
42a80d1b8b
⬆️ Update ggerganov/llama.cpp ( #1375 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-12-02 00:09:48 +00:00
Ettore Di Giacinto
d6073ac18e
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-01 20:05:58 +01:00
Ettore Di Giacinto
1c450d46cf
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-12-01 20:01:07 +01:00
lunamidori5
6b312a8522
Site Clean up - How to Clean up ( #1342 )
...
* Create easy-request.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-request.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-request.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-request.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-request.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-request.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-request-curl.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-request-openai-v0.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-request-openai-v1.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-request.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-request-openai-v1.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-request-openai-v0.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-request-curl.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update and rename easy-model-import-downloaded.md to easy-model.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-gpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-setup-docker-cpu.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/autogen-setup.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Delete docs/content/howtos/easy-request-autogen.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update easy-model.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.en.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.en.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.en.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.en.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
* Update _index.md
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
---------
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2023-12-01 19:12:21 +01:00
Ettore Di Giacinto
2b2007ae9e
docs: add fine-tuning example ( #1374 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-12-01 19:11:45 +01:00
Dave
e94a34be8c
fix: OSX Build Fix Part 1: Metal ( #1365 )
...
* Make Metal the default on OSX, simplify osx-specific code, and fix the file copy error.
* fix endif / comment
2023-11-30 19:50:50 +01:00
Ettore Di Giacinto
c3fb4b1d8e
ci: rename workflow
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-11-30 19:25:33 +01:00
Ettore Di Giacinto
e3ca1a7dbe
ci: split into reusable workflows ( #1366 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-30 19:24:37 +01:00
B4ckslash
2d64d8b444
fix/docs: Python backend dependencies ( #1360 )
...
* Update docs for new requirements.txt path
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com >
* Fix typo (.PONY -> .PHONY) in python backend makefiles
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com >
---------
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com >
2023-11-30 17:46:55 +01:00
Ettore Di Giacinto
9b98be160a
ci: limit concurrent jobs ( #1364 )
...
* ci: limit concurrent image push
* docs: mention core images
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-30 17:45:20 +01:00
LocalAI [bot]
9f708ff318
⬆️ Update ggerganov/llama.cpp ( #1363 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-30 00:06:28 +01:00
Ettore Di Giacinto
4e0ad33d92
docs: Update getting started and GPU section ( #1362 )
2023-11-29 18:51:57 +01:00
LocalAI [bot]
519285bf38
⬆️ Update ggerganov/llama.cpp ( #1351 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-29 08:29:03 +01:00
Ettore Di Giacinto
fd1b7b3f22
docs: Add docker instructions, add community projects section in README ( #1359 )
...
docs: Add docker instructions
2023-11-28 23:14:16 +01:00
Gianluca Boiano
687730a7f5
fix: go-piper add libucd at linking time ( #1357 )
...
Signed-off-by: Gianluca Boiano <morf3089@gmail.com >
2023-11-28 19:55:09 +00:00
Ettore Di Giacinto
b7821361c3
feat(petals): add backend ( #1350 )
...
* feat(petals): add backend
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fixups
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-28 09:01:46 +01:00
LocalAI [bot]
63e1f8fffd
⬆️ Update ggerganov/llama.cpp ( #1345 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-27 09:02:19 +01:00
Ettore Di Giacinto
824612f1b4
feat: initial watchdog implementation ( #1341 )
...
* feat: initial watchdog implementation
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
* fiuxups
* Add more output
* wip: idletime checker
* wire idle watchdog checks
* enlarge watchdog time window
* small fixes
* Use stopmodel
* Always delete process
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-26 18:36:23 +01:00
LocalAI [bot]
9482acfdfc
⬆️ Update ggerganov/llama.cpp ( #1340 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-26 09:27:42 +01:00
Ettore Di Giacinto
c75bdd99e4
fix: rename transformers.py to avoid circular import ( #1337 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-26 08:49:43 +01:00
Ettore Di Giacinto
6f34e8f044
fix: propagate CMAKE_ARGS when building grpc ( #1334 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-25 13:53:51 +01:00
Ettore Di Giacinto
6d187af643
fix: handle grpc and llama-cpp with REBUILD=true ( #1328 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-25 08:48:24 +01:00
LocalAI [bot]
97e9598c79
⬆️ Update ggerganov/llama.cpp ( #1330 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-24 23:45:05 +01:00
B4ckslash
5a6a6de3d7
docs: Update Features->Embeddings page to reflect backend restructuring ( #1325 )
...
* Update path to sentencetransformers backend for local execution
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com >
* Rename huggingface-embeddings -> sentencetransformers in embeddings.md for consistency with the backend structure
The Dockerfile still knows the "huggingface-embeddings"
backend (I assume for compatibility reasons) but uses the
sentencetransformers backend under the hood anyway.
I figured it would be good to update the docs to use the new naming to
make it less confusing moving forward. As the docker container knows
both the "huggingface-embeddings" and the "sentencetransformers"
backend, this should not break anything.
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com >
---------
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com >
2023-11-24 18:21:04 +01:00
LocalAI [bot]
b1a20effde
⬆️ Update ggerganov/llama.cpp ( #1323 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-24 08:32:36 +01:00
Ettore Di Giacinto
ba5ab26f2e
docs: Add llava, update hot topics ( #1322 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-23 18:54:55 +01:00
Dave
69f53211a1
Feat: OSX Local Codesigning ( #1319 )
...
* stage makefile
* OSX local code signing and entitlements file to fix incoming connections prompt
2023-11-23 15:22:54 +01:00
B4ckslash
9dddd1134d
fix: move python header comments below shebang in some backends ( #1321 )
...
* Fix python header comments for some extra gRPC backends
When a Python script is to be executed directly via exec(3), either the platform knows how to execute
the file itself (i.e. special configuration is necessary) or the first line
contains a shebang (#!) specifying the interpreter to run it (similar to
shell scripts).
The shebang MUST be on the first line for the script to work on all platforms,
so any header comments need to be in the lines following it. Otherwise
executing these scripts as extra backends will yield an "exec format
error" message.
Changes:
* Move introductory comments below the shebang line
* Change header comment in transformers.py to refer to the correct
python module
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com >
* Make header comment in ttsbark.py more specific
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com >
---------
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com >
2023-11-23 15:22:37 +01:00
Ettore Di Giacinto
c5c77d2b0d
docs: Initial import from localai-website ( #1312 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-22 18:13:50 +01:00
LocalAI [bot]
763f94ca80
⬆️ Update ggerganov/llama.cpp ( #1313 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-22 08:37:11 +01:00
ok2sh
20d637e7b7
fix: ExLlama Backend Context Size & Rope Scaling ( #1311 )
...
* fix: context_size not propagated to exllama backend
* fix: exllama rope scaling
2023-11-21 19:26:39 +01:00
LocalAI [bot]
480b14c8dc
⬆️ Update ggerganov/llama.cpp ( #1310 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-21 00:20:37 +01:00
Ettore Di Giacinto
999db4301a
ci(core): add -core images without python deps ( #1309 )
...
* ci(core): add -core images without python deps
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci(core): use public runners
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-20 23:01:31 +01:00
Ettore Di Giacinto
92cbc4d516
feat(transformers): add embeddings with Automodel ( #1308 )
...
* Update huggingface.py
Switch SentenceTransformer for AutoModel in order to set trust_remote_code needed to use the encode method with embeddings models like jinai-v2
Signed-off-by: Lucas Hänke de Cansino <lhc@next-boss.eu >
* feat(transformers): split in separate backend
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Lucas Hänke de Cansino <lhc@next-boss.eu >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: Lucas Hänke de Cansino <lhc@next-boss.eu >
2023-11-20 21:21:17 +01:00
LocalAI [bot]
ff9afdb0fe
⬆️ Update ggerganov/llama.cpp ( #1306 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-20 08:16:00 +01:00
LocalAI [bot]
3e35b20a02
⬆️ Update mudler/go-piper ( #1305 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-19 09:01:40 +01:00
LocalAI [bot]
9ea371d6cd
⬆️ Update ggerganov/llama.cpp ( #1304 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-19 08:49:05 +01:00
Ettore Di Giacinto
7a0f9767da
docs: fix heading
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-11-18 15:04:00 +01:00
Ettore Di Giacinto
9d7363f2a7
docs: update configuration readme
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-11-18 15:03:15 +01:00
Ettore Di Giacinto
8ee5cf38fd
Delete examples/configurations/llava/README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-11-18 15:01:39 +01:00
Ettore Di Giacinto
a6b788d220
docs: update LLaVa instructions
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-11-18 15:01:16 +01:00
lunamidori5
ccd87cd9f0
llava.yaml (yaml format standardization) ( #1303 )
...
Signed-off-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
2023-11-18 14:48:54 +01:00
LocalAI [bot]
b5af87fc6c
⬆️ Update ggerganov/llama.cpp ( #1300 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-18 08:19:10 +01:00
Ettore Di Giacinto
3c9544b023
refactor: rename llama-stable to llama-ggml ( #1287 )
...
* refactor: rename llama-stable to llama-ggml
* Makefile: get sources in sources/
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fixup path
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fixup sources
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fixups sd
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* update SD
* fixup
* fixup: create piper libdir also when not built
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix make target on linux test
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-18 08:18:43 +01:00
Mathias
2f65671070
fix(api/config): allow YAML config with .yml ( #1299 )
...
This commit allow to use both `.yml` and `.yaml` extensions for YAML configuration files as
it is usually expected.
2023-11-17 22:47:30 +01:00
LocalAI [bot]
8c5436cbed
⬆️ Update ggerganov/llama.cpp ( #1297 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-17 08:45:22 +01:00
Ettore Di Giacinto
548959b50f
feat: queue up requests if not running parallel requests ( #1296 )
...
Return a GRPC which handles a lock in case it is not meant to be
parallel.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-16 22:20:16 +01:00
LocalAI [bot]
2addb9f99a
⬆️ Update ggerganov/llama.cpp ( #1291 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-16 08:20:26 +01:00
Ettore Di Giacinto
fdd95d1d86
feat: allow to run parallel requests ( #1290 )
...
* feat: allow to run parallel requests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fixup
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-16 08:20:05 +01:00
Ettore Di Giacinto
66a558ff41
fix: respect OpenAI spec for response format ( #1289 )
...
fix: properly respect OpenAI spec for response format
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-15 19:36:23 +01:00
LocalAI [bot]
733b612eb2
⬆️ Update ggerganov/llama.cpp ( #1288 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-15 18:41:09 +01:00
LocalAI [bot]
991ecce004
⬆️ Update ggerganov/llama.cpp ( #1285 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-14 18:23:09 +01:00
Ettore Di Giacinto
ad0e30bca5
refactor: move backends into the backends directory ( #1279 )
...
* refactor: move backends into the backends directory
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* refactor: move main close to implementation for every backend
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-13 22:40:16 +01:00
LocalAI [bot]
55461188a4
⬆️ Update ggerganov/llama.cpp ( #1282 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-13 00:48:26 +00:00
LocalAI [bot]
5d2405fdef
⬆️ Update ggerganov/llama.cpp ( #1280 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-11 23:26:54 +00:00
LocalAI [bot]
e9f1268225
⬆️ Update ggerganov/llama.cpp ( #1272 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-11 20:00:28 +00:00
Ettore Di Giacinto
803a0ac02a
feat(llama.cpp): support lora with scale and yarn ( #1277 )
...
* feat(llama.cpp): support lora with scale
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(llama.cpp): support yarn
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-11 18:40:48 +01:00
Gianluca Boiano
bde87d00b9
deps(go-piper): update to 2023.11.6-3 ( #1257 )
...
Signed-off-by: Gianluca Boiano <morf3089@gmail.com >
2023-11-11 18:40:26 +01:00
Ettore Di Giacinto
0eae727366
🔥 add LaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types ( #1254 )
...
* wip
* wip
* Make it functional
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* wip
* Small fixups
* do not inject space on role encoding, encode img at beginning of messages
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add examples/config defaults
* Add include dir of current source dir
* cleanup
* fixes
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fixups
* Revert "fixups"
This reverts commit f1a4731cca .
* fixes
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-11 13:14:59 +01:00
LocalAI [bot]
3b4c5d54d8
⬆️ Update ggerganov/llama.cpp ( #1265 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-10 08:50:42 +01:00
LocalAI [bot]
4e16bc2f13
⬆️ Update ggerganov/llama.cpp ( #1256 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-08 08:21:12 +01:00
LocalAI [bot]
562ac62f59
⬆️ Update ggerganov/llama.cpp ( #1242 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-07 08:37:55 +01:00
Diego
e7fa2e06f8
Fixes the bug 1196 ( #1232 )
...
* Current state of the branch.
* Now gRPC is build only when the BUILD_GRPC_FOR_BACKEND_LLAMA variable is defined.
* Now the local compilation of gRPC is executed on BUILD_GRPC_FOR_BACKEND_LLAMA.
* Revised the Makefile.
* Removed replace directives in go.mod.
---------
Signed-off-by: Diego <38375572+diego-minguzzi@users.noreply.github.com >
Co-authored-by: lunamidori5 <118759930+lunamidori5@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-11-06 19:07:46 +01:00
Ettore Di Giacinto
8123f009d0
dockerfile: fixup duplicate
...
This should have been "exllama"
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-05 14:09:31 +01:00
Ettore Di Giacinto
622aaa9f7d
dockerfile: avoid pushing a big layer
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-05 10:31:33 +01:00
Ettore Di Giacinto
7b1ee203ce
tests: re-add flake-attempts
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-11-05 09:01:03 +01:00
Ettore Di Giacinto
f347e51927
feat(conda): conda environments ( #1144 )
...
* feat(autogptq): add a separate conda environment for autogptq (#1137 )
**Description**
This PR related to #1117
**Notes for Reviewers**
Here we lock down the version of the dependencies. Make sure it can be
used all the time without failed if the version of dependencies were
upgraded.
I change the order of importing packages according to the pylint, and no
change the logic of code. It should be ok.
I will do more investigate on writing some test cases for every backend.
I can run the service in my environment, but there is not exist a way to
test it. So, I am not confident on it.
Add a README.md in the `grpc` root. This is the common commands for
creating `conda` environment. And it can be used to the reference file
for creating extral gRPC backend document.
Signed-off-by: GitHub <noreply@github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* [Extra backend] Add seperate environment for ttsbark (#1141 )
**Description**
This PR relates to #1117
**Notes for Reviewers**
Same to the latest PR:
* The code is also changed, but only the order of the import package
parts. And some code comments are also added.
* Add a configuration of the `conda` environment
* Add a simple test case for testing if the service can be startup in
current `conda` environment. It is succeed in VSCode, but the it is not
out of box on terminal. So, it is hard to say the test case really
useful.
**[Signed
commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)**
- [x] Yes, I signed my commits.
<!--
Thank you for contributing to LocalAI!
Contributing Conventions
-------------------------
The draft above helps to give a quick overview of your PR.
Remember to remove this comment and to at least:
1. Include descriptive PR titles with [<component-name>] prepended. We
use [conventional
commits](https://www.conventionalcommits.org/en/v1.0.0/ ).
2. Build and test your changes before submitting a PR (`make build`).
3. Sign your commits
4. **Tag maintainer:** for a quicker response, tag the relevant
maintainer (see below).
5. **X/Twitter handle:** we announce bigger features on X/Twitter. If
your PR gets announced, and you'd like a mention, we'll gladly shout you
out!
By following the community's contribution conventions upfront, the
review process will
be accelerated and your PR merged more quickly.
If no one reviews your PR within a few days, please @-mention @mudler.
-->
Signed-off-by: GitHub <noreply@github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(conda): add make target and entrypoints for the dockerfile
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(conda): Add seperate conda env for diffusers (#1145 )
**Description**
This PR relates to #1117
**Notes for Reviewers**
* Add `conda` env `diffusers.yml`
* Add Makefile to create it automatically
* Add `run.sh` to support running as a extra backend
* Also adding it to the main Dockerfile
* Add make command in the root Makefile
* Testing the server, it can start up under the env
Signed-off-by: GitHub <noreply@github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(conda):Add seperate env for vllm (#1148 )
**Description**
This PR is related to #1117
**Notes for Reviewers**
* The gRPC server can be started as normal
* The test case can be triggered in VSCode
* Same to other this kind of PRs, add `vllm.yml` Makefile and add
`run.sh` to the main Dockerfile, and command to the main Makefile
**[Signed
commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)**
- [x] Yes, I signed my commits.
<!--
Thank you for contributing to LocalAI!
Contributing Conventions
-------------------------
The draft above helps to give a quick overview of your PR.
Remember to remove this comment and to at least:
1. Include descriptive PR titles with [<component-name>] prepended. We
use [conventional
commits](https://www.conventionalcommits.org/en/v1.0.0/ ).
2. Build and test your changes before submitting a PR (`make build`).
3. Sign your commits
4. **Tag maintainer:** for a quicker response, tag the relevant
maintainer (see below).
5. **X/Twitter handle:** we announce bigger features on X/Twitter. If
your PR gets announced, and you'd like a mention, we'll gladly shout you
out!
By following the community's contribution conventions upfront, the
review process will
be accelerated and your PR merged more quickly.
If no one reviews your PR within a few days, please @-mention @mudler.
-->
Signed-off-by: GitHub <noreply@github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(conda):Add seperate env for huggingface (#1146 )
**Description**
This PR is related to #1117
**Notes for Reviewers**
* Add conda env `huggingface.yml`
* Change the import order, and also remove the no-used packages
* Add `run.sh` and `make command` to the main Dockerfile and Makefile
* Add test cases for it. It can be triggered and succeed under VSCode
Python extension but it is hang by using `python -m unites
test_huggingface.py` in the terminal
```
Running tests (unittest): /workspaces/LocalAI/extra/grpc/huggingface
Running tests: /workspaces/LocalAI/extra/grpc/huggingface/test_huggingface.py::TestBackendServicer::test_embedding
/workspaces/LocalAI/extra/grpc/huggingface/test_huggingface.py::TestBackendServicer::test_load_model
/workspaces/LocalAI/extra/grpc/huggingface/test_huggingface.py::TestBackendServicer::test_server_startup
./test_huggingface.py::TestBackendServicer::test_embedding Passed
./test_huggingface.py::TestBackendServicer::test_load_model Passed
./test_huggingface.py::TestBackendServicer::test_server_startup Passed
Total number of tests expected to run: 3
Total number of tests run: 3
Total number of tests passed: 3
Total number of tests failed: 0
Total number of tests failed with errors: 0
Total number of tests skipped: 0
Finished running tests!
```
**[Signed
commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)**
- [x] Yes, I signed my commits.
<!--
Thank you for contributing to LocalAI!
Contributing Conventions
-------------------------
The draft above helps to give a quick overview of your PR.
Remember to remove this comment and to at least:
1. Include descriptive PR titles with [<component-name>] prepended. We
use [conventional
commits](https://www.conventionalcommits.org/en/v1.0.0/ ).
2. Build and test your changes before submitting a PR (`make build`).
3. Sign your commits
4. **Tag maintainer:** for a quicker response, tag the relevant
maintainer (see below).
5. **X/Twitter handle:** we announce bigger features on X/Twitter. If
your PR gets announced, and you'd like a mention, we'll gladly shout you
out!
By following the community's contribution conventions upfront, the
review process will
be accelerated and your PR merged more quickly.
If no one reviews your PR within a few days, please @-mention @mudler.
-->
Signed-off-by: GitHub <noreply@github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(conda): Add the seperate conda env for VALL-E X (#1147 )
**Description**
This PR is related to #1117
**Notes for Reviewers**
* The gRPC server cannot start up
```
(ttsvalle) @Aisuko ➜ /workspaces/LocalAI (feat/vall-e-x) $ /opt/conda/envs/ttsvalle/bin/python /workspaces/LocalAI/extra/grpc/vall-e-x/ttsvalle.py
Traceback (most recent call last):
File "/workspaces/LocalAI/extra/grpc/vall-e-x/ttsvalle.py", line 14, in <module>
from utils.generation import SAMPLE_RATE, generate_audio, preload_models
ModuleNotFoundError: No module named 'utils'
```
The installation steps follow
https://github.com/Plachtaa/VALL-E-X#-installation below:
* Under the `ttsvalle` conda env
```
git clone https://github.com/Plachtaa/VALL-E-X.git
cd VALL-E-X
pip install -r requirements.txt
```
**[Signed
commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)**
- [x] Yes, I signed my commits.
<!--
Thank you for contributing to LocalAI!
Contributing Conventions
-------------------------
The draft above helps to give a quick overview of your PR.
Remember to remove this comment and to at least:
1. Include descriptive PR titles with [<component-name>] prepended. We
use [conventional
commits](https://www.conventionalcommits.org/en/v1.0.0/ ).
2. Build and test your changes before submitting a PR (`make build`).
3. Sign your commits
4. **Tag maintainer:** for a quicker response, tag the relevant
maintainer (see below).
5. **X/Twitter handle:** we announce bigger features on X/Twitter. If
your PR gets announced, and you'd like a mention, we'll gladly shout you
out!
By following the community's contribution conventions upfront, the
review process will
be accelerated and your PR merged more quickly.
If no one reviews your PR within a few days, please @-mention @mudler.
-->
Signed-off-by: GitHub <noreply@github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: set image type
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(conda):Add seperate conda env for exllama (#1149 )
Add seperate env for exllama
Signed-off-by: Aisuko <urakiny@gmail.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Setup conda
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Set image_type arg
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci: prepare only conda env in tests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Dockerfile: comment manual pip calls
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* conda: add conda to PATH
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fixes
* add shebang
* Fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* file perms
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* debug
* Install new conda in the worker
* Disable GPU tests for now until the worker is back
* Rename workflows
* debug
* Fixup conda install
* fixup(wrapper): pass args
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: GitHub <noreply@github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: Aisuko <urakiny@gmail.com >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: Aisuko <urakiny@gmail.com >
2023-11-04 15:30:32 +01:00
LocalAI [bot]
9b17af18b3
⬆️ Update ggerganov/llama.cpp ( #1236 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-03 19:23:53 +01:00
Samuel Walker
23c7fbfe6b
chianlit example ( #1238 )
2023-11-02 22:56:46 +01:00
Samuel Walker
035fea676a
llama index example ( #1237 )
2023-11-02 13:35:06 -07:00
Vitor Oliveira
6e1a234d15
feat(certificates): add support for custom CA certificates ( #880 )
...
This change facilitates users working behind corporate firewalls or proxies. By allowing the integration of custom CA certificates, users can handle SSL connections that are intercepted by company infrastructure.
2023-11-01 20:10:14 +01:00
LocalAI [bot]
5b596ea605
⬆️ Update ggerganov/llama.cpp ( #1231 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-11-01 12:44:34 +00:00
Dave
6bd56460de
Update .gitignore for backend/llama.cpp ( #1235 )
...
Signed-off-by: Dave <dave@gray101.com >
2023-11-01 09:52:02 +01:00
LocalAI [bot]
6ef7ea2635
⬆️ Update ggerganov/llama.cpp ( #1207 )
...
Signed-off-by: GitHub <noreply@github.com >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-10-30 08:00:36 +00:00
Ettore Di Giacinto
f8c00fbaf1
ci: enlarge download timeout window
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-29 22:09:35 +01:00
Ettore Di Giacinto
d9a42cc4c5
ci: run only cublas on selfhosted ( #1224 )
...
* ci: run only cublas on selfhosted
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* debug
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* update git
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* change testing embeddings model link
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-29 22:04:43 +01:00
Ettore Di Giacinto
fc0bc32814
ci: use self-hosted to build container images ( #1206 )
...
ci: use self-hosted
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-10-26 21:13:40 +02:00
Ettore Di Giacinto
c62504ac92
cleanup: drop bloomz and ggllm as now supported by llama.cpp ( #1217 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-26 07:43:31 +02:00
Ettore Di Giacinto
f227e918f9
feat(llama.cpp): Bump llama.cpp, adapt grpc server ( #1211 )
...
* feat(llama.cpp): Bump llama.cpp, adapt grpc server
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci: fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-25 20:56:25 +02:00
Ettore Di Giacinto
c132dbadce
docs(examples): Add mistral example ( #1214 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-25 20:56:12 +02:00
Dave
b839eb80a1
Fix backend/cpp/llama CMakeList.txt on OSX ( #1212 )
...
* Fix backend/cpp/llama CMakeList.txt on OSX - detect OSX and use homebrew libraries
* sneak a logging fix in too for gallery debugging
* additional logging
2023-10-25 20:53:26 +02:00
renovate[bot]
23b03a7f03
fix(deps): update module github.com/onsi/gomega to v1.28.1 ( #1205 )
...
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2023-10-24 09:16:02 +02:00
LocalAI [bot]
9196583651
⬆️ Update ggerganov/llama.cpp ( #1204 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-10-23 19:06:39 +02:00
Ettore Di Giacinto
fd28252e55
fix(Dockerfile): try to save some space
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-22 17:13:39 +02:00
renovate[bot]
94f20e2eb7
fix(deps): update github.com/nomic-ai/gpt4all/gpt4all-bindings/golang digest to c25dc51 ( #1191 )
...
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2023-10-22 16:58:45 +02:00
Ettore Di Giacinto
5ced99a8e7
ci: more cleanup for workers
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2023-10-22 12:27:04 +02:00
LocalAI [bot]
c377e61ff0
⬆️ Update go-skynet/go-llama.cpp ( #1156 )
...
Signed-off-by: GitHub <noreply@github.com >
Co-authored-by: mudler <mudler@users.noreply.github.com >
2023-10-22 08:55:44 +02:00
Ettore Di Giacinto
a6fe0a020a
feat(llama.cpp): update ( #1200 )
...
**Description**
This PR updates llama.cpp to
465219b914
Supersedes #1195
2023-10-21 18:44:37 +02:00
Ettore Di Giacinto
bf2ed3d752
fix(Dockerfile): piper phonemize is required during build
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-21 16:40:41 +02:00
Ettore Di Giacinto
d17a92eef3
example(bruno): add image generation
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-21 11:38:23 +02:00
Ettore Di Giacinto
1a7be035d3
fix(Makefile): build all backends if none is specified
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-21 11:34:59 +02:00
Ettore Di Giacinto
004baaa30f
feat(llama.cpp): update
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-21 11:04:03 +02:00
renovate[bot]
ef19268418
chore(deps): update actions/checkout action to v4 ( #1006 )
...
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2023-10-21 08:55:44 +02:00
renovate[bot]
e82470341f
fix(deps): update module google.golang.org/grpc to v1.59.0 ( #1189 )
...
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2023-10-20 17:04:14 +02:00
renovate[bot]
88fa42de75
fix(deps): update github.com/tmc/langchaingo digest to c636b3d ( #1188 )
...
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2023-10-20 17:03:01 +02:00
Ettore Di Giacinto
432513c3ba
ci: add GPU tests ( #1095 )
...
* ci: test GPU
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci: show logs
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Debug
* debug
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* split extra/core images
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* split extra/core images
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* consider runner host dir
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-19 13:50:40 +02:00
renovate[bot]
45370c212b
fix(deps): update github.com/nomic-ai/gpt4all/gpt4all-bindings/golang digest to 9a19c74 ( #1179 )
...
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2023-10-17 18:37:27 +02:00
Jesús Espino
e91f660eb1
feat(metrics): Adding initial support for prometheus metrics ( #1176 )
...
* feat(metrics): Adding initial support for prometheus metrics
* Fixing CI
* run go mod tidy
2023-10-17 18:22:53 +02:00
renovate[bot]
3f3162e57c
fix(deps): update module github.com/gofiber/fiber/v2 to v2.50.0 ( #1177 )
...
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2023-10-16 21:47:44 +02:00
renovate[bot]
208d1fce58
fix(deps): update github.com/tmc/langchaingo digest to a02d4fd ( #1175 )
...
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2023-10-16 21:46:53 +02:00
Ettore Di Giacinto
128694213f
feat: llama.cpp gRPC C++ backend ( #1170 )
...
* wip: llama.cpp c++ gRPC server
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* make it work, attach it to the build process
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* update deps
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: add protobuf dep
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* try fix protobuf on cmake
* cmake: workarounds
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* add packages
* cmake: use fixed version of grpc
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* cmake(grpc): install locally
* install grpc
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* install required deps for grpc on debian bullseye
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* debug
* debug
* Fixups
* no need to install cmake manually
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* ci: fixup macOS
* use brew whenever possible
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* macOS fixups
* debug
* fix container build
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* workaround
* try mac
https://stackoverflow.com/questions/23905661/on-mac-g-clang-fails-to-search-usr-local-include-and-usr-local-lib-by-def
* Disable temp. arm64 docker image builds
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2023-10-16 21:46:29 +02:00