mirror of
https://github.com/mudler/LocalAI.git
synced 2026-01-22 21:31:20 -05:00
23 lines
1.0 KiB
Bash
23 lines
1.0 KiB
Bash
#!/bin/bash
|
|
|
|
cp -r CMakeLists.txt llama.cpp/examples/grpc-server/
|
|
cp -r grpc-server.cpp llama.cpp/examples/grpc-server/
|
|
cp -rfv json.hpp llama.cpp/examples/grpc-server/
|
|
cp -rfv utils.hpp llama.cpp/examples/grpc-server/
|
|
|
|
if grep -q "grpc-server" llama.cpp/examples/CMakeLists.txt; then
|
|
echo "grpc-server already added"
|
|
else
|
|
echo "add_subdirectory(grpc-server)" >> llama.cpp/examples/CMakeLists.txt
|
|
fi
|
|
|
|
## XXX: In some versions of CMake clip wasn't being built before llama.
|
|
## This is an hack for now, but it should be fixed in the future.
|
|
cp -rfv llama.cpp/examples/llava/clip.h llama.cpp/examples/grpc-server/clip.h
|
|
cp -rfv llama.cpp/examples/llava/llava.cpp llama.cpp/examples/grpc-server/llava.cpp
|
|
echo '#include "llama.h"' > llama.cpp/examples/grpc-server/llava.h
|
|
cat llama.cpp/examples/llava/llava.h >> llama.cpp/examples/grpc-server/llava.h
|
|
cp -rfv llama.cpp/examples/llava/clip.cpp llama.cpp/examples/grpc-server/clip.cpp
|
|
|
|
# https://github.com/ggerganov/llama.cpp/issues/8665
|
|
cp -rfv CMakeLists.txt.rpc-8662 llama.cpp/examples/rpc/CMakeLists.txt |