local-ai: 2.17.1 -> 2.18.1
This commit is contained in:
parent
5065d861d0
commit
e1f3b1a53c
@ -97,21 +97,17 @@ let
|
||||
llama-cpp-rpc = (llama-cpp-grpc.overrideAttrs (prev: {
|
||||
name = "llama-cpp-rpc";
|
||||
cmakeFlags = prev.cmakeFlags ++ [
|
||||
(lib.cmakeBool "LLAMA_AVX" false)
|
||||
(lib.cmakeBool "LLAMA_AVX2" false)
|
||||
(lib.cmakeBool "LLAMA_AVX512" false)
|
||||
(lib.cmakeBool "LLAMA_FMA" false)
|
||||
(lib.cmakeBool "LLAMA_F16C" false)
|
||||
(lib.cmakeBool "LLAMA_RPC" true)
|
||||
(lib.cmakeBool "GGML_AVX" false)
|
||||
(lib.cmakeBool "GGML_AVX2" false)
|
||||
(lib.cmakeBool "GGML_AVX512" false)
|
||||
(lib.cmakeBool "GGML_FMA" false)
|
||||
(lib.cmakeBool "GGML_F16C" false)
|
||||
];
|
||||
postPatch = prev.postPatch + ''
|
||||
sed -i examples/rpc/CMakeLists.txt \
|
||||
-e '$a\install(TARGETS rpc-server RUNTIME)'
|
||||
'';
|
||||
})).override {
|
||||
cudaSupport = false;
|
||||
openclSupport = false;
|
||||
blasSupport = false;
|
||||
rpcSupport = true;
|
||||
};
|
||||
|
||||
llama-cpp-grpc = (llama-cpp.overrideAttrs (final: prev: {
|
||||
@ -119,8 +115,8 @@ let
|
||||
src = fetchFromGitHub {
|
||||
owner = "ggerganov";
|
||||
repo = "llama.cpp";
|
||||
rev = "37bef8943312d91183ff06d8f1214082a17344a5";
|
||||
hash = "sha256-E3kCMDK5TXozBsprp4D581WHTVP9aljxB1KZUKug1pM=";
|
||||
rev = "cb5fad4c6c2cbef92e9b8b63449e1cb7664e4846";
|
||||
hash = "sha256-cIJuDC+MFLd5hkA1kUxuaw2dZagHqn5fi5Q2XKvDEII=";
|
||||
fetchSubmodules = true;
|
||||
};
|
||||
postPatch = prev.postPatch + ''
|
||||
@ -137,14 +133,12 @@ let
|
||||
'';
|
||||
cmakeFlags = prev.cmakeFlags ++ [
|
||||
(lib.cmakeBool "BUILD_SHARED_LIBS" false)
|
||||
(lib.cmakeBool "LLAMA_AVX" enable_avx)
|
||||
(lib.cmakeBool "LLAMA_AVX2" enable_avx2)
|
||||
(lib.cmakeBool "LLAMA_AVX512" enable_avx512)
|
||||
(lib.cmakeBool "LLAMA_FMA" enable_fma)
|
||||
(lib.cmakeBool "LLAMA_F16C" enable_f16c)
|
||||
(lib.cmakeBool "GGML_AVX" enable_avx)
|
||||
(lib.cmakeBool "GGML_AVX2" enable_avx2)
|
||||
(lib.cmakeBool "GGML_AVX512" enable_avx512)
|
||||
(lib.cmakeBool "GGML_FMA" enable_fma)
|
||||
(lib.cmakeBool "GGML_F16C" enable_f16c)
|
||||
];
|
||||
postInstall = null;
|
||||
|
||||
buildInputs = prev.buildInputs ++ [
|
||||
protobuf # provides also abseil_cpp as propagated build input
|
||||
grpc
|
||||
@ -394,18 +388,18 @@ let
|
||||
stdenv;
|
||||
|
||||
pname = "local-ai";
|
||||
version = "2.17.1";
|
||||
version = "2.18.1";
|
||||
src = fetchFromGitHub {
|
||||
owner = "go-skynet";
|
||||
repo = "LocalAI";
|
||||
rev = "v${version}";
|
||||
hash = "sha256-G9My4t3vJ1sWyD+vxUgON4ezXURVAAgu1nAtTjd3ZR8=";
|
||||
hash = "sha256-hRrbGUUawQV4fqxAn3eFBvn4/lZ+NrKhxnGHqpljrec=";
|
||||
};
|
||||
|
||||
self = buildGoModule.override { stdenv = effectiveStdenv; } {
|
||||
inherit pname version src;
|
||||
|
||||
vendorHash = "sha256-Hu7aJFi40CKNWAxYOR47VBZI1A/9SlBIVQVcB8iqcxA=";
|
||||
vendorHash = "sha256-uvko1PQWW5P+6cgmwVKocKBm5GndszqCsSbxlXANqJs=";
|
||||
|
||||
env.NIX_CFLAGS_COMPILE = lib.optionalString with_stablediffusion " -isystem ${opencv}/include/opencv4";
|
||||
|
||||
@ -441,7 +435,7 @@ let
|
||||
cp ${llama-cpp-rpc}/bin/grpc-server backend-assets/grpc/llama-cpp-grpc
|
||||
|
||||
mkdir -p backend-assets/util
|
||||
cp ${llama-cpp-rpc}/bin/rpc-server backend-assets/util/llama-cpp-rpc-server
|
||||
cp ${llama-cpp-rpc}/bin/llama-rpc-server backend-assets/util/llama-cpp-rpc-server
|
||||
'';
|
||||
|
||||
buildInputs = [ ]
|
||||
|
Loading…
Reference in New Issue
Block a user