mirror of https://github.com/mudler/LocalAI.git
fix: add some more missing dependencies to python backends
Signed-off-by: Chris Jowett <421501+cryptk@users.noreply.github.com>
This commit is contained in:
parent
7627e69997
commit
444dc35e4f
|
@ -2,4 +2,5 @@ accelerate
|
|||
auto-gptq==0.7.1
|
||||
grpcio==1.63.0
|
||||
protobuf
|
||||
torch
|
||||
torch
|
||||
certifi
|
|
@ -1,4 +1,5 @@
|
|||
accelerate
|
||||
bark==0.1.5
|
||||
grpcio==1.63.0
|
||||
protobuf
|
||||
protobuf
|
||||
certifi
|
|
@ -1,4 +1,5 @@
|
|||
accelerate
|
||||
TTS==0.22.0
|
||||
grpcio==1.63.0
|
||||
protobuf
|
||||
protobuf
|
||||
certifi
|
|
@ -6,4 +6,5 @@ opencv-python
|
|||
pillow
|
||||
protobuf
|
||||
torch
|
||||
transformers
|
||||
transformers
|
||||
certifi
|
|
@ -1,4 +1,5 @@
|
|||
grpcio==1.63.0
|
||||
protobuf
|
||||
torch
|
||||
transformers
|
||||
transformers
|
||||
certifi
|
|
@ -1,3 +1,4 @@
|
|||
accelerate
|
||||
grpcio==1.63.0
|
||||
protobuf
|
||||
certifi
|
|
@ -2,4 +2,5 @@ accelerate
|
|||
causal-conv1d==1.2.0.post2
|
||||
mamba-ssm==1.2.0.post1
|
||||
grpcio==1.63.0
|
||||
torch==2.1.2
|
||||
torch==2.1.2
|
||||
certifi
|
|
@ -2,4 +2,5 @@ accelerate
|
|||
grpcio==1.63.0
|
||||
protobuf
|
||||
torch
|
||||
git+https://github.com/huggingface/parler-tts.git@10016fb0300c0dc31a0fb70e26f3affee7b62f16
|
||||
git+https://github.com/huggingface/parler-tts.git@10016fb0300c0dc31a0fb70e26f3affee7b62f16
|
||||
certifi
|
|
@ -25,6 +25,7 @@ class TestBackendServicer(unittest.TestCase):
|
|||
"""
|
||||
This method tears down the gRPC service by terminating the server
|
||||
"""
|
||||
print("stopping service")
|
||||
self.service.terminate()
|
||||
self.service.wait()
|
||||
|
||||
|
@ -53,6 +54,8 @@ class TestBackendServicer(unittest.TestCase):
|
|||
with grpc.insecure_channel("localhost:50051") as channel:
|
||||
stub = backend_pb2_grpc.BackendStub(channel)
|
||||
response = stub.LoadModel(backend_pb2.ModelOptions(Model="parler-tts/parler_tts_mini_v0.1"))
|
||||
print("response:")
|
||||
print(response)
|
||||
self.assertTrue(response.success)
|
||||
self.assertEqual(response.message, "Model loaded successfully")
|
||||
except Exception as err:
|
||||
|
|
|
@ -1 +1,2 @@
|
|||
git+https://github.com/bigscience-workshop/petals
|
||||
git+https://github.com/bigscience-workshop/petals
|
||||
certifi
|
|
@ -1,4 +1,5 @@
|
|||
accelerate
|
||||
rerankers[transformers]
|
||||
grpcio==1.63.0
|
||||
protobuf
|
||||
protobuf
|
||||
certifi
|
|
@ -1,4 +1,5 @@
|
|||
accelerate
|
||||
sentence-transformers==2.5.1
|
||||
grpcio==1.63.0
|
||||
protobuf
|
||||
protobuf
|
||||
certifi
|
|
@ -3,4 +3,5 @@ transformers>=4.38.2
|
|||
grpcio==1.63.0
|
||||
protobuf
|
||||
torch
|
||||
scipy==1.13.0
|
||||
scipy==1.13.0
|
||||
certifi
|
|
@ -2,4 +2,5 @@ accelerate
|
|||
transformers>=4.38.2
|
||||
grpcio==1.63.0
|
||||
protobuf
|
||||
torch
|
||||
torch
|
||||
certifi
|
|
@ -1,3 +1,4 @@
|
|||
accelerate
|
||||
grpcio==1.63.0
|
||||
protobuf
|
||||
protobuf
|
||||
certifi
|
|
@ -2,3 +2,4 @@ accelerate
|
|||
vllm
|
||||
grpcio==1.63.0
|
||||
protobuf
|
||||
certifi
|
Loading…
Reference in New Issue