LocalAI/backend/python/exllama2
Ludovic Leroux 939411300a
Bump vLLM version + more options when loading models in vLLM (#1782)
* Bump vLLM version to 0.3.2

* Add vLLM model loading options

* Remove transformers-exllama

* Fix install exllama
2024-03-01 22:48:53 +01:00
..
backend_pb2_grpc.py exllama(v2): fix exllamav1, add exllamav2 (#1384) 2023-12-05 08:15:37 +01:00
backend_pb2.py Bump vLLM version + more options when loading models in vLLM (#1782) 2024-03-01 22:48:53 +01:00
exllama2_backend.py fix: exllama2 backend (#1484) 2023-12-24 08:32:12 +00:00
exllama2.yml exllama(v2): fix exllamav1, add exllamav2 (#1384) 2023-12-05 08:15:37 +01:00
install.sh fix(python): pin exllama2 (#1711) 2024-02-14 21:44:12 +01:00
Makefile deps(conda): use transformers-env with vllm,exllama(2) (#1554) 2024-01-06 13:32:28 +01:00
run.sh deps(conda): use transformers-env with vllm,exllama(2) (#1554) 2024-01-06 13:32:28 +01:00