Ludovic Leroux
939411300a
Bump vLLM version + more options when loading models in vLLM ( #1782 )
...
* Bump vLLM version to 0.3.2
* Add vLLM model loading options
* Remove transformers-exllama
* Fix install exllama
2024-03-01 22:48:53 +01:00
Ettore Di Giacinto
cb7512734d
transformers: correctly load automodels ( #1643 )
...
* backends(transformers): use AutoModel with LLM types
* examples: animagine-xl
* Add codellama examples
2024-01-26 00:13:21 +01:00
Ettore Di Giacinto
e19d7226f8
feat: more embedded models, coqui fixes, add model usage and description ( #1556 )
...
* feat: add model descriptions and usage
* remove default model gallery
* models: add embeddings and tts
* docs: update table
* docs: updates
* images: cleanup pip cache after install
* images: always run apt-get clean
* ux: improve gRPC connection errors
* ux: improve some messages
* fix: fix coqui when no AudioPath is passed by
* embedded: add more models
* Add usage
* Reorder table
2024-01-08 00:37:02 +01:00
Ettore Di Giacinto
949da7792d
deps(conda): use transformers-env with vllm,exllama(2) ( #1554 )
...
* deps(conda): use transformers with vllm
* join vllm, exllama, exllama2, split petals
2024-01-06 13:32:28 +01:00
Ettore Di Giacinto
a1aa6cb7c2
fix(entrypoint): cd to backend dir before start ( #1530 )
...
Certain backends as vall-e-x are not meant to be used as a library, so
we want to start the process in the same folder where the backend and
all the assets are fixes #1394
2024-01-01 22:02:48 +01:00
Ettore Di Giacinto
7641f92cde
feat(diffusers): update, add autopipeline, controlnet ( #1432 )
...
* feat(diffusers): update, add autopipeline, controlenet
* tests with AutoPipeline
* simplify logic
2023-12-13 19:20:22 +01:00
Dave
8b6e601405
Feat: new backend: transformers-musicgen ( #1387 )
...
Transformers-MusicGen
---------
Signed-off-by: Dave <dave@gray101.com>
2023-12-08 10:01:02 +01:00
Ettore Di Giacinto
2b2d6673ff
exllama(v2): fix exllamav1, add exllamav2 ( #1384 )
...
* fix(exllama): fix exllama deps with anaconda
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(exllamav2): add exllamav2 backend
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2023-12-05 08:15:37 +01:00
B4ckslash
2d64d8b444
fix/docs: Python backend dependencies ( #1360 )
...
* Update docs for new requirements.txt path
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com>
* Fix typo (.PONY -> .PHONY) in python backend makefiles
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com>
---------
Signed-off-by: Marcus Köhler <khler.marcus@gmail.com>
2023-11-30 17:46:55 +01:00
ok2sh
20d637e7b7
fix: ExLlama Backend Context Size & Rope Scaling ( #1311 )
...
* fix: context_size not propagated to exllama backend
* fix: exllama rope scaling
2023-11-21 19:26:39 +01:00
Ettore Di Giacinto
ad0e30bca5
refactor: move backends into the backends directory ( #1279 )
...
* refactor: move backends into the backends directory
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* refactor: move main close to implementation for every backend
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2023-11-13 22:40:16 +01:00