Ludovic Leroux
|
939411300a
|
Bump vLLM version + more options when loading models in vLLM (#1782)
* Bump vLLM version to 0.3.2
* Add vLLM model loading options
* Remove transformers-exllama
* Fix install exllama
|
2024-03-01 22:48:53 +01:00 |
|
Ettore Di Giacinto
|
cb7512734d
|
transformers: correctly load automodels (#1643)
* backends(transformers): use AutoModel with LLM types
* examples: animagine-xl
* Add codellama examples
|
2024-01-26 00:13:21 +01:00 |
|
Ettore Di Giacinto
|
9e653d6abe
|
feat: 🐍 add mamba support (#1589)
feat(mamba): Initial import
This is a first iteration of the mamba backend, loosely based on
mamba-chat(https://github.com/havenhq/mamba-chat).
|
2024-01-19 23:42:50 +01:00 |
|
Ettore Di Giacinto
|
e19d7226f8
|
feat: more embedded models, coqui fixes, add model usage and description (#1556)
* feat: add model descriptions and usage
* remove default model gallery
* models: add embeddings and tts
* docs: update table
* docs: updates
* images: cleanup pip cache after install
* images: always run apt-get clean
* ux: improve gRPC connection errors
* ux: improve some messages
* fix: fix coqui when no AudioPath is passed by
* embedded: add more models
* Add usage
* Reorder table
|
2024-01-08 00:37:02 +01:00 |
|
Ettore Di Giacinto
|
95eb72bfd3
|
feat: add 🐸 coqui (#1489)
* feat: add coqui
* docs: update news
|
2023-12-24 19:38:54 +01:00 |
|