LocalAI/backend/python/transformers
fakezeta e7cbe32601
feat: Openvino runtime for transformer backend and streaming support for Openvino and CUDA (#1892)
* fixes #1775 and #1774

Add BitsAndBytes Quantization and fixes embedding on CUDA devices

* Manage 4bit and 8 bit quantization

Manage different BitsAndBytes options with the quantization: parameter in yaml

* fix compilation errors on non CUDA environment

* OpenVINO draft

First draft of OpenVINO integration in transformer backend

* first working implementation

* Streaming working

* Small fix for regression on CUDA and XPU

* use pip version of optimum[openvino]

* Update backend/python/transformers/transformers_server.py

Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>

---------

Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
2024-03-26 23:31:43 +00:00
..
backend_pb2_grpc.py feat(transformers): add embeddings with Automodel (#1308) 2023-11-20 21:21:17 +01:00
backend_pb2.py feat(tts): add Elevenlabs and OpenAI TTS compatibility layer (#1834) 2024-03-14 23:08:34 +01:00
Makefile feat(conda): share envs with transformer-based backends (#1465) 2023-12-21 08:35:15 +01:00
README.md feat(transformers): add embeddings with Automodel (#1308) 2023-11-20 21:21:17 +01:00
run.sh feat(intel): add diffusers/transformers support (#1746) 2024-03-07 14:37:45 +01:00
test_transformers_server.py tests: add diffusers tests (#1419) 2023-12-11 08:20:34 +01:00
test.sh fix: rename transformers.py to avoid circular import (#1337) 2023-11-26 08:49:43 +01:00
transformers_server.py feat: Openvino runtime for transformer backend and streaming support for Openvino and CUDA (#1892) 2024-03-26 23:31:43 +00:00

Creating a separate environment for the transformers project

make transformers