LocalAI/backend/python/autogptq
Sebastian.W d23e73b118
fix(autogptq): do not use_triton with qwen-vl (#1985)
* Enhance autogptq backend to support VL models

* update dependencies for autogptq

* remove redundant auto-gptq dependency

* Convert base64 to image_url for Qwen-VL model

* implemented model inference for qwen-vl

* remove user prompt from generated answer

* fixed write image error

* fixed use_triton issue when loading Qwen-VL model

---------

Co-authored-by: Binghua Wu <bingwu@estee.com>
2024-04-10 10:36:10 +00:00
..
autogptq.py fix(autogptq): do not use_triton with qwen-vl (#1985) 2024-04-10 10:36:10 +00:00
autogptq.yml Enhance autogptq backend to support VL models (#1860) 2024-03-26 18:48:14 +01:00
backend_pb2_grpc.py refactor: move backends into the backends directory (#1279) 2023-11-13 22:40:16 +01:00
backend_pb2.py feat(tts): add Elevenlabs and OpenAI TTS compatibility layer (#1834) 2024-03-14 23:08:34 +01:00
Makefile deps(conda): use transformers environment with autogptq (#1555) 2024-01-06 15:30:53 +01:00
README.md refactor: move backends into the backends directory (#1279) 2023-11-13 22:40:16 +01:00
run.sh deps(conda): use transformers environment with autogptq (#1555) 2024-01-06 15:30:53 +01:00

Creating a separate environment for the autogptq project

make autogptq