LocalAI/backend
fakezeta 66b002458d
Transformer Backend: Implementing use_tokenizer_template and stop_prompts options (#2090)
* fix regression #1971

fixes regression #1971 introduced by intel_extension_for_transformers==1.4

* UseTokenizerTemplate and StopPrompt

Implementation of use_tokenizer_template and stopwords options
2024-04-21 16:20:25 +00:00
..
cpp feat(grpc): return consumed token count and update response accordingly (#2035) 2024-04-15 19:47:11 +02:00
go Revert #1963 (#2056) 2024-04-17 23:33:49 +02:00
python Transformer Backend: Implementing use_tokenizer_template and stop_prompts options (#2090) 2024-04-21 16:20:25 +00:00
backend.proto Add tensor_parallel_size setting to vllm setting items (#2085) 2024-04-20 14:37:02 +00:00