LocalAI/backend
fakezeta 8e36fe9b6f
Transformers Backend: max_tokens adherence to OpenAI API (#2108)
max token adherence to OpenAI API

improve adherence to OpenAI API when max tokens is omitted or equal to 0 in the request
2024-04-23 18:42:17 +02:00
..
cpp feat(grpc): return consumed token count and update response accordingly (#2035) 2024-04-15 19:47:11 +02:00
go Revert #1963 (#2056) 2024-04-17 23:33:49 +02:00
python Transformers Backend: max_tokens adherence to OpenAI API (#2108) 2024-04-23 18:42:17 +02:00
backend.proto Add tensor_parallel_size setting to vllm setting items (#2085) 2024-04-20 14:37:02 +00:00