v1 api currently uses a harcoded alpaca prompt template with support for specifying adapters to change it: #466
This method is not supported in for example Open-WebUI. It would be great if koboldcpp could automatically apply the chat template specified in tokenizer.chat_template of the gguf model.
v1 api currently uses a harcoded alpaca prompt template with support for specifying adapters to change it: #466
This method is not supported in for example Open-WebUI. It would be great if koboldcpp could automatically apply the chat template specified in tokenizer.chat_template of the gguf model.