Fix code typo in llama-cli#8198
Conversation
|
I don't think it is completely right yet, there are still extra new lines added to the assistant messages randomly. I suspect that at least one issue is that the chat template expects a new line after the |
|
Hmm, that's strange. My result is pretty consistent (with |
|
That case is also fixed for me, but I still see many messages ending in double or triple lines during random chat. |
|
Probably there is some thing more specific for gemma template (or the model itself). In any case, I'll merge this PR now and have a deeper look on gemma later. |
I see what you mean. In this case, we must either patch the template behavior with |
Fix a small typo that breaks chat template support on
llama-cli -cnv