User Story
As a user
I want that the most common used models are cached and hold in memory with Ollama
So that I can benefit from faster model reply circles.
Description
the most common used model should stay in the memory cache of Ollama so that fast replies can be used. normally Ollama has to load the model in memory and the model stays for a time in memory. most common used models should stay in memory for a longer time so that fast replies can be enabled. also improve model caching effectiveness.
Acceptance Criteria
User Story
As a user
I want that the most common used models are cached and hold in memory with Ollama
So that I can benefit from faster model reply circles.
Description
the most common used model should stay in the memory cache of Ollama so that fast replies can be used. normally Ollama has to load the model in memory and the model stays for a time in memory. most common used models should stay in memory for a longer time so that fast replies can be enabled. also improve model caching effectiveness.
Acceptance Criteria