chatGPT has an API that allows you to easily integrate it into your own applications. alpaca.cpp currently runs in the terminal, but is there an existing way to call it from our own programs?
The only way I've seen this to be possible currently is via a terminal command similar to below (#103):
.\Release\chat.exe -p "write a python script that prints the current time" --interactive
The problem with this solution is that it requires restarting/reloading the entire model for every prompt you ask it. This becomes increasingly time consuming and compute intensive the larger the weight you use as well.
Is there a better way to do this currently? If not, how can it be added?
chatGPT has an API that allows you to easily integrate it into your own applications. alpaca.cpp currently runs in the terminal, but is there an existing way to call it from our own programs?
The only way I've seen this to be possible currently is via a terminal command similar to below (#103):
.\Release\chat.exe -p "write a python script that prints the current time" --interactiveThe problem with this solution is that it requires restarting/reloading the entire model for every prompt you ask it. This becomes increasingly time consuming and compute intensive the larger the weight you use as well.
Is there a better way to do this currently? If not, how can it be added?