Some users may use OpenAI or HF inference that we currently support as hubs, but other users may be working on a custom LLM and have the model running on a different hub. We could extend the API implementation to be generic such that the user can specify the parameters to connect to any API and how to parse the results.
For example, the user may define the URL and parameters of their API, and a function to parse the results. In this way, any API system could be supported by LangTest.
Two tools to have in mind: vLLM and HugginFace Text Generation Inference
Some users may use OpenAI or HF inference that we currently support as hubs, but other users may be working on a custom LLM and have the model running on a different hub. We could extend the API implementation to be generic such that the user can specify the parameters to connect to any API and how to parse the results.
For example, the user may define the URL and parameters of their API, and a function to parse the results. In this way, any API system could be supported by LangTest.
Two tools to have in mind: vLLM and HugginFace Text Generation Inference