Skip to content

Connecting Nemoclaw to self-hosted vLLM model on same host #893

@jhthompson12

Description

@jhthompson12

Description

using https url

I've successfully nemoclaw onboard my self-hosted vllm instance of gpt-oss-120b using the 3) Other OpenAI-compatible endpoint option with:

OpenAI-compatible base URL (e.g., https://openrouter.ai/api/v1): https://gpt-oss-120b.ai-lab.my-company.com/v1
Other OpenAI-compatible endpoint model []: openai/gpt-oss-120b

But when I try to interact with this model through openclaw tui i get:

run error: LLM request timed out.
connected | error
agent main | session main (openclaw-tui) | inference/openai/gpt-oss-120b | tokens ?/131k

and within the logs (nemoclaw ai-lab-test logs --follow) I see this each time i attempt inference:

[1774449585.801] [sandbox] [WARN ] [openshell_sandbox::proxy] inference endpoint detected but upstream service failed error=upstream unavailable: failed to connect to https://gpt-oss-120b.ai-lab.my-company.com/v1/responses: error sending request for url (https://gpt-oss-120b.ai-lab.my-company.com/v1/responses)

I know the /v1/responses endpoint is working because I've tested it with cURL from my local machine. but If i send the same test cURL request from inside the ai-lab-test sandbox I see this in the nemoclaw logs:

[1774450226.358] [sandbox] [WARN ] [openshell_sandbox::proxy] TLS relay error error=invalid peer certificate: UnknownIssuer host=gpt-oss-120b.ai-lab.my-compant.com port=443

Is the self-signed cert the issue causing all the problems here? How can I add this custom cert to the nemoclaw instance?

using localhost url

I've also tried adding a new provider for the same service, but using the localhost url to the vLLM container which is running on the same host (``)

openshell provider create --name ai-lab-gpt-oss-120b-localhost --type openai --credential "OPENAI_API_KEY=dummy" --config "OPENAI_BASE_URL=http://localhost:8001/v1"

openshell inference set --provider ai-lab-gpt-oss-120b-localhost --model openai/gpt-oss-120b

Error:   × failed to verify inference endpoint for provider 'ai-lab-gpt-oss-120b-localhost' and model 'openai/gpt-oss-120b' at 'http://localhost:8001/v1': failed to
  │ connect to http://localhost:8001/v1/chat/completions: error sending request for url (http://localhost:8001 v1/chat/completions). Next steps: check that the
  │ service is running, confirm the base URL and protocol, and verify credentials, or retry with '--no-verify' if you want to skip verification

but i know this endpoint is also working.

Any help on verifying this is possible or getting this configured would be greatly appreciated!

Reproduction Steps

nemoclaw onboard my model provider successfully. test the model by saying "hello" in the openclaw tui. No inference and receive run error: LLM request timed out. message

Environment

  • OS: Ubuntu 22.04.5 LTS
  • Node: v22.22.1
  • docker: 29.0.4
  • nemoclaw: v0.1.0

Debug Output

Logs

Checklist

  • I confirmed this bug is reproducible
  • I searched existing issues and this is not a duplicate

Metadata

Metadata

Assignees

No one assigned

    Labels

    Getting StartedUse this label to identify setup, installation, or onboarding issues.Provider: OpenAIUse this label to identify issues with the OpenAI provider integration.bugSomething isn't workingenhancement: providerUse this label to identify requests to add a new AI provider to NemoClaw.priority: highImportant issue that should be resolved in the next release

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions