Skip to content

Conversation

@andreaschari
Copy link
Contributor

No description provided.

@andreaschari andreaschari changed the title remove hard-coded model length Search-O1: remove hard-coded model length Dec 15, 2025
@cmacdonald
Copy link
Collaborator

what problem does this solve? Also doesnt removing the length limit prevent some VLLM optimisation?

@andreaschari
Copy link
Contributor Author

I found that there were cases where search-O1 would crash with:

ValueError: The decoder prompt (length 17848) is longer than the maximum model length of 16384. Make sure that max_model_len is no smaller than the number of text tokens.

So the choice was removing any constraint on the length or adding some truncation in generation_args with truncate_prompt_tokens. I guess not having any constraint by default made more sense and people can define their own backend if they wish to set a different max_model_len and truncate_prompt_tokens?

@ZhlSh17
Copy link
Collaborator

ZhlSh17 commented Dec 15, 2025

Unlike the other two (SearchR1 and R1-Searcher), SearchO1 is model-agnostic so we can't have a one-fits-all max_model_len. Removing it will leave it to the default based on the model's config. Users may manually create a backend to specify their desired max length and/or do truncation.

@cmacdonald
Copy link
Collaborator

Thanks both for the justification!

@cmacdonald cmacdonald merged commit e244707 into terrierteam:main Dec 15, 2025
5 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants