-
Notifications
You must be signed in to change notification settings - Fork 24
Added data generation & training support for qwen3 vl model #228
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: shanjiaz <zsjwpianpian@gmail.com>
|
📦 Build Artifacts Available |
Summary
Errors per inputErrors in docs/developer/code-of-conduct.md
Redirects per inputRedirects in CONTRIBUTING.md
Redirects in docs/developer/contributing.md
|
661718b to
8125605
Compare
Signed-off-by: shanjiaz <zsjwpianpian@gmail.com>
fynnsu
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good!
Signed-off-by: shanjiaz <zsjwpianpian@gmail.com>
rahul-tuli
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! great job
vl model support:
text_configfrom multimodal model configs in vocab mapping, training, and model initializationkwargsto support vision language parameters likedeepstack_input_embedsin custom worker for Qwen3-VL's multimodal processingget_language_model()method to access the text model in vision-language architecturesUtil changes:
Add suffix pattern matching for weight keys (e.g., "embed_tokens.weight" matches "model.embed_tokens.weight")
Support single-file safetensors models (no index.json) by building virtual weight maps
Run e2e test for a qwen 3 vl model