-
Notifications
You must be signed in to change notification settings - Fork 693
[SOT] Add sot warmup (NVIDIA GPU Only) #2929
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
Thanks for your contribution! |
SigureMo
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
SigureMo
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
merge 一下最新的 develop 并跑一下 pre-commit
pre-commit run --files fastdeploy/model_executor/graph_optimization/utils.py fastdeploy/worker/gpu_model_runner.py fastdeploy/worker/gpu_worker.py
SigureMo
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| sot_warmup_guard, in_sot_warmup_mode = create_guard(False) | ||
| profile_run_guard, in_profile_run_mode = create_guard(False) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
是不是保留一个profile_run_guard 就够了,sot 和 cuda graph 都能通过 in_profile_run_mode 判断
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
目前是酱紫的(按时间顺序):
profile_run_guard这个用来判断是否在profile_run阶段,如果是这个阶段,则确定跑动态图sot_warmup_guard这个用来判断是否在 SOT 的 warmup 阶段,这个阶段转静,并标记动态shape- 服务启动之后,如果是启动了SOT转静,则跑静态图,不再标记动态shape
如果只留下 profile_run_guard 的话,就没办法区分是在 SOT warmup 阶段跑的假数据,还是服务启动之后跑的真数据(服务启动后不再标记动态shape,只在 warmup 阶段标记)
|
SOT 计划支持多硬件吗,gcu 那些model runner 需要适配吗 |
@gongshaotian 其他硬件也会支持 warmup,下一个PR统一修改吧 |
gongshaotian
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
* [MTP Fix] Fix code and register cpp operators (PaddlePaddle#2965) * fix rl config local rank (PaddlePaddle#2957) * [FIX]fix rejection sampling when topp=0 using _SAMPLING_EPS (PaddlePaddle#2967) * fix rejection sampling when topp=0 * fix * [SOT] Add sot warmup (NVIDIA GPU Only) (PaddlePaddle#2929) * add sot warmup * fix code style * change batch_size list * add param to config * rm free_list settings && set sot_warmup_sizes * finish debug with dynamic dims by type annotations * add profile_run guard * rm sth useless * support chunk_prefill in fa3 * 【Infer】Improve the performance block_wise_fp8 of triton_moe_backend (PaddlePaddle#2942) * Update README.md * Update README.md * delete max-len (PaddlePaddle#2959) * [CI] add codestyle_check action (PaddlePaddle#2972) * [CI] add codestyle_check action * [CI] Integrate codestyle check via pre-commit in GitHub Actions * fix mtp bug in pd-split mode (PaddlePaddle#2970) * [BugFix] Add prefill restrictions for chunked_prefill+VL (PaddlePaddle#2983) * Fix performance degradation bug of custom_all_reduce (PaddlePaddle#2981) * FA3 fix bug (PaddlePaddle#2987) * polish code for prefill restrictions (PaddlePaddle#2991) * [Feature] Support block scheduler v1 for FD (PaddlePaddle#2928) * Support FD block scheduler v1 * Support FD block scheduler v1 * Support FD block scheduler v1 * Fix according to copilot review * Fix according to review * Remove is_dummy * Fix bug when real_bsz=1 * Fix infer first token cost time --------- Co-authored-by: Jiang-Jia-Jun <163579578+Jiang-Jia-Jun@users.noreply.github.com> * update (PaddlePaddle#2978) * [Code Simplification] fix init_distributed_environment() (PaddlePaddle#2982) * support c4 attn && fix cache * fix chunk_prefill * [benchmark] add quantization for benchmark yaml (PaddlePaddle#2995) * [Fix] fix mm ep empty run (PaddlePaddle#2999) * add ci reuse action (PaddlePaddle#2968) * add ci reuse action * fix code formatting * update * [Feature] multi-source download (PaddlePaddle#2986) * multi-source download * multi-source download * huggingface download revision * requirement * style * add revision arg * test * pre-commit * [LLM] update function name (PaddlePaddle#2985) * [LLM] update function name * [BugFix] fix multinode deployment (PaddlePaddle#2977) * Update benchmark tools (PaddlePaddle#3004) * update benchmark tools * update benchmark tools * update flake8 version to support pre-commit in python3.12 (PaddlePaddle#3000) * update flake8 version to support pre-commit in python3.12 * polish code * [Feature] multi source download (PaddlePaddle#3005) * multi-source download * multi-source download * huggingface download revision * requirement * style * add revision arg * test * pre-commit * Change default download * change requirements.txt * modify English Documentation * documentation * [GCU] Update to develop (PaddlePaddle#2988) * [Model] Provide clearer error for missing KV cache quantization scales (PaddlePaddle#3007) * [Feature] Support_eplb (PaddlePaddle#2997) * [Feature] support_eplb * [Feature] support_eplb * [Fix] fix mm ep * Update setup.py * [feat] add disable_chat_template in chat api as a substitute for previous raw_request (PaddlePaddle#3023) * [feat] add disable_chat_template in chat api as a substitute for previous raw_request * [fix] pre-commit code check --------- Co-authored-by: GoldPancake <56388518+Deleter-D@users.noreply.github.com> Co-authored-by: gaoziyuan <88373061+gzy19990617@users.noreply.github.com> Co-authored-by: Sunny-bot1 <68891411+Sunny-bot1@users.noreply.github.com> Co-authored-by: Ryan <zihaohuang@aliyun.com> Co-authored-by: lizhenyun01 <1500424927@qq.com> Co-authored-by: chen <103103266+ckl117@users.noreply.github.com> Co-authored-by: Jiang-Jia-Jun <163579578+Jiang-Jia-Jun@users.noreply.github.com> Co-authored-by: lizexu123 <39205361+lizexu123@users.noreply.github.com> Co-authored-by: YuBaoku <49938469+EmmonsCurse@users.noreply.github.com> Co-authored-by: freeliuzc <lzc842650834@gmail.com> Co-authored-by: Zero Rains <linjunlu@zerorains.top> Co-authored-by: zhink <33270771+zhink@users.noreply.github.com> Co-authored-by: chenjian <1435317881@qq.com> Co-authored-by: bukejiyu <52310069+bukejiyu@users.noreply.github.com> Co-authored-by: xiegegege <46314656+xiegegege@users.noreply.github.com> Co-authored-by: xiaoxiaohehe001 <49090790+xiaoxiaohehe001@users.noreply.github.com> Co-authored-by: YUNSHEN XIE <1084314248@qq.com> Co-authored-by: Yzc216 <101054010+Yzc216@users.noreply.github.com> Co-authored-by: ltd0924 <32387785+ltd0924@users.noreply.github.com> Co-authored-by: Zhang Yulong <35552275+ZhangYulongg@users.noreply.github.com> Co-authored-by: EnflameGCU <118410644+EnflameGCU@users.noreply.github.com> Co-authored-by: littledgg <61149469+littledgg@users.noreply.github.com> Co-authored-by: 李泳桦 <39643373+liyonghua0910@users.noreply.github.com>
给SOT添加 Warm Up 操作
cc @SigureMo