Skip to content

[infer]Add and optimize vllm Continous Batching and Pageattention#4755

Merged
isky-cd merged 10 commits intohpcaitech:feature/vllm-continupus-batchingfrom
isky-cd:vllm_continous_batching
Sep 22, 2023
Merged

[infer]Add and optimize vllm Continous Batching and Pageattention#4755
isky-cd merged 10 commits intohpcaitech:feature/vllm-continupus-batchingfrom
isky-cd:vllm_continous_batching

Conversation

@isky-cd
Copy link
Copy Markdown
Contributor

@isky-cd isky-cd commented Sep 19, 2023

📌 Checklist before creating the PR

  • I have created an issue for this PR for traceability
  • The title follows the standard format: [doc/gemini/tensor/...]: A concise description
  • I have added relevant tags if possible for us to better distinguish different PRs

🚨 Issue number

Link this PR to your issue with words like fixed to automatically close the linked issue upon merge

e.g. fixed #1234, closed #1234, resolved #1234

📝 What does this PR do?

Summarize your work here.
if you have any plots/diagrams/screenshots/tables, please attach them here.

💥 Checklist before requesting a review

  • I have linked my PR to an issue (instruction)
  • My issue clearly describes the problem/feature/proposal, with diagrams/charts/table/code if possible
  • I have performed a self-review of my code
  • I have added thorough tests.
  • I have added docstrings for all the functions/methods I implemented

⭐️ Do you enjoy contributing to Colossal-AI?

  • 🌝 Yes, I do.
  • 🌚 No, I don't.

Tell us more if you don't enjoy contributing to Colossal-AI.

@isky-cd isky-cd marked this pull request as ready for review September 20, 2023 10:30
Comment thread colossalai/inference/tensor_parallel/utils.py Outdated
Comment thread colossalai/inference/tensor_parallel/engine.py Outdated
Comment thread colossalai/inference/tensor_parallel/engine.py Outdated
Comment thread colossalai/inference/tensor_parallel/engine.py Outdated
Comment thread colossalai/inference/tensor_parallel/utils.py Outdated
Comment thread colossalai/inference/tensor_parallel/utils.py Outdated
Comment thread colossalai/inference/tensor_parallel/engine.py Outdated
Comment thread colossalai/inference/tensor_parallel/engine.py
Comment thread colossalai/inference/tensor_parallel/engine.py Outdated
Comment thread examples/inference/test_continous_batching.py
Comment thread examples/inference/bench_llama_continous_batching.py
Comment thread tests/test_infer/test_bloom_infer.py
Comment thread tests/test_infer/test_infer_engine.py
Comment thread tests/test_infer/test_llama_infer.py
@github-actions
Copy link
Copy Markdown
Contributor

The code coverage for the changed files is 33%.

Click me to view the complete report
Name                                             Stmts   Miss  Cover
--------------------------------------------------------------------
colossalai/inference/tensor_parallel/engine.py     203    165    19%
colossalai/inference/tensor_parallel/utils.py       36     26    28%
tests/test_infer/test_bloom_infer.py                43     16    63%
tests/test_infer/test_infer_engine.py               64     35    45%
tests/test_infer/test_llama_infer.py                63     33    48%
--------------------------------------------------------------------
TOTAL                                              409    275    33%

@isky-cd isky-cd merged commit 78158cc into hpcaitech:feature/vllm-continupus-batching Sep 22, 2023
@isky-cd isky-cd changed the title [infer]Add Continous batching and pageattention [infer]Add vllm Continous batching and pageattention Oct 16, 2023
@isky-cd isky-cd changed the title [infer]Add vllm Continous batching and pageattention [infer]Add and optimize vllm Continous batching and pageattention Oct 16, 2023
@isky-cd isky-cd changed the title [infer]Add and optimize vllm Continous batching and pageattention [infer]Add and optimize vllm Continous Batching and Pageattention Oct 16, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants