Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Rename to vLLM #150

Merged
merged 27 commits into from
Jun 17, 2023
Merged

Rename to vLLM #150

merged 27 commits into from
Jun 17, 2023

Conversation

WoosukKwon
Copy link
Collaborator

@WoosukKwon WoosukKwon commented Jun 17, 2023

The current plan is

  1. Merge this PR
  2. Change the repo name to vllm
  3. Move the repo to an organization
  4. Fix readthedocs URL

Copy link
Member

@zhuohan123 zhuohan123 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM! vLLM rocks!

benchmarks/benchmark_latency.py Show resolved Hide resolved
@WoosukKwon
Copy link
Collaborator Author

@zhuohan123 FYI, I've just renamed the attention classes: GPTPagedAttention to PagedAttention, and GPTNeoXPagedAttention to PagedAttentionWithRoPE.

@WoosukKwon WoosukKwon merged commit 0b98ba1 into main Jun 17, 2023
@WoosukKwon WoosukKwon deleted the rename-vllm branch June 17, 2023 10:08
hongxiayang pushed a commit to hongxiayang/vllm that referenced this pull request Feb 13, 2024
sjchoi1 pushed a commit to casys-kaist-internal/vllm that referenced this pull request May 7, 2024
yukavio pushed a commit to yukavio/vllm that referenced this pull request Jul 3, 2024
Summary:
The 2024-03-25 nightly benchmarks failed due to performance regressions.
We find that this is either due to,
- the inherent flakiness in the benchmark experiment itself (experiments
with small work loads), or
  - the inherent flakiness in the metrics. 
Please look at
https://docs.google.com/document/d/1478BMToQIcpSCloiEWqmHoZVrVOZVV-1u4gCyqtjkKE/edit?usp=sharing
for more details.

Updates in this PR:
 - Serving case : Remove the 3000 num prompts at 10 qps experiments. 
- Serving case : Mark the p90, p99 statistics as "Observation" metrics
so they dont trigger failure.
- Engine case (benchmark_throughput.py) : Remove the 16 and 32 prefill
cases.

Test: 
Some local testing

---------

Co-authored-by: Varun Sundar Rabindranath <varun@neuralmagic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants