Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bugfix] Clean up some cruft in mamba.py #9343

Merged
merged 4 commits into from
Oct 15, 2024

Conversation

tlrmchlsmth
Copy link
Collaborator

Delete some unnecessary/useless code I noticed while reviewing #7478

Copy link

👋 Hi! Thank you for contributing to the vLLM project.
Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your fastcheck build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping simon-mo or khluu to add you in our Buildkite org.

Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.

To run CI, PR reviewers can do one of these:

  • Add ready label to the PR
  • Enable auto-merge.

🚀

@DarkLight1337 DarkLight1337 enabled auto-merge (squash) October 14, 2024 14:59
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Oct 14, 2024
vllm/model_executor/models/mamba.py Outdated Show resolved Hide resolved
vllm/model_executor/models/mamba.py Outdated Show resolved Hide resolved
@tlrmchlsmth
Copy link
Collaborator Author

Ok, this was extremely boneheaded -- just removed an entire feed forwards section that I had copy-pasted from Jamba, that isn't present in the Mamba model.

Copy link
Collaborator

@mgoin mgoin left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Now that is clean 😎

@@ -332,40 +282,20 @@ def forward(
current_ssm_state = ssm_state[i]
current_conv_state = conv_state[i]

hidden_states, residual = layer(
hidden_states = layer(
positions=positions,
hidden_states=hidden_states,
attn_metadata=attn_metadata,
residual=residual,
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Residual is still passed in here. I personally think it is fine to keep the previous structure of passing residual into rmsnorm, but up to you

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I saw modeling_mamba.py do it this way and like that it's super clean, but you're right -- going to revert to the previous way so that we fuse. Sidenote: this is the kind of optimization we should be relying on torch.compile to do for us IMO

@tlrmchlsmth tlrmchlsmth enabled auto-merge (squash) October 14, 2024 19:10
@tlrmchlsmth tlrmchlsmth merged commit 169b530 into vllm-project:main Oct 15, 2024
56 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ready ONLY add when PR is ready to merge/full CI is needed
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants