Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Q4_1 quantization compiling to vmfb megacommit #2

Merged

Conversation

Max191
Copy link
Contributor

@Max191 Max191 commented Feb 22, 2024

No description provided.

@Max191
Copy link
Contributor Author

Max191 commented Feb 22, 2024

I should split this into multiple PRs, but my commits accidentally got too jumbled up so I just squashed everything :P

I'll split it up tomorrow, but I'll leave this PR here in case anyone wants to see it or cherry pick it

@Max191
Copy link
Contributor Author

Max191 commented Feb 22, 2024

nod-ai/SHARK-ModelDev#473 is also needed for this PR

Copy link
Owner

@stellaraccident stellaraccident left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It's fine. Prototype code we'll clean it in a future revision.

@stellaraccident stellaraccident marked this pull request as ready for review February 23, 2024 00:14
@stellaraccident stellaraccident merged commit e2189c7 into stellaraccident:main Feb 23, 2024
dmahurin pushed a commit to persimmonsai/mlir-llm-runner that referenced this pull request Jun 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants