Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Conversation templates for Qwen and StableLM #285

Merged
merged 3 commits into from
Feb 3, 2024

Conversation

DavidGOrtega
Copy link
Contributor

@DavidGOrtega DavidGOrtega commented Feb 3, 2024

  • Adds qwen
  • Adds stablels-3b
  • Adds stablelm-2
  • Fixes chatml not able to be extended with config json

Copy link
Contributor

@CharlieFRuan CharlieFRuan left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@CharlieFRuan CharlieFRuan merged commit 3319d1c into mlc-ai:main Feb 3, 2024
CharlieFRuan added a commit that referenced this pull request Feb 15, 2024
The new version includes two main changes:
1. We now support models compiled with `PagedKVCache` (only
Llama-variants for now)
- This is breaking in the sense that, in order to use the update Llama
WASMs, an update to the WebLLM npm is required
- WASMs updated here:
mlc-ai/binary-mlc-llm-libs#90
  - For more see #293
2. We now support `GenerationConfig`, allowing configuring each
generation (e.g. repetition penalty, temperature, etc.)
- All WASMs needed to be recompiled since we included new function
`ApplyPresenceAndRequencyPenalty()` in tvmjs
  - For more see #298

Other changes include: 
- #285
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants