Skip to content

GPTQModel v1.2.1

Compare
Choose a tag to compare
@Qubitium Qubitium released this 11 Nov 08:11
· 529 commits to main since this release
297b0a9

What's Changed

  • Meta MobileLLM model support added.
  • lm-eval[gptqmodel] integration merged upstream.
  • Intel/IPEX cpu inference merged replacing QBits (deprecated).
  • Auto-fix/patch ChatGLM-3/GLM-4 compat with latest transformers.
  • New .load() and .save() api. 100% model CI coverage and unit testing.

Note that 1.2.1 and 1.2.0 are the same. 1.2.0 release had a bad version name: 1.2.1-dev embedded in code/release.

PR Logs

New Contributors

Full Changelog: v1.1.0...v1.2.1