Releases: bobazooba/xllm
Releases · bobazooba/xllm
Single GPU DeepSpeed, download model without loading to RAM
What's Changed
- DeepSpeed on single GPU by @bobazooba in #8
- Add project_name by @bobazooba in #11
- set transformers version by @bobazooba in #12
- Download model without loading to RAM by @bobazooba in #13
Full Changelog: 0.1.7...0.1.8
Tokenizer hotfix
What's Changed
- Tokenizer trust remote code by @bobazooba in #2
Int8 fusing
Stable release
🎉 Code and documents done!
Useful materials
- Quickstart: basics of
xllm
- Examples: minimal examples of using
xllm
- Guide: here, we go into detail about everything the library can
do - Demo project: here's a minimal step-by-step example of how to use X—LLM and fit it
into your own project - WeatherGPT: this repository features an example of how to utilize the xllm library. Included is a solution for a common type of assessment given to LLM engineers, who typically earn between $120,000 to $140,000 annually
- Shurale: project with the finetuned 7B Mistal model
Init Release
Code done and tested. Time to docs