- Singapore
-
11:48
- 8h ahead
Stars
Zero Bubble Pipeline Parallelism
Summary about Video-to-Text datasets. This repository is part of the review paper *Bridging Vision and Language from the Video-to-Text Perspective: A Comprehensive Review*
Robust Speech Recognition via Large-Scale Weak Supervision
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficie…
Stable Diffusion web UI
WebUI extension for ControlNet
Fast and memory-efficient exact attention
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackable.
Accessible large language models via k-bit quantization for PyTorch.
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
4 bits quantization of LLaMA using GPTQ
A Gradio web UI for Large Language Models with support for multiple inference backends.
A collection of libraries to optimise AI model performances
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
A cloud-native Go microservices framework with cli tool for productivity.
A simple, efficient and powerful micro front-end framework. 一款简约、高效、功能强大的微前端框架
Workrave is a program that assists in the recovery and prevention of Repetitive Strain Injury (RSI). The program frequently alerts you to take micro-pauses, rest breaks and restricts you to your da…
Implementation examples of module federation , by the creators of module federation
Create React App boilerplate with React 17, Webpack 5, Tailwind 2, Module Federation, and TypeScript.
A GUI client for Windows, Linux and macOS, support Xray core and sing-box-core and others
Pretrain, finetune ANY AI model of ANY size on multiple GPUs, TPUs with zero code changes.
Profiling and inspecting memory in pytorch
Making large AI models cheaper, faster and more accessible
Scalable PaLM implementation of PyTorch
Optimizing AlphaFold Training and Inference on GPU Clusters