Skip to content

Hannibal046/Awesome-LLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Awesome-LLM Awesome

🔥 Large Language Models(LLM) have taken the NLP community AI community the Whole World by storm. Here is a curated list of papers about large language models, especially relating to ChatGPT. It also contains frameworks for LLM training, tools to deploy LLM, courses and tutorials about LLM and all publicly available LLM checkpoints and APIs.

Trending LLM Projects

  • TinyZero - Clean, minimal, accessible reproduction of DeepSeek R1-Zero
  • open-r1 - Fully open reproduction of DeepSeek-R1
  • DeepSeek-R1 - First-generation reasoning models from DeepSeek.
  • Qwen2.5-Max - Exploring the Intelligence of Large-scale MoE Model.
  • OpenAI o3-mini - Pushing the frontier of cost-effective reasoning.
  • DeepSeek-V3 - First open-sourced GPT-4o level model.

Table of Content

Milestone Papers

milestone papers
Date keywords Institute Paper
2017-06 Transformers Google Attention Is All You Need
2018-06 GPT 1.0 OpenAI Improving Language Understanding by Generative Pre-Training
2018-10 BERT Google BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
2019-02 GPT 2.0 OpenAI Language Models are Unsupervised Multitask Learners
2019-09 Megatron-LM NVIDIA Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
2019-10 T5 Google Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
2019-10 ZeRO Microsoft ZeRO: Memory Optimizations Toward Training Trillion Parameter Models
2020-01 Scaling Law OpenAI Scaling Laws for Neural Language Models
2020-05 GPT 3.0 OpenAI Language models are few-shot learners
2021-01 Switch Transformers Google Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
2021-08 Codex OpenAI Evaluating Large Language Models Trained on Code
2021-08 Foundation Models Stanford On the Opportunities and Risks of Foundation Models
2021-09 FLAN Google Finetuned Language Models are Zero-Shot Learners
2021-10 T0 HuggingFace et al. Multitask Prompted Training Enables Zero-Shot Task Generalization
2021-12 GLaM Google GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
2021-12 WebGPT OpenAI WebGPT: Browser-assisted question-answering with human feedback
2021-12 Retro DeepMind Improving language models by retrieving from trillions of tokens
2021-12 Gopher DeepMind Scaling Language Models: Methods, Analysis & Insights from Training Gopher
2022-01 COT Google Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
2022-01 LaMDA Google LaMDA: Language Models for Dialog Applications
2022-01 Minerva Google Solving Quantitative Reasoning Problems with Language Models
2022-01 Megatron-Turing NLG Microsoft&NVIDIA Using Deep and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model
2022-03 InstructGPT OpenAI Training language models to follow instructions with human feedback
2022-04 PaLM Google PaLM: Scaling Language Modeling with Pathways
2022-04 Chinchilla DeepMind An empirical analysis of compute-optimal large language model training
2022-05 OPT Meta OPT: Open Pre-trained Transformer Language Models
2022-05 UL2 Google Unifying Language Learning Paradigms
2022-06 Emergent Abilities Google Emergent Abilities of Large Language Models
2022-06 BIG-bench Google Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
2022-06 METALM Microsoft Language Models are General-Purpose Interfaces
2022-09 Sparrow DeepMind Improving alignment of dialogue agents via targeted human judgements
2022-10 Flan-T5/PaLM Google Scaling Instruction-Finetuned Language Models
2022-10 GLM-130B Tsinghua GLM-130B: An Open Bilingual Pre-trained Model
2022-11 HELM Stanford Holistic Evaluation of Language Models
2022-11 BLOOM BigScience BLOOM: A 176B-Parameter Open-Access Multilingual Language Model
2022-11 Galactica Meta Galactica: A Large Language Model for Science
2022-12 OPT-IML Meta OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization
2023-01 Flan 2022 Collection Google The Flan Collection: Designing Data and Methods for Effective Instruction Tuning
2023-02 LLaMA Meta LLaMA: Open and Efficient Foundation Language Models
2023-02 Kosmos-1 Microsoft Language Is Not All You Need: Aligning Perception with Language Models
2023-03 LRU DeepMind Resurrecting Recurrent Neural Networks for Long Sequences
2023-03 PaLM-E Google PaLM-E: An Embodied Multimodal Language Model
2023-03 GPT 4 OpenAI GPT-4 Technical Report
2023-04 LLaVA UW–Madison&Microsoft Visual Instruction Tuning
2023-04 Pythia EleutherAI et al. Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
2023-05 Dromedary CMU et al. Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision
2023-05 PaLM 2 Google PaLM 2 Technical Report
2023-05 RWKV Bo Peng RWKV: Reinventing RNNs for the Transformer Era
2023-05 DPO Stanford Direct Preference Optimization: Your Language Model is Secretly a Reward Model
2023-05 ToT Google&Princeton Tree of Thoughts: Deliberate Problem Solving with Large Language Models
2023-07 LLaMA2 Meta Llama 2: Open Foundation and Fine-Tuned Chat Models
2023-10 Mistral 7B Mistral Mistral 7B
2023-12 Mamba CMU&Princeton Mamba: Linear-Time Sequence Modeling with Selective State Spaces
2024-01 DeepSeek-v2 DeepSeek DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model
2024-02 OLMo Ai2 OLMo: Accelerating the Science of Language Models
2024-05 Mamba2 CMU&Princeton Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality
2024-05 Llama3 Meta The Llama 3 Herd of Models
2024-06 FineWeb HuggingFace The FineWeb Datasets: Decanting the Web for the Finest Text Data at Scale
2024-09 OLMoE Ai2 OLMoE: Open Mixture-of-Experts Language Models
2024-12 Qwen2.5 Alibaba Qwen2.5 Technical Report
2024-12 DeepSeek-V3 DeepSeek DeepSeek-V3 Technical Report
2025-01 DeepSeek-R1 DeepSeek DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning

Other Papers

Note

If you're interested in the field of LLM, you may find the above list of milestone papers helpful to explore its history and state-of-the-art. However, each direction of LLM offers a unique set of insights and contributions, which are essential to understanding the field as a whole. For a detailed list of papers in various subfields, please refer to the following link:

other papers
  • Awesome-LLM-hallucination - LLM hallucination paper list.

  • awesome-hallucination-detection - List of papers on hallucination detection in LLMs.

  • LLMsPracticalGuide - A curated list of practical guide resources of LLMs

  • Awesome ChatGPT Prompts - A collection of prompt examples to be used with the ChatGPT model.

  • awesome-chatgpt-prompts-zh - A Chinese collection of prompt examples to be used with the ChatGPT model.

  • Awesome ChatGPT - Curated list of resources for ChatGPT and GPT-3 from OpenAI.

  • Chain-of-Thoughts Papers - A trend starts from "Chain of Thought Prompting Elicits Reasoning in Large Language Models.

  • Awesome Deliberative Prompting - How to ask LLMs to produce reliable reasoning and make reason-responsive decisions.

  • Instruction-Tuning-Papers - A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).

  • LLM Reading List - A paper & resource list of large language models.

  • Reasoning using Language Models - Collection of papers and resources on Reasoning using Language Models.

  • Chain-of-Thought Hub - Measuring LLMs' Reasoning Performance

  • Awesome GPT - A curated list of awesome projects and resources related to GPT, ChatGPT, OpenAI, LLM, and more.

  • Awesome GPT-3 - a collection of demos and articles about the OpenAI GPT-3 API.

  • Awesome LLM Human Preference Datasets - a collection of human preference datasets for LLM instruction tuning, RLHF and evaluation.

  • RWKV-howto - possibly useful materials and tutorial for learning RWKV.

  • ModelEditingPapers - A paper & resource list on model editing for large language models.

  • Awesome LLM Security - A curation of awesome tools, documents and projects about LLM Security.

  • Awesome-Align-LLM-Human - A collection of papers and resources about aligning large language models (LLMs) with human.

  • Awesome-Code-LLM - An awesome and curated list of best code-LLM for research.

  • Awesome-LLM-Compression - Awesome LLM compression research papers and tools.

  • Awesome-LLM-Systems - Awesome LLM systems research papers.

  • awesome-llm-webapps - A collection of open source, actively maintained web apps for LLM applications.

  • awesome-japanese-llm - 日本語LLMまとめ - Overview of Japanese LLMs.

  • Awesome-LLM-Healthcare - The paper list of the review on LLMs in medicine.

  • Awesome-LLM-Inference - A curated list of Awesome LLM Inference Paper with codes.

  • Awesome-LLM-3D - A curated list of Multi-modal Large Language Model in 3D world, including 3D understanding, reasoning, generation, and embodied agents.

  • LLMDatahub - a curated collection of datasets specifically designed for chatbot training, including links, size, language, usage, and a brief description of each dataset

  • Awesome-Chinese-LLM - 整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。

  • LLM4Opt - Applying Large language models (LLMs) for diverse optimization tasks (Opt) is an emerging research area. This is a collection of references and papers of LLM4Opt.

  • awesome-language-model-analysis - This paper list focuses on the theoretical or empirical analysis of language models, e.g., the learning dynamics, expressive capacity, interpretability, generalization, and other interesting topics.

LLM Leaderboard

  • Chatbot Arena Leaderboard - a benchmark platform for large language models (LLMs) that features anonymous, randomized battles in a crowdsourced manner.
  • LiveBench - A Challenging, Contamination-Free LLM Benchmark.
  • Open LLM Leaderboard - aims to track, rank, and evaluate LLMs and chatbots as they are released.
  • AlpacaEval - An Automatic Evaluator for Instruction-following Language Models using Nous benchmark suite.
other leaderboards
  • ACLUE - an evaluation benchmark focused on ancient Chinese language comprehension.
  • BeHonest - A pioneering benchmark specifically designed to assess honesty in LLMs comprehensively.
  • Berkeley Function-Calling Leaderboard - evaluates LLM's ability to call external functions/tools.
  • Chinese Large Model Leaderboard - an expert-driven benchmark for Chineses LLMs.
  • CompassRank - CompassRank is dedicated to exploring the most advanced language and visual models, offering a comprehensive, objective, and neutral evaluation reference for the industry and research.
  • CompMix - a benchmark evaluating QA methods that operate over a mixture of heterogeneous input sources (KB, text, tables, infoboxes).
  • DreamBench++ - a benchmark for evaluating the performance of large language models (LLMs) in various tasks related to both textual and visual imagination.
  • FELM - a meta-benchmark that evaluates how well factuality evaluators assess the outputs of large language models (LLMs).
  • InfiBench - a benchmark designed to evaluate large language models (LLMs) specifically in their ability to answer real-world coding-related questions.
  • LawBench - a benchmark designed to evaluate large language models in the legal domain.
  • LLMEval - focuses on understanding how these models perform in various scenarios and analyzing results from an interpretability perspective.
  • M3CoT - a benchmark that evaluates large language models on a variety of multimodal reasoning tasks, including language, natural and social sciences, physical and social commonsense, temporal reasoning, algebra, and geometry.
  • MathEval - a comprehensive benchmarking platform designed to evaluate large models' mathematical abilities across 20 fields and nearly 30,000 math problems.
  • MixEval - a ground-truth-based dynamic benchmark derived from off-the-shelf benchmark mixtures, which evaluates LLMs with a highly capable model ranking (i.e., 0.96 correlation with Chatbot Arena) while running locally and quickly (6% the time and cost of running MMLU).
  • MMedBench - a benchmark that evaluates large language models' ability to answer medical questions across multiple languages.
  • MMToM-QA - a multimodal question-answering benchmark designed to evaluate AI models' cognitive ability to understand human beliefs and goals.
  • OlympicArena - a benchmark for evaluating AI models across multiple academic disciplines like math, physics, chemistry, biology, and more.
  • PubMedQA - a biomedical question-answering benchmark designed for answering research-related questions using PubMed abstracts.
  • SciBench - benchmark designed to evaluate large language models (LLMs) on solving complex, college-level scientific problems from domains like chemistry, physics, and mathematics.
  • SuperBench - a benchmark platform designed for evaluating large language models (LLMs) on a range of tasks, particularly focusing on their performance in different aspects such as natural language understanding, reasoning, and generalization.
  • SuperLim - a Swedish language understanding benchmark that evaluates natural language processing (NLP) models on various tasks such as argumentation analysis, semantic similarity, and textual entailment.
  • TAT-DQA - a large-scale Document Visual Question Answering (VQA) dataset designed for complex document understanding, particularly in financial reports.
  • TAT-QA - a large-scale question-answering benchmark focused on real-world financial data, integrating both tabular and textual information.
  • VisualWebArena - a benchmark designed to assess the performance of multimodal web agents on realistic visually grounded tasks.
  • We-Math - a benchmark that evaluates large multimodal models (LMMs) on their ability to perform human-like mathematical reasoning.
  • WHOOPS! - a benchmark dataset testing AI's ability to reason about visual commonsense through images that defy normal expectations.

Open LLM

DeepSeek
Alibaba
Meta
Mistral AI
Google
Apple
Microsoft
AllenAI
xAI
Cohere
01-ai
Baichuan
Nvidia
BLOOM
Zhipu AI
OpenBMB
RWKV Foundation
ElutherAI
Stability AI
BigCode
DataBricks
Shanghai AI Laboratory - [InternLM2-1.8|7|20B](https://huggingface.co/collections/internlm/internlm2-65b0ce04970888799707893c) - [InternLM-Math-7B|20B](https://huggingface.co/collections/internlm/internlm2-math-65b0ce88bf7d3327d0a5ad9f) - [InternLM-XComposer2-1.8|7B](https://huggingface.co/collections/internlm/internlm-xcomposer2-65b3706bf5d76208998e7477) - [InternVL-2|6|14|26](https://huggingface.co/collections/OpenGVLab/internvl-65b92d6be81c86166ca0dde4)

LLM Data

Reference: LLMDataHub

  • IBM data-prep-kit - Open-Source Toolkit for Efficient Unstructured Data Processing with Pre-built Modules and Local to Cluster Scalability.
  • Datatrove - Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.

LLM Evaluation:

  • lm-evaluation-harness - A framework for few-shot evaluation of language models.
  • lighteval - a lightweight LLM evaluation suite that Hugging Face has been using internally.
  • simple-evals - Eval tools by OpenAI.
other evaluation frameworks
  • OLMO-eval - a repository for evaluating open language models.
  • MixEval - A reliable click-and-go evaluation suite compatible with both open-source and proprietary models, supporting MixEval and other benchmarks.
  • HELM - Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models.
  • instruct-eval - This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
  • Giskard - Testing & evaluation library for LLM applications, in particular RAGs
  • LangSmith - a unified platform from LangChain framework for: evaluation, collaboration HITL (Human In The Loop), logging and monitoring LLM applications.
  • Ragas - a framework that helps you evaluate your Retrieval Augmented Generation (RAG) pipelines.

LLM Training Frameworks

  • Meta Lingua - a lean, efficient, and easy-to-hack codebase to research LLMs.
  • Litgpt - 20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
  • nanotron - Minimalistic large language model 3D-parallelism training.
  • DeepSpeed - DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
  • Megatron-LM - Ongoing research training transformer models at scale.
  • torchtitan - A native PyTorch Library for large model training.
other frameworks
  • Megatron-DeepSpeed - DeepSpeed version of NVIDIA's Megatron-LM that adds additional support for several features such as MoE model training, Curriculum Learning, 3D Parallelism, and others.
  • torchtune - A Native-PyTorch Library for LLM Fine-tuning.
  • veRL - veRL is a flexible and efficient RL framework for LLMs.
  • NeMo Framework - Generative AI framework built for researchers and PyTorch developers working on Large Language Models (LLMs), Multimodal Models (MMs), Automatic Speech Recognition (ASR), Text to Speech (TTS), and Computer Vision (CV) domains.
  • Colossal-AI - Making large AI models cheaper, faster, and more accessible.
  • BMTrain - Efficient Training for Big Models.
  • Mesh Tensorflow - Mesh TensorFlow: Model Parallelism Made Easier.
  • maxtext - A simple, performant and scalable Jax LLM!
  • GPT-NeoX - An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
  • Transformer Engine - A library for accelerating Transformer model training on NVIDIA GPUs.
  • OpenRLHF - An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT).
  • TRL - TRL is a full stack library where we provide a set of tools to train transformer language models with Reinforcement Learning, from the Supervised Fine-tuning step (SFT), Reward Modeling step (RM) to the Proximal Policy Optimization (PPO) step.
  • unslothai - A framework that specializes in efficient fine-tuning. On its GitHub page, you can find ready-to-use fine-tuning templates for various LLMs, allowing you to easily train your own data for free on the Google Colab cloud.

LLM Inference

Reference: llm-inference-solutions

  • SGLang - SGLang is a fast serving framework for large language models and vision language models.
  • vLLM - A high-throughput and memory-efficient inference and serving engine for LLMs.
  • llama.cpp - LLM inference in C/C++.
  • ollama - Get up and running with Llama 3, Mistral, Gemma, and other large language models.
  • TGI - a toolkit for deploying and serving Large Language Models (LLMs).
  • TensorRT-LLM - Nvidia Framework for LLM Inference
other deployment tools
  • FasterTransformer - NVIDIA Framework for LLM Inference(Transitioned to TensorRT-LLM)
  • MInference - To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which reduces inference latency by up to 10x for pre-filling on an A100 while maintaining accuracy.
  • exllama - A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
  • FastChat - A distributed multi-model LLM serving system with web UI and OpenAI-compatible RESTful APIs.
  • mistral.rs - Blazingly fast LLM inference.
  • SkyPilot - Run LLMs and batch jobs on any cloud. Get maximum cost savings, highest GPU availability, and managed execution -- all with a simple interface.
  • Haystack - an open-source NLP framework that allows you to use LLMs and transformer-based models from Hugging Face, OpenAI and Cohere to interact with your own data.
  • OpenLLM - Fine-tune, serve, deploy, and monitor any open-source LLMs in production. Used in production at BentoML for LLMs-based applications.
  • DeepSpeed-Mii - MII makes low-latency and high-throughput inference, similar to vLLM powered by DeepSpeed.
  • Text-Embeddings-Inference - Inference for text-embeddings in Rust, HFOIL Licence.
  • Infinity - Inference for text-embeddings in Python
  • LMDeploy - A high-throughput and low-latency inference and serving framework for LLMs and VLs
  • Liger-Kernel - Efficient Triton Kernels for LLM Training.

LLM Applications

Reference: awesome-llm-apps

  • dspy - DSPy: The framework for programming—not prompting—foundation models.
  • LangChain — A popular Python/JavaScript library for chaining sequences of language model prompts.
  • LlamaIndex — A Python library for augmenting LLM apps with data.
more applications
  • MLflow - MLflow: An open-source framework for the end-to-end machine learning lifecycle, helping developers track experiments, evaluate models/prompts, deploy models, and add observability with tracing.

  • Swiss Army Llama - Comprehensive set of tools for working with local LLMs for various tasks.

  • LiteChain - Lightweight alternative to LangChain for composing LLMs

  • magentic - Seamlessly integrate LLMs as Python functions

  • wechat-chatgpt - Use ChatGPT On Wechat via wechaty

  • promptfoo - Test your prompts. Evaluate and compare LLM outputs, catch regressions, and improve prompt quality.

  • Agenta - Easily build, version, evaluate and deploy your LLM-powered apps.

  • Serge - a chat interface crafted with llama.cpp for running Alpaca models. No API keys, entirely self-hosted!

  • Langroid - Harness LLMs with Multi-Agent Programming

  • Embedchain - Framework to create ChatGPT like bots over your dataset.

  • Opik - Confidently evaluate, test, and ship LLM applications with a suite of observability tools to calibrate language model outputs across your dev and production lifecycle.

  • IntelliServer - simplifies the evaluation of LLMs by providing a unified microservice to access and test multiple AI models.

  • Langchain-Chatchat - Formerly langchain-ChatGLM, local knowledge based LLM (like ChatGLM) QA app with langchain.

  • Search with Lepton - Build your own conversational search engine using less than 500 lines of code by LeptonAI.

  • Robocorp - Create, deploy and operate Actions using Python anywhere to enhance your AI agents and assistants. Batteries included with an extensive set of libraries, helpers and logging.

  • Tune Studio - Playground for devs to finetune & deploy LLMs

  • LLocalSearch - Locally running websearch using LLM chains

  • AI Gateway — Gateway streamlines requests to 100+ open & closed source models with a unified API. It is also production-ready with support for caching, fallbacks, retries, timeouts, loadbalancing, and can be edge-deployed for minimum latency.

  • talkd.ai dialog - Simple API for deploying any RAG or LLM that you want adding plugins.

  • Wllama - WebAssembly binding for llama.cpp - Enabling in-browser LLM inference

  • GPUStack - An open-source GPU cluster manager for running LLMs

  • MNN-LLM -- A Device-Inference framework, including LLM Inference on device(Mobile Phone/PC/IOT)

  • CAMEL - First LLM Multi-agent framework.

  • QA-Pilot - An interactive chat project that leverages Ollama/OpenAI/MistralAI LLMs for rapid understanding and navigation of GitHub code repository or compressed file resources.

  • Shell-Pilot - Interact with LLM using Ollama models(or openAI, mistralAI)via pure shell scripts on your Linux(or MacOS) system, enhancing intelligent system management without any dependencies.

  • MindSQL - A python package for Txt-to-SQL with self hosting functionalities and RESTful APIs compatible with proprietary as well as open source LLM.

  • Langfuse - Open Source LLM Engineering Platform 🪢 Tracing, Evaluations, Prompt Management, Evaluations and Playground.

  • AdalFlow - AdalFlow: The library to build&auto-optimize LLM applications.

  • Guidance — A handy looking Python library from Microsoft that uses Handlebars templating to interleave generation, prompting, and logical control.

  • Evidently — An open-source framework to evaluate, test and monitor ML and LLM-powered systems.

  • Chainlit — A Python library for making chatbot interfaces.

  • Guardrails.ai — A Python library for validating outputs and retrying failures. Still in alpha, so expect sharp edges and bugs.

  • Semantic Kernel — A Python/C#/Java library from Microsoft that supports prompt templating, function chaining, vectorized memory, and intelligent planning.

  • Prompttools — Open-source Python tools for testing and evaluating models, vector DBs, and prompts.

  • Outlines — A Python library that provides a domain-specific language to simplify prompting and constrain generation.

  • Promptify — A small Python library for using language models to perform NLP tasks.

  • Scale Spellbook — A paid product for building, comparing, and shipping language model apps.

  • PromptPerfect — A paid product for testing and improving prompts.

  • Weights & Biases — A paid product for tracking model training and prompt engineering experiments.

  • OpenAI Evals — An open-source library for evaluating task performance of language models and prompts.

  • Arthur Shield — A paid product for detecting toxicity, hallucination, prompt injection, etc.

  • LMQL — A programming language for LLM interaction with support for typed prompting, control flow, constraints, and tools.

  • ModelFusion - A TypeScript library for building apps with LLMs and other ML models (speech-to-text, text-to-speech, image generation).

  • OneKE — A bilingual Chinese-English knowledge extraction model with knowledge graphs and natural language processing technologies.

  • llm-ui - A React library for building LLM UIs.

  • Wordware - A web-hosted IDE where non-technical domain experts work with AI Engineers to build task-specific AI agents. We approach prompting as a new programming language rather than low/no-code blocks.

  • Wallaroo.AI - Deploy, manage, optimize any model at scale across any environment from cloud to edge. Let's you go from python notebook to inferencing in minutes.

  • Dify - An open-source LLM app development platform with an intuitive interface that streamlines AI workflows, model management, and production deployment.

  • LazyLLM - An open-source LLM app for building multi-agent LLMs applications in an easy and lazy way, supports model deployment and fine-tuning.

  • MemFree - Open Source Hybrid AI Search Engine, Instantly Get Accurate Answers from the Internet, Bookmarks, Notes, and Docs. Support One-Click Deployment

  • AutoRAG - Open source AutoML tool for RAG. Optimize the RAG answer quality automatically. From generation evaluation datset to deploying optimized RAG pipeline.

  • Epsilla - An all-in-one LLM Agent platform with your private data and knowledge, delivers your production-ready AI Agents on Day 1.

  • Arize-Phoenix - Open-source tool for ML observability that runs in your notebook environment. Monitor and fine tune LLM, CV and Tabular Models.

LLM Tutorials and Courses

LLM Books

Great thoughts about LLM

Miscellaneous

  • Emergent Mind - The latest AI news, curated & explained by GPT-4.
  • ShareGPT - Share your wildest ChatGPT conversations with one click.
  • Major LLMs + Data Availability
  • 500+ Best AI Tools
  • Cohere Summarize Beta - Introducing Cohere Summarize Beta: A New Endpoint for Text Summarization
  • chatgpt-wrapper - ChatGPT Wrapper is an open-source unofficial Python API and CLI that lets you interact with ChatGPT.
  • Cursor - Write, edit, and chat about your code with a powerful AI.
  • AutoGPT - an experimental open-source application showcasing the capabilities of the GPT-4 language model.
  • OpenAGI - When LLM Meets Domain Experts.
  • EasyEdit - An easy-to-use framework to edit large language models.
  • chatgpt-shroud - A Chrome extension for OpenAI's ChatGPT, enhancing user privacy by enabling easy hiding and unhiding of chat history. Ideal for privacy during screen shares.

Contributing

This is an active repository and your contributions are always welcome!

I will keep some pull requests open if I'm not sure if they are awesome for LLM, you could vote for them by adding 👍 to them.


If you have any question about this opinionated list, do not hesitate to contact me chengxin1998@stu.pku.edu.cn.

About

Awesome-LLM: a curated list of Large Language Model

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published