Skip to content
Change the repository type filter

All

    Repositories list

    • Python
      0000Updated Mar 6, 2025Mar 6, 2025
    • huatuogpt-o1-70b

      Public template
      A medical LLM built on LLaMA-3.1-70B, employing detailed step-by-step reasoning for complex medical problem-solving. <metadata> gpu: A100 | collections: ["HF Transformers","Variable Inputs"] </metadata>
      Python
      0000Updated Mar 5, 2025Mar 5, 2025
    • Python
      1002Updated Mar 5, 2025Mar 5, 2025
    • Python
      1000Updated Mar 5, 2025Mar 5, 2025
    • Gemma-7B

      Public
      Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants.
      Python
      2100Updated Mar 5, 2025Mar 5, 2025
    • Mixtral is a large language model developed by Mistral AI, a French artificial intelligence company. It is a sparse Mixture of Experts (MoE) model with 8 experts per MLP, totaling 45 billion parameters. Mixtral is designed to handle contexts of up to 32,000 tokens.
      Python
      1000Updated Mar 5, 2025Mar 5, 2025
    • Python
      1000Updated Mar 5, 2025Mar 5, 2025
    • Python
      2000Updated Mar 5, 2025Mar 5, 2025
    • Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 34 billion parameters. This is the repository for the 13 instruct-tuned version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding.
      Python
      3000Updated Mar 5, 2025Mar 5, 2025
    • Python
      0000Updated Mar 5, 2025Mar 5, 2025
    • Python
      0000Updated Mar 5, 2025Mar 5, 2025
    • TRELLIS

      Public
      Python
      3000Updated Mar 5, 2025Mar 5, 2025
    • Python
      4100Updated Mar 5, 2025Mar 5, 2025
    • Python
      2200Updated Mar 5, 2025Mar 5, 2025
    • Python
      0000Updated Mar 5, 2025Mar 5, 2025
    • BART model pre-trained on English language, and fine-tuned on CNN Daily Mail. It was introduced in the paper BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension by Lewis et al. and first released in [this repository (https://github.com/pytorch/fairseq/tree/master/examples/bart).
      Python
      2800Updated Mar 5, 2025Mar 5, 2025
    • Python
      2000Updated Mar 5, 2025Mar 5, 2025
    • Smaug-72B

      Public
      Smaug-72B - which topped the Hugging Face LLM leaderboard and it’s the first model with an average score of 80, making it the world’s best open-source foundation model.
      Python
      51600Updated Mar 5, 2025Mar 5, 2025
    • IDEFICS (Image-aware Decoder Enhanced à la Flamingo with Interleaved Cross-attentionS) is an open-access reproduction of Flamingo, a closed-source visual language model developed by Deepmind. Like GPT-4, the multimodal model accepts arbitrary sequences of image and text inputs and produces text outputs.
      Python
      3000Updated Mar 5, 2025Mar 5, 2025
    • Distilled model which is 49% smaller and 6.3× faster while maintaining near accuracy, especially on long-form transcription. <metadata> gpu: T4 | collections: ["HF Transformers"] </metadata>
      Python
      6300Updated Mar 5, 2025Mar 5, 2025
    • Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. Links to other models can be found in the index at the bottom.
      Python
      3000Updated Mar 5, 2025Mar 5, 2025
    • Python
      0000Updated Mar 5, 2025Mar 5, 2025
    • 0000Updated Mar 5, 2025Mar 5, 2025
    • Command-r-v01

      Public template
      35B model delivering high performance in reasoning, summarization, and question answering. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>
      Python
      3200Updated Mar 5, 2025Mar 5, 2025
    • Python
      0000Updated Mar 5, 2025Mar 5, 2025
    • This is a semantic search application build using Inferless and Pinecone.
      Python
      4000Updated Mar 5, 2025Mar 5, 2025
    • Flan-UL2

      Public
      Python
      1000Updated Mar 5, 2025Mar 5, 2025
    • Python
      1000Updated Mar 5, 2025Mar 5, 2025
    • stable-diffusion-2-1

      Public template
      Text-to-image model, refined for higher fidelity, improved prompt adherence, and diverse creative outputs. <metadata> gpu: T4 | collections: ["Diffusers"] </metadata>
      Python
      8000Updated Mar 5, 2025Mar 5, 2025
    • Python
      2000Updated Mar 5, 2025Mar 5, 2025