Change the repository type filter
All
Repositories list
153 repositories
huatuogpt-o1-70b
Public templateA medical LLM built on LLaMA-3.1-70B, employing detailed step-by-step reasoning for complex medical problem-solving. <metadata> gpu: A100 | collections: ["HF Transformers","Variable Inputs"] </metadata>Multi_LoRA_Adapter
PublicYouTube-Video-Summarizer
PublicGemma-7B
PublicGemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants.Mixral-8x7B
PublicMixtral is a large language model developed by Mistral AI, a French artificial intelligence company. It is a sparse Mixture of Experts (MoE) model with 8 experts per MLP, totaling 45 billion parameters. Mixtral is designed to handle contexts of up to 32,000 tokens.CodeLlama-70B
Publicbge-base-en-v1.5
Public- Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 34 billion parameters. This is the repository for the 13 instruct-tuned version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding.
TRELLIS
PublicLogo-Generator
PublicTenyxChat-8x7B-v1
PublicFacebook-bart-cnn
PublicBART model pre-trained on English language, and fine-tuned on CNN Daily Mail. It was introduced in the paper BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension by Lewis et al. and first released in [this repository (https://github.com/pytorch/fairseq/tree/master/examples/bart).Llama3-TenyxChat-70B
PublicSmaug-72B
PublicSmaug-72B - which topped the Hugging Face LLM leaderboard and it’s the first model with an average score of 80, making it the world’s best open-source foundation model.idefics-9b-instruct-8bit
PublicIDEFICS (Image-aware Decoder Enhanced à la Flamingo with Interleaved Cross-attentionS) is an open-access reproduction of Flamingo, a closed-source visual language model developed by Deepmind. Like GPT-4, the multimodal model accepts arbitrary sequences of image and text inputs and produces text outputs.distil-whisper-large-v3
Public templateDistilled model which is 49% smaller and 6.3× faster while maintaining near accuracy, especially on long-form transcription. <metadata> gpu: T4 | collections: ["HF Transformers"] </metadata>Llama-2-13b-chat-AWQ
PublicLlama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. Links to other models can be found in the index at the bottom.Command-r-v01
Public template35B model delivering high performance in reasoning, summarization, and question answering. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>Document-RAG-QnA
PublicDocument-RAG-Upload
PublicThis is a semantic search application build using Inferless and Pinecone.Flan-UL2
PublicDonut-docVQA
Publicstable-diffusion-2-1
Public templateText-to-image model, refined for higher fidelity, improved prompt adherence, and diverse creative outputs. <metadata> gpu: T4 | collections: ["Diffusers"] </metadata>Llama-3.1-70B-awq
Public