Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

TinyLlama-1.1B-Chat-v0.6-GGUF · Hugging Face #166

Open
1 task
irthomasthomas opened this issue Dec 22, 2023 · 0 comments
Open
1 task

TinyLlama-1.1B-Chat-v0.6-GGUF · Hugging Face #166

irthomasthomas opened this issue Dec 22, 2023 · 0 comments
Labels
llm Large Language Models llm-experiments experiments with large language models Models LLM and ML model repos and links

Comments

@irthomasthomas
Copy link
Owner

  • afrideva/TinyLlama-1.1B-Chat-v0.6-GGUF · Hugging Face

    This is the chat model finetuned on top of TinyLlama/TinyLlama-1.1B-intermediate-step-955k-2T. We follow HF's Zephyr's training recipe. The model was " initially fine-tuned on a variant of the UltraChat dataset, which contains a diverse range of synthetic dialogues generated by ChatGPT. We then further aligned the model with 🤗 TRL's DPOTrainer on the openbmb/UltraFeedback dataset, which contain 64k prompts and model completions that are ranked by GPT-4."

How to use

You will need the transformers>=4.34 Do check the TinyLlama github page for more information.

@irthomasthomas irthomasthomas added inbox-url unclassified Choose this if none of the other labels (bar New Label) fit the content. llm Large Language Models llm-experiments experiments with large language models Models LLM and ML model repos and links and removed unclassified Choose this if none of the other labels (bar New Label) fit the content. labels Dec 22, 2023
@ShellLM ShellLM removed the llama label May 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
llm Large Language Models llm-experiments experiments with large language models Models LLM and ML model repos and links
Projects
None yet
Development

No branches or pull requests

2 participants