Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

deepseek-coder-6.7b-instruct-8.0bpw-h8-exl2-2 · Hugging Face #189

Open
1 of 2 tasks
irthomasthomas opened this issue Dec 30, 2023 · 0 comments
Open
1 of 2 tasks
Labels
llm Large Language Models MachineLearning ML Models, Training and Inference Models LLM and ML model repos and links

Comments

@irthomasthomas
Copy link
Owner

irthomasthomas commented Dec 30, 2023

  1. Introduction of Deepseek Coder

Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese. We provide various sizes of the code model, ranging from 1B to 33B versions. Each model is pre-trained on project-level code corpus by employing a window size of 16K and a extra fill-in-the-blank task, to support project-level code completion and infilling. For coding capabilities, Deepseek Coder achieves state-of-the-art performance among open-source code models on multiple programming languages and various benchmarks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
llm Large Language Models MachineLearning ML Models, Training and Inference Models LLM and ML model repos and links
Projects
None yet
Development

No branches or pull requests

1 participant