DeepSeek Coder:Let the Code Write Itself
DeepSeek Coder comprises a series of code language models trained from scratch on both 87% code and 13% natural language in English and Chinese, with each model pre-trained on 2T tokens.
We provide various sizes of the code model, ranging from 1B to 33B versions.
Each model is pre-trained on repo-level code corpus by employing a window size of 16K and a extra fill-in-the-blank task, resulting in foundational models (DeepSeek-Coder-Base).
We further fine-tune the base model with 2B tokens of inst...
Read more at deepseekcoder.github.io