Skip to content

Commit ecc33eb

Browse files
authored
Update README.md
1 parent 959403f commit ecc33eb

File tree

1 file changed

+26
-2
lines changed

1 file changed

+26
-2
lines changed

README.md

Lines changed: 26 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,10 @@
1111

1212
Ling-Coder-Lite is a MoE LLM provided and open-sourced by InclusionAI, which has 16.8 billion parameters with 2.75 billion activated parameters. Ling-Coder-Lite performs impressively on coding tasks compared to existing models in the industry. Specifically, Ling-Coder-Lite further pre-training from an intermediate checkpoint of Ling-Lite, incorporating an additional 3 trillion tokens. This extended pre-training significantly boosts the coding abilities of Ling-Lite, while preserving its strong performance in general language tasks.
1313

14+
<p align="center">
15+
<img src="./figures/data-accuracy-efficiency.png" width="1500"/>
16+
<p>
17+
1418
## Model Downloads
1519

1620
You can download the following table to see the various parameters for your use case. If you are located in mainland China, we also provide the model on ModelScope.cn to speed up the download process.
@@ -28,9 +32,21 @@ You can download the following table to see the various parameters for your use
2832

2933
</div>
3034

35+
## Dataset Downloads
36+
37+
<div align="center">
38+
39+
| **Model** | **Samples** | **Download** |
40+
| :------------: | :----------------: | :--------------------------------------------------------------------------------------------------------------------------------------------------: |
41+
| Ling-Coder-SyntheticQA | 24M | [🤗 HuggingFace](https://huggingface.co/datasets/inclusionAI/Ling-Coder-SyntheticQA) <br>[🤖 ModelScope](https://modelscope.cn/datasets/inclusionAI/Ling-Coder-SyntheticQA) |
42+
| Ling-Coder-SFT | 5M | [🤗 HuggingFace](https://huggingface.co/datasets/inclusionAI/Ling-Coder-SFT) <br>[🤖 ModelScope](https://modelscope.cn/datasets/inclusionAI/Ling-Coder-SFT) |
43+
| Ling-Coder-DPO | 250K | [🤗 HuggingFace](https://huggingface.co/datasets/inclusionAI/Ling-Coder-DPO) <br>[🤖 ModelScope](https://modelscope.cn/datasets/inclusionAI/Ling-Coder-DPO) |
44+
45+
</div>
46+
3147
## Evaluation
3248

33-
Detailed evaluation results are reported in our [technical report](https://github.com/codefuse-ai/Ling-Coder-Lite/blob/master/Ling_Coder_Lite_Technique_Report.pdf).
49+
Detailed evaluation results are reported in our [technical report](https://arxiv.org/abs/2503.17793).
3450

3551
## Quickstart
3652

@@ -162,5 +178,13 @@ llamafactory-cli train examples/sft/ling_full_sft.yaml
162178
This code repository is licensed under [the MIT License](https://github.com/codefuse-ai/Ling-Coder-Lite/blob/master/LICENCE).
163179

164180
## Citation
181+
If you find our work is useful or helpful, please feel free to cite our paper as below.
165182

166-
[TBD]
183+
```
184+
@article{team2025every,
185+
title={Every Sample Matters: Leveraging Mixture-of-Experts and High-Quality Data for Efficient and Accurate Code LLM},
186+
author={Codefuse, Ling Team: Wenting Cai, Yuchen Cao, Chaoyu Chen, Chen Chen, Siba Chen, Qing Cui, Peng Di, Junpeng Fang, Zi Gong, Ting Guo, Zhengyu He, Yang Huang, Cong Li, Jianguo Li, Zheng Li, Shijie Lian, BingChang Liu, Songshan Luo, Shuo Mao, Min Shen, Jian Wu, Jiaolong Yang, Wenjie Yang, Tong Ye, Hang Yu, Wei Zhang, Zhenduo Zhang, Hailin Zhao, Xunjin Zheng, Jun Zhou},
187+
journal={arXiv preprint arXiv:2503.17793},
188+
year={2025}
189+
}
190+
```

0 commit comments

Comments
 (0)