Skywork-13B-Base
: The model was trained on a high-quality cleaned dataset consisting of 3.2 trillion multilingual data (mainly Chinese and English) and code. It has demonstrated the best performance among models of similar scale in various evaluations and benchmark tests.
We have developed a data cleaning pipeline with great care to effectively clean and filter low-quality data and eliminate harmful information from text data. Our Skywork-13B-Base model is trained on a dataset with 3.2TB tokens that consists of high-quality Chinese, English, and code data, all of which have been thoroughly cleaned. The English data comprises 52.2% of the dataset, the Chinese data accounts for 39.6%, and the code data makes up 8%. This comprehensive approach ensures optimal performance for both Chinese and English while also maintaining the ability to handle code.
Compared to the Llama2-13B model, the Skywork-13B model adopts a relatively thinner and deeper network structure with 52 layers. At the same time, the FFN Dim and Hidden Dim are reduced to 12288 and 4608, respectively, to ensure that the model has a similar number of parameters as the original Llama-13B model. Based on our preliminary experimental results, a relatively thinner and deeper network structure can achieve better generalization performance under large batch size training. The detailed comparison between the Skywork-13B and Llama-2-13B models is as follows:
We use Byte-Pair Encoding (BPE) to tokenize the data, with a vocabulary size of 65536. Among them, there are 32000 Latin characters and subwords, 8000 Chinese characters and Unicode symbols, 25519 Chinese words, and the remaining 17 are reserved words.
We have chosen several hundred to thousands of high-quality articles that were published after September 1, 2023 across various fields. We have manually verified these articles to ensure their quality. It is important to note that none of the test data used in evaluating the Skywork model or any other models is included in their training set. Furthermore, the test data is diverse and of high quality, making it challenging for the models to gain an unfair advantage.
The figure below displays the performance of different open source models. Skywork-13B-Base achieves the best results.
Tech
Movie
Gov.
Game
Finance
General
Average
MOSS-7B
20.83
39.66
11.08
31.24
10.59
13.25
18.50
InternLM-7B
13.43
24.90
5.88
19.78
6.17
8.10
11.17
Qwen-7B
13.39
25.16
5.55
19.26
5.76
7.78
10.83
Baichuan2-7B
12.89
23.26
5.34
18.36
5.68
7.62
10.41
LLaMA2-13B
23.26
50.66
18.09
32.52
14.85
16.55
23.54
Xverse-13B
12.55
23.49
5.20
17.69
5.54
7.46
10.19
Baichuan-13B
12.38
22.46
5.21
17.59
5.42
7.37
10.03
Baichuan2-13B
12.14
21.85
5.05
17.15
5.35
7.24
9.81
Qwen-14B
11.90
22.43
4.89
16.94
5.24
7.03
9.67
InternLM-20B
12.34
22.06
5.75
17.45
5.73
7.78
10.34
Aquila2-34B
14.62
29.09
5.72
21.78
5.83
8.45
11.73
Skywork-13B-Base
11.58
21.84
4.76
17.28
4.92
6.82
9.42
评测数据和评测脚本(Loss Evaluation)
我们将评测数据和评测脚本也进行了开源,下载github上的代码运行下面命令则可以复现我们的结果。
We have also open-sourced the data and evaluation scripts. You can reproduce our results by running the following command.
We evaluated Skywork-13B-Base on several popular benchmarks, including C-Eval, MMLU, CMMLU, and GSM8K. Following the previous evaluation process, we tested the 5-shot results of C-Eval, MMLU, and CMMLU, and the 8-shot results of GSM8K. It can be seen that the Skywork-13B-Base model is among the top models in the Chinese open source model community, performing at an optimal level with the same parameter scale.
## preprocess continue pretraining data## Because pre-training data is usually large, we use a script to process the training data separately.
python train/pt_data_preprocess.py \
-t $MODEL_PATH \
-i data/pt_train.jsonl \
-o data_cache/pt_train_demo
## launch trainingexport WANDB_API_KEY=YOUR_WANDB_KEY
export WANDB_ENTITY=skywork
export WANDB_PROJECT=skywork-13b-opensource
export MODEL_PATH=skywork-13b-models/skywork-13b-base
export DATA_CACHE_DIR=data_cache/pt_train_demo/pt_train
bash bash_scripts/skywork_13b_pt.sh
## preprocess data and launch trainingexport WANDB_API_KEY=YOUR_WANDB_KEY
export WANDB_ENTITY=skywork
export WANDB_PROJECT=skywork-13b-opensource
export SFT_DATA_DIR=data/sft_data
export DATA_CACHE_DIR=data_cache/sft_train_demo
bash bash_scripts/skywork_13b_sft.sh
LoRA微调(PEFT)
使用Skywork-13B-Base模型以及LoRA进行预训练微调
## preprocess continue pretraining data## Because pre-training data is usually large, we use a script to process the training data separately.
python train/pt_data_preprocess.py \
-t $MODEL_PATH \
-i data/pt_train.jsonl \
-o data_cache/pt_train_demo
export WANDB_API_KEY=YOUR_WANDB_KEY
export WANDB_ENTITY=skywork
export WANDB_PROJECT=skywork-13b-opensource
export MODEL_PATH=skywork-13b-models/skywork-13b-base
export DATA_CACHE_DIR=data_cache/pt_train_demo/pt_train
bash bash_scripts/skywork_13b_pt_lora.sh
We hereby declare that the Skywork model should not be used for any activities that pose a threat to national or societal security or engage in unlawful actions. Additionally, we request users not to deploy the Skywork model for internet services without appropriate security reviews and records. We hope that all users will adhere to this principle to ensure that technological advancements occur in a regulated and lawful environment.
We have done our utmost to ensure the compliance of the data used during the model's training process. However, despite our extensive efforts, due to the complexity of the model and data, there may still be unpredictable risks and issues. Therefore, if any problems arise as a result of using the Skywork open-source model, including but not limited to data security issues, public opinion risks, or any risks and problems arising from the model being misled, abused, disseminated, or improperly utilized, we will not assume any responsibility.
The community usage of Skywork model requires
Skywork Community License
. The Skywork model supports commercial use. If you plan to use the Skywork model or its derivatives for commercial purposes, you must abide by terms and conditions within
Skywork Community License
.
引用和联系我们(Contact Us and Citation)
如果您觉得我们的工作对您有帮助,欢迎引用我们的论文~
If you find our work helpful, please feel free to cite our paper~
@misc{wei2023skywork,
title={Skywork: A More Open Bilingual Foundation Model},
author={Tianwen Wei and Liang Zhao and Lichang Zhang and Bo Zhu and Lijie Wang and Haihua Yang and Biye Li and Cheng Cheng and Weiwei Lü and Rui Hu and Chenxia Li and Liu Yang and Xilin Luo and Xuejie Wu and Lunan Liu and Wenjun Cheng and Peng Cheng and Jianhao Zhang and Xiaoyu Zhang and Lei Lin and Xiaokun Wang and Yutuan Ma and Chuanhai Dong and Yanqi Sun and Yifu Chen and Yongyi Peng and Xiaojuan Liang and Shuicheng Yan and Han Fang and Yahui Zhou},
year={2023},
eprint={2310.19341},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
@article{Skywork_Multi-Modal_Group_Empirical_Study_Towards_2023,
author = {Skywork Multi-Modal Group},
month = sep,
title = {{Empirical Study Towards Building An Effective Multi-Modal Large Language Model}},
year = {2023}
}
Runs of Skywork Skywork-13B-base on huggingface.co
1.5K
Total runs
0
24-hour runs
299
3-day runs
738
7-day runs
738
30-day runs
More Information About Skywork-13B-base huggingface.co Model
Skywork-13B-base huggingface.co is an AI model on huggingface.co that provides Skywork-13B-base's model effect (), which can be used instantly with this Skywork Skywork-13B-base model. huggingface.co supports a free trial of the Skywork-13B-base model, and also provides paid use of the Skywork-13B-base. Support call Skywork-13B-base model through api, including Node.js, Python, http.
Skywork-13B-base huggingface.co is an online trial and call api platform, which integrates Skywork-13B-base's modeling effects, including api services, and provides a free online trial of Skywork-13B-base, you can try Skywork-13B-base online for free by clicking the link below.
Skywork Skywork-13B-base online free url in huggingface.co:
Skywork-13B-base is an open source model from GitHub that offers a free installation service, and any user can find Skywork-13B-base on GitHub to install. At the same time, huggingface.co provides the effect of Skywork-13B-base install, users can directly use Skywork-13B-base installed effect in huggingface.co for debugging and trial. It also supports api for free installation.