Model Card for Model ID

This a Chinese LLaMA2, built upon the original LLaMA2 with continue pre-training on 12B corpus.

Continue Pre-training Data:

  • WuDaoCorpora
  • BaiduBaike
  • Baidu News
  • Tiger Dataset
  • C4 samples

SFT Data:

  • Chinese SFT Data:
    • Alpaca-GPT4-zh
    • InstinWild-ch
    • Psychology-instruction
    • Firefly dataset
  • English SFT Data:
    • Code-Alpaca
    • InstinWild-en
    • School-math dataset
    • Ultrachat dataset

Results

  • CEval: 34.47%
  • CMMLU: 34.58%
  • MMLU: 41.23%
Downloads last month
6
Safetensors
Model size
6.89B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.