Add pipeline tag, link to paper
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -9,8 +9,12 @@ datasets:
|
|
| 9 |
- codingsteven/Llama-3-8B-chat
|
| 10 |
language:
|
| 11 |
- zh
|
|
|
|
|
|
|
| 12 |
base_model:
|
| 13 |
- meta-llama/Llama-3.1-8B
|
|
|
|
|
|
|
| 14 |
model-index:
|
| 15 |
- name: Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
|
| 16 |
results:
|
|
@@ -74,10 +78,10 @@ model-index:
|
|
| 74 |
---
|
| 75 |
|
| 76 |
# Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
|
| 77 |
-
This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Dlerp.
|
| 78 |
|
| 79 |
## Linked Paper
|
| 80 |
-
This model is associated with the paper: [Control-LLM](https://arxiv.org/abs/
|
| 81 |
|
| 82 |
## Evaluation Results
|
| 83 |
Here is an overview of the evaluation results and findings:
|
|
@@ -108,4 +112,4 @@ The table below summarizes evaluation results across Chinese tasks and original
|
|
| 108 |
- **MLU**: MMLU (Massive Multitask Language Understanding)
|
| 109 |
- **MLUP**: MMLU Pro
|
| 110 |
- **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
|
| 111 |
-
- **Overall**: Combined average across all tasks
|
|
|
|
| 9 |
- codingsteven/Llama-3-8B-chat
|
| 10 |
language:
|
| 11 |
- zh
|
| 12 |
+
metrics:
|
| 13 |
+
- accuracy
|
| 14 |
base_model:
|
| 15 |
- meta-llama/Llama-3.1-8B
|
| 16 |
+
pipeline_tag: text-generation
|
| 17 |
+
library_name: transformers
|
| 18 |
model-index:
|
| 19 |
- name: Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
|
| 20 |
results:
|
|
|
|
| 78 |
---
|
| 79 |
|
| 80 |
# Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
|
| 81 |
+
This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Dlerp, as described in [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979).
|
| 82 |
|
| 83 |
## Linked Paper
|
| 84 |
+
This model is associated with the paper: [Control-LLM](https://arxiv.org/abs/2410.14745).
|
| 85 |
|
| 86 |
## Evaluation Results
|
| 87 |
Here is an overview of the evaluation results and findings:
|
|
|
|
| 112 |
- **MLU**: MMLU (Massive Multitask Language Understanding)
|
| 113 |
- **MLUP**: MMLU Pro
|
| 114 |
- **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
|
| 115 |
+
- **Overall**: Combined average across all tasks
|