apply quentin's update here
Browse files
bonus-unit1/bonus-unit1.ipynb
CHANGED
@@ -184,7 +184,7 @@
|
|
184 |
"import torch\n",
|
185 |
"import json\n",
|
186 |
"\n",
|
187 |
-
"from transformers import AutoModelForCausalLM, AutoTokenizer,
|
188 |
"from datasets import load_dataset\n",
|
189 |
"from trl import SFTConfig, SFTTrainer\n",
|
190 |
"from peft import LoraConfig, TaskType\n",
|
@@ -319,10 +319,7 @@
|
|
319 |
"source": [
|
320 |
"dataset = dataset.map(preprocess, remove_columns=\"messages\")\n",
|
321 |
"dataset = dataset[\"train\"].train_test_split(0.1)\n",
|
322 |
-
"print(dataset)
|
323 |
-
"\n",
|
324 |
-
"dataset[\"train\"] = dataset[\"train\"].select(range(100))\n",
|
325 |
-
"dataset[\"test\"] = dataset[\"test\"].select(range(10))"
|
326 |
]
|
327 |
},
|
328 |
{
|
|
|
184 |
"import torch\n",
|
185 |
"import json\n",
|
186 |
"\n",
|
187 |
+
"from transformers import AutoModelForCausalLM, AutoTokenizer, set_seed\n",
|
188 |
"from datasets import load_dataset\n",
|
189 |
"from trl import SFTConfig, SFTTrainer\n",
|
190 |
"from peft import LoraConfig, TaskType\n",
|
|
|
319 |
"source": [
|
320 |
"dataset = dataset.map(preprocess, remove_columns=\"messages\")\n",
|
321 |
"dataset = dataset[\"train\"].train_test_split(0.1)\n",
|
322 |
+
"print(dataset)"
|
|
|
|
|
|
|
323 |
]
|
324 |
},
|
325 |
{
|