woodwardmw commited on
Commit
b268a54
·
verified ·
1 Parent(s): 35b31d4

End of training

Browse files
Files changed (3) hide show
  1. README.md +5 -9
  2. generation_config.json +11 -0
  3. tokenizer_config.json +1 -1
README.md CHANGED
@@ -1,13 +1,10 @@
1
  ---
2
- library_name: peft
3
  license: apache-2.0
4
  base_model: unsloth/orpheus-3b-0.1-pretrained
5
  tags:
6
- - base_model:adapter:unsloth/orpheus-3b-0.1-pretrained
7
- - lora
8
- - transformers
9
  - unsloth
10
- pipeline_tag: text-generation
11
  model-index:
12
  - name: nyiha_nt-tts-lora
13
  results: []
@@ -43,7 +40,7 @@ The following hyperparameters were used during training:
43
  - seed: 3407
44
  - gradient_accumulation_steps: 4
45
  - total_train_batch_size: 4
46
- - optimizer: Use OptimizerNames.ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_steps: 5
49
  - num_epochs: 5.0
@@ -54,8 +51,7 @@ The following hyperparameters were used during training:
54
 
55
  ### Framework versions
56
 
57
- - PEFT 0.16.0
58
- - Transformers 4.53.2
59
  - Pytorch 2.7.1+cu126
60
  - Datasets 4.0.0
61
- - Tokenizers 0.21.2
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: unsloth/orpheus-3b-0.1-pretrained
5
  tags:
 
 
 
6
  - unsloth
7
+ - generated_from_trainer
8
  model-index:
9
  - name: nyiha_nt-tts-lora
10
  results: []
 
40
  - seed: 3407
41
  - gradient_accumulation_steps: 4
42
  - total_train_batch_size: 4
43
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
44
  - lr_scheduler_type: linear
45
  - lr_scheduler_warmup_steps: 5
46
  - num_epochs: 5.0
 
51
 
52
  ### Framework versions
53
 
54
+ - Transformers 4.53.3
 
55
  - Pytorch 2.7.1+cu126
56
  - Datasets 4.0.0
57
+ - Tokenizers 0.21.2
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 128000,
4
+ "do_sample": true,
5
+ "eos_token_id": 128009,
6
+ "max_length": 131072,
7
+ "pad_token_id": 128004,
8
+ "temperature": 0.6,
9
+ "top_p": 0.9,
10
+ "transformers_version": "4.53.3"
11
+ }
tokenizer_config.json CHANGED
@@ -231535,7 +231535,7 @@
231535
  ],
231536
  "model_max_length": 131072,
231537
  "pad_token": "<|finetune_right_pad_id|>",
231538
- "padding_side": "right",
231539
  "tokenizer_class": "PreTrainedTokenizerFast",
231540
  "unk_token": null
231541
  }
 
231535
  ],
231536
  "model_max_length": 131072,
231537
  "pad_token": "<|finetune_right_pad_id|>",
231538
+ "padding_side": "left",
231539
  "tokenizer_class": "PreTrainedTokenizerFast",
231540
  "unk_token": null
231541
  }