s-emanuilov commited on
Commit
f686dc9
·
verified ·
1 Parent(s): 3c12b87

Model update

Browse files
adapter_config.json CHANGED
@@ -13,22 +13,22 @@
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
- "lora_alpha": 32,
17
  "lora_bias": false,
18
  "lora_dropout": 0,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
- "r": 32,
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "q_proj",
28
  "o_proj",
 
29
  "v_proj",
30
  "k_proj",
31
- "gate_proj",
32
  "down_proj",
33
  "up_proj"
34
  ],
 
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
+ "lora_alpha": 16,
17
  "lora_bias": false,
18
  "lora_dropout": 0,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
+ "r": 16,
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "gate_proj",
28
  "o_proj",
29
+ "q_proj",
30
  "v_proj",
31
  "k_proj",
 
32
  "down_proj",
33
  "up_proj"
34
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68a60b9d6bd223b7f2f4b0520acb474a7d496d2e210e1fe24f4a8c2149c089cd
3
- size 166182480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cac36a378e18d7633646e1854777384570ea9f9bdfc062c834902bf8ef8fac47
3
+ size 83115256
chat_template.jinja CHANGED
@@ -1,4 +1,8 @@
1
- {{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '
 
 
2
  ' + message['content'] | trim + '<end_of_turn>
3
- ' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model
4
- '}}{% endif %}
 
 
 
1
+ {{ bos_token }}{% if messages[0]['role'] == 'system' %}{{'<start_of_turn>user
2
+ ' + messages[0]['content'] | trim + ' ' + messages[1]['content'] | trim + '<end_of_turn>
3
+ '}}{% set messages = messages[2:] %}{% endif %}{% for message in messages %}{% if message['role'] == 'user' %}{{'<start_of_turn>user
4
  ' + message['content'] | trim + '<end_of_turn>
5
+ '}}{% elif message['role'] == 'assistant' %}{{'<start_of_turn>model
6
+ ' + message['content'] | trim + '<end_of_turn>
7
+ ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<start_of_turn>model
8
+ ' }}{% endif %}
special_tokens_map.json CHANGED
@@ -1,34 +1,6 @@
1
  {
2
- "additional_special_tokens": [
3
- "<start_of_turn>",
4
- "<end_of_turn>"
5
- ],
6
- "bos_token": {
7
- "content": "<bos>",
8
- "lstrip": false,
9
- "normalized": false,
10
- "rstrip": false,
11
- "single_word": false
12
- },
13
- "eos_token": {
14
- "content": "<eos>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false
19
- },
20
- "pad_token": {
21
- "content": "<pad>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false
26
- },
27
- "unk_token": {
28
- "content": "<unk>",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false
33
- }
34
  }
 
1
  {
2
+ "bos_token": "<bos>",
3
+ "eos_token": "<end_of_turn>",
4
+ "pad_token": "<pad>",
5
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6ce83119bb404f7f0a6e621b76759d476357dcd01241a90f9ca136ae2b3c11c
3
  size 34362972
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f4ec9e12434bbdee07008abfb928df9354ad1aee62d988e7bb41e55b989f4e6
3
  size 34362972
tokenizer_config.json CHANGED
@@ -11,7 +11,7 @@
11
  "special": true
12
  },
13
  "1": {
14
- "content": "<eos>",
15
  "lstrip": false,
16
  "normalized": false,
17
  "rstrip": false,
@@ -859,7 +859,7 @@
859
  "special": true
860
  },
861
  "107": {
862
- "content": "<end_of_turn>",
863
  "lstrip": false,
864
  "normalized": false,
865
  "rstrip": false,
@@ -1995,24 +1995,12 @@
1995
  "special": false
1996
  }
1997
  },
1998
- "additional_special_tokens": [
1999
- "<start_of_turn>",
2000
- "<end_of_turn>"
2001
- ],
2002
  "bos_token": "<bos>",
2003
  "clean_up_tokenization_spaces": false,
2004
- "eos_token": "<eos>",
2005
  "extra_special_tokens": {},
2006
- "max_length": 8192,
2007
- "model_max_length": 8192,
2008
  "pad_token": "<pad>",
2009
- "padding_side": "right",
2010
- "sp_model_kwargs": {},
2011
- "spaces_between_special_tokens": false,
2012
- "stride": 0,
2013
- "tokenizer_class": "GemmaTokenizer",
2014
- "truncation_side": "right",
2015
- "truncation_strategy": "longest_first",
2016
- "unk_token": "<unk>",
2017
- "use_default_system_prompt": false
2018
  }
 
11
  "special": true
12
  },
13
  "1": {
14
+ "content": "<end_of_turn>",
15
  "lstrip": false,
16
  "normalized": false,
17
  "rstrip": false,
 
859
  "special": true
860
  },
861
  "107": {
862
+ "content": "<eos>",
863
  "lstrip": false,
864
  "normalized": false,
865
  "rstrip": false,
 
1995
  "special": false
1996
  }
1997
  },
 
 
 
 
1998
  "bos_token": "<bos>",
1999
  "clean_up_tokenization_spaces": false,
2000
+ "eos_token": "<end_of_turn>",
2001
  "extra_special_tokens": {},
2002
+ "model_max_length": 1000000000000000019884624838656,
 
2003
  "pad_token": "<pad>",
2004
+ "tokenizer_class": "GemmaTokenizerFast",
2005
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
2006
  }