Tima5 commited on
Commit
3b8bead
·
verified ·
1 Parent(s): 5accf0d

Update configuration_timaai.py

Browse files
Files changed (1) hide show
  1. configuration_timaai.py +10 -10
configuration_timaai.py CHANGED
@@ -3,12 +3,12 @@ from transformers.utils import logging
3
 
4
  logger = logging.get_logger(__name__)
5
 
6
- DEEPSEEK_PRETRAINED_CONFIG_ARCHIVE_MAP = {}
7
- class DeepseekV3Config(PretrainedConfig):
8
  r"""
9
- This is the configuration class to store the configuration of a [`DeepseekV3Model`]. It is used to instantiate an DeepSeek
10
  model according to the specified arguments, defining the model architecture. Instantiating a configuration with the
11
- defaults will yield a similar configuration to that of the DeepSeek-V3.
12
 
13
  Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the
14
  documentation from [`PretrainedConfig`] for more information.
@@ -17,7 +17,7 @@ class DeepseekV3Config(PretrainedConfig):
17
  Args:
18
  vocab_size (`int`, *optional*, defaults to 129280):
19
  Vocabulary size of the Deep model. Defines the number of different tokens that can be represented by the
20
- `inputs_ids` passed when calling [`DeepseekV3Model`]
21
  hidden_size (`int`, *optional*, defaults to 4096):
22
  Dimension of the hidden representations.
23
  intermediate_size (`int`, *optional*, defaults to 11008):
@@ -27,7 +27,7 @@ class DeepseekV3Config(PretrainedConfig):
27
  num_hidden_layers (`int`, *optional*, defaults to 32):
28
  Number of hidden layers in the Transformer decoder.
29
  num_nextn_predict_layers (`int`, *optional*, defaults to 1):
30
- Number of nextn predict layers in the DeepSeekV3 Model.
31
  num_attention_heads (`int`, *optional*, defaults to 32):
32
  Number of attention heads for each attention layer in the Transformer decoder.
33
  n_shared_experts (`int`, *optional*, defaults to None):
@@ -102,16 +102,16 @@ class DeepseekV3Config(PretrainedConfig):
102
  The dropout ratio for the attention probabilities.
103
 
104
  ```python
105
- >>> from transformers import DeepseekV3Model, DeepseekV3Config
106
 
107
- >>> # Initializing a Deepseek-V3 style configuration
108
- >>> configuration = DeepseekV3Config()
109
 
110
  >>> # Accessing the model configuration
111
  >>> configuration = model.config
112
  ```"""
113
 
114
- model_type = "deepseek_v3"
115
  keys_to_ignore_at_inference = ["past_key_values"]
116
 
117
  def __init__(
 
3
 
4
  logger = logging.get_logger(__name__)
5
 
6
+ timaai_PRETRAINED_CONFIG_ARCHIVE_MAP = {}
7
+ class timaaiV3Config(PretrainedConfig):
8
  r"""
9
+ This is the configuration class to store the configuration of a [`timaaiV3Model`]. It is used to instantiate an timaai
10
  model according to the specified arguments, defining the model architecture. Instantiating a configuration with the
11
+ defaults will yield a similar configuration to that of the timaai-V3.
12
 
13
  Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the
14
  documentation from [`PretrainedConfig`] for more information.
 
17
  Args:
18
  vocab_size (`int`, *optional*, defaults to 129280):
19
  Vocabulary size of the Deep model. Defines the number of different tokens that can be represented by the
20
+ `inputs_ids` passed when calling [`timaaiV3Model`]
21
  hidden_size (`int`, *optional*, defaults to 4096):
22
  Dimension of the hidden representations.
23
  intermediate_size (`int`, *optional*, defaults to 11008):
 
27
  num_hidden_layers (`int`, *optional*, defaults to 32):
28
  Number of hidden layers in the Transformer decoder.
29
  num_nextn_predict_layers (`int`, *optional*, defaults to 1):
30
+ Number of nextn predict layers in the timaaiV3 Model.
31
  num_attention_heads (`int`, *optional*, defaults to 32):
32
  Number of attention heads for each attention layer in the Transformer decoder.
33
  n_shared_experts (`int`, *optional*, defaults to None):
 
102
  The dropout ratio for the attention probabilities.
103
 
104
  ```python
105
+ >>> from transformers import timaaiV3Model, timaaiV3Config
106
 
107
+ >>> # Initializing a timaai-V3 style configuration
108
+ >>> configuration = timaaiV3Config()
109
 
110
  >>> # Accessing the model configuration
111
  >>> configuration = model.config
112
  ```"""
113
 
114
+ model_type = "timaai_v3"
115
  keys_to_ignore_at_inference = ["past_key_values"]
116
 
117
  def __init__(