fix docs
Browse files- config.json +4 -4
    	
        config.json
    CHANGED
    
    | @@ -1,6 +1,6 @@ | |
| 1 | 
             
            {
         | 
| 2 | 
             
              "architectures": [
         | 
| 3 | 
            -
                " | 
| 4 | 
             
              ],
         | 
| 5 | 
             
              "attention_dropout": 0.0,
         | 
| 6 | 
             
              "attn_type_list": [
         | 
| @@ -86,8 +86,8 @@ | |
| 86 | 
             
                1
         | 
| 87 | 
             
              ],
         | 
| 88 | 
             
              "auto_map": {
         | 
| 89 | 
            -
                "AutoConfig": "configuration_minimax_m1. | 
| 90 | 
            -
                "AutoModelForCausalLM": "modeling_minimax_m1. | 
| 91 | 
             
              },
         | 
| 92 | 
             
              "bos_token_id": null,
         | 
| 93 | 
             
              "eos_token_id": null,
         | 
| @@ -103,7 +103,7 @@ | |
| 103 | 
             
              "layernorm_mlp_alpha": 3.5565588200778455,
         | 
| 104 | 
             
              "layernorm_mlp_beta": 1.0,
         | 
| 105 | 
             
              "max_position_embeddings": 10240000,
         | 
| 106 | 
            -
              "model_type": " | 
| 107 | 
             
              "num_attention_heads": 64,
         | 
| 108 | 
             
              "num_experts_per_tok": 2,
         | 
| 109 | 
             
              "num_hidden_layers": 80,
         | 
|  | |
| 1 | 
             
            {
         | 
| 2 | 
             
              "architectures": [
         | 
| 3 | 
            +
                "MiniMaxM1ForCausalLM"
         | 
| 4 | 
             
              ],
         | 
| 5 | 
             
              "attention_dropout": 0.0,
         | 
| 6 | 
             
              "attn_type_list": [
         | 
|  | |
| 86 | 
             
                1
         | 
| 87 | 
             
              ],
         | 
| 88 | 
             
              "auto_map": {
         | 
| 89 | 
            +
                "AutoConfig": "configuration_minimax_m1.MiniMaxM1Config",
         | 
| 90 | 
            +
                "AutoModelForCausalLM": "modeling_minimax_m1.MiniMaxM1ForCausalLM"
         | 
| 91 | 
             
              },
         | 
| 92 | 
             
              "bos_token_id": null,
         | 
| 93 | 
             
              "eos_token_id": null,
         | 
|  | |
| 103 | 
             
              "layernorm_mlp_alpha": 3.5565588200778455,
         | 
| 104 | 
             
              "layernorm_mlp_beta": 1.0,
         | 
| 105 | 
             
              "max_position_embeddings": 10240000,
         | 
| 106 | 
            +
              "model_type": "minimax_m1",
         | 
| 107 | 
             
              "num_attention_heads": 64,
         | 
| 108 | 
             
              "num_experts_per_tok": 2,
         | 
| 109 | 
             
              "num_hidden_layers": 80,
         | 

