qwen72l

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the Passthrough merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

slices:
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [0,30]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [30,31]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [30,31]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [30,31]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [31,32]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [31,32]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [31,32]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [32,33]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [32,33]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [32,33]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [33,34]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [33,34]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [33,34]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [34,35]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [34,35]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [34,35]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [35,36]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [35,36]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [35,36]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [36,37]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [36,37]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [36,37]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [37,38]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [37,38]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [37,38]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [38,39]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [38,39]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [38,39]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [39,40]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [39,40]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [39,40]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [40,41]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [40,41]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [40,41]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [41,42]
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [41,42]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [41,42]
      parameters:
        scale:
        - filter: o_proj
          value: 0.0
        - filter: down_proj
          value: 0.0
        - value: 1.0
  - sources:
    - model: ToastyPigeon/qwen2.5-14b-1m-unalign-v2
      layer_range: [42,48]
merge_method: passthrough
Downloads last month
53
Safetensors
Model size
21.4B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for Columbidae/qwen-depth-upscaled-72

Finetuned
(4)
this model