merge_method: ties base_model: yamatazen/LorablatedStock-12B models: - model: yamatazen/LorablatedStock-12B parameters: weight: 0.65 density: 1.0 - model: yamatazen/EtherealAurora-12B-v2 parameters: weight: 0.35 density: 1.0 parameters: normalize: false int8_mask: false dtype: bfloat16 layer_parameters: - filter: "attn" sources: - model: yamatazen/LorablatedStock-12B weight: 0.6 - model: yamatazen/EtherealAurora-12B-v2 weight: 0.4 - filter: "mlp" sources: - model: yamatazen/LorablatedStock-12B weight: 0.55 - model: yamatazen/EtherealAurora-12B-v2 weight: 0.45 - filter: "embed_tokens" sources: - model: yamatazen/LorablatedStock-12B weight: 1.0 - model: yamatazen/EtherealAurora-12B-v2 weight: 0.0 - filter: "layer_norm" sources: - model: yamatazen/LorablatedStock-12B weight: 0.7 - model: yamatazen/EtherealAurora-12B-v2 weight: 0.3