models: - model: mlabonne/gemma-3-4b-it-abliterated - model: mshojaei77/gemma-3-4b-persian-v0 base_model: mlabonne/gemma-3-4b-it-abliterated merge_method: slerp dtype: bfloat16 # Better stability for precision-sensitive merges parameters: density: 0.5 weight: - filter: "self_attn" value: [0.75, 0.4, 0.25, 0.4, 0.75] # U-shaped attention weighting - filter: "mlp" value: [0.25, 0.6, 0.9, 0.6, 0.25] # Λ-shaped MLP weighting t: [0.15, 0.35, 0.65, 0.35, 0.15] # Optimized linguistic injection