File size: 579 Bytes
aff15f9 |
1 2 3 4 5 6 7 8 9 10 11 |
this is a test.... trying to get Qwen3 30B to behave. I started by fine-tuning SuperbEmphasis/Black-Eclipse-Test-ERP-RP-v2 with my improved dataset. however this time, I cranked the activated experts to 64/128 to force more parameters to get trained. Then I cranked the experts back down to 24/128. it is surprisingly coherent. the biggest issue I have run into so far is that it REALLY doesnt like roleplay scenarios where '{{char}}' can be multiple characters. It wants to stay as the single initial characters This tells me that I need more scenario based examples... |