this is a test.... trying to get Qwen3 30B to behave. | |
I started by fine-tuning | |
SuperbEmphasis/Black-Eclipse-Test-ERP-RP-v2 with my improved dataset. | |
however this time, I cranked the activated experts to 64/128 to force more parameters to get trained. | |
Then I cranked the experts back down to 24/128. | |
it is surprisingly coherent. the biggest issue I have run into so far is that it REALLY doesnt like roleplay scenarios where '{{char}}' can be multiple characters. It wants to stay as the single initial characters | |
This tells me that I need more scenario based examples... |