SuperbEmphasis's picture
Create README.md
aff15f9 verified

this is a test.... trying to get Qwen3 30B to behave.

I started by fine-tuning SuperbEmphasis/Black-Eclipse-Test-ERP-RP-v2 with my improved dataset.

however this time, I cranked the activated experts to 64/128 to force more parameters to get trained.

Then I cranked the experts back down to 24/128.

it is surprisingly coherent. the biggest issue I have run into so far is that it REALLY doesnt like roleplay scenarios where '{{char}}' can be multiple characters. It wants to stay as the single initial characters This tells me that I need more scenario based examples...