File size: 579 Bytes
aff15f9
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
this is a test.... trying to get Qwen3 30B to behave.

I started by fine-tuning
SuperbEmphasis/Black-Eclipse-Test-ERP-RP-v2 with my improved dataset.

however this time, I cranked the activated experts to 64/128 to force more parameters to get trained.

Then I cranked the experts back down to 24/128. 

it is surprisingly coherent.  the biggest issue I have run into so far is that it REALLY doesnt like roleplay scenarios where '{{char}}' can be multiple characters.  It wants to stay as the single initial characters
  This tells me that I need more scenario based examples...