Presentation

Piaget, a language model finetuned on 15k psychological and philosophical reasoning traces.

Piaget is based on Qwen3 and was finetuned on a subset of open reasoning traces from Dolphin R1 and General Reasoning.

Available sizes are: 0.6B, 1.7B, 4B, 8B.

How to use

from transformers import AutoTokenizer, AutoModelForCausalLM
from transformers.pipelines import pipeline
import torch

repo = "gustavecortal/Piaget-8B"
tokenizer = AutoTokenizer.from_pretrained(repo, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
   repo, torch_dtype=torch.bfloat16, device_map="auto", trust_remote_code=True
)

pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)

prompt = tokenizer.apply_chat_template(
   [
       {
           "role": "user",
           "content": "Create a new psychotherapeutic technique based on cybernetic principles",
       }
   ],
   tokenize=False,
   add_generation_prompt=True,
   enable_thinking=True,
)

print(pipe(prompt, max_new_tokens=2048, do_sample=True)[0]["generated_text"])

Methodology

We perform domain filtering on Dolphin R1 and General Reasoning.

Prompts are embedded, clustered with k-means (k=20 000) and majority-voted for domain labels using Qwen3-1.7B, following the Intelligent Internet pipeline.

Clusters tagged psychology or philosophy were retained for LoRA finetuning (rank=8, alpha=16, max length=2048, epoch=1, batch size=16).

This work was performed using HPC resources (Jean Zay supercomputer) from GENCI-IDRIS (Grant 20XX-AD011014205).

Inspiration

Piaget aims to reason about psychological and philosophical concepts such as self-image, emotion, and existence.

Piaget was inspired by my position paper on emotion analysis: Improving Language Models for Emotion Analysis: Insights from Cognitive Science.

Contact

Mail: [email protected]

Website: gustavecortal.com

Downloads last month
3
Safetensors
Model size
8.19B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for gustavecortal/Piaget-8B

Base model

Qwen/Qwen3-8B-Base
Finetuned
Qwen/Qwen3-8B
Finetuned
(129)
this model
Quantizations
2 models

Datasets used to train gustavecortal/Piaget-8B

Collection including gustavecortal/Piaget-8B