TokenHD-1.7B

TokenHD is a token-level hallucination detector trained on top of Qwen/Qwen3-1.7B using the TokenHD pipeline. It assigns a hallucination probability to each token in an LLM-generated response, enabling fine-grained localization of errors without requiring predefined step segmentation.

Paper: arxiv.org/abs/2605.12384
Code: github.com/rmin2000/TokenHD
Training Data: mr233/TokenHD-training-data


Model Details

Property Value
Base model Qwen/Qwen3-1.7B
Architecture AutoModelForTokenClassification (num_labels=1)
Training domain Mathematics (competition-level problems)
Output Per-token hallucination probability (sigmoid of logits)

Usage

from transformers import AutoTokenizer, AutoModelForTokenClassification
import torch

model_id = "mr233/TokenHD-1.7B"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForTokenClassification.from_pretrained(model_id, num_labels=1)
model.eval()

problem = "What is the capital of France?"
response = "The capital of France is London."

messages = [
    {"role": "user", "content": problem},
    {"role": "assistant", "content": response},
]
input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=False)[:-2]
input_tensor = torch.tensor(input_ids).unsqueeze(0)

with torch.no_grad():
    logits = model(input_ids=input_tensor).logits  # shape: (1, seq_len, 1)

# scores for response tokens only
response_ids = tokenizer.encode(response, add_special_tokens=False)
scores = torch.sigmoid(logits.squeeze(-1).squeeze(0))[-len(response_ids):]
# scores[i] is the hallucination probability for the i-th response token

Evaluation

TokenHD models are evaluated with two metrics:

  • S_incor: Token-level F1 on hallucinated (incorrect) responses — measures how precisely the detector localizes errors.
  • S_cor: Recall on hallucination-free (correct) responses — measures how rarely the detector raises false alarms.
Downloads last month
18
Safetensors
Model size
2B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for mr233/TokenHD-1.7B

Finetuned
Qwen/Qwen3-1.7B
Finetuned
(717)
this model

Collection including mr233/TokenHD-1.7B

Paper for mr233/TokenHD-1.7B