Intermediate Checkpoints Release

For the first time among Korean-targeted LLMs, we’re releasing intermediate checkpoints from the Tri family—0.5B, 1.9B, and 7B—to advance research on LLM training dynamics. We release checkpoints at regular step intervals— ≈20B tokens (0.5B), ≈40B (1.9B), and ≈160B (7B & 70B) —enabling consistent analysis of training dynamics. Each step’s release is distinguished by its branch name. We’re also sharing the 0.5B and 1.9B runs—originally produced for system bring-up but now available as valuable artifacts for analyzing training behavior at smaller scales.

You can browse all intermediate checkpoints here:

Feel free to check out the full Tri-series collection here:

Dive into the full details—including training configuration and loss curves —on our blog.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

INTERMEDIATE_STEP = "0000020000"
model = AutoModelForCausalLM.from_pretrained('trillionlabs/Tri-70B-Intermediate-Checkpoints', revision=INTERMEDIATE_STEP, trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained('trillionlabs/Tri-70B-Intermediate-Checkpoints', revision=INTERMEDIATE_STEP, trust_remote_code=True)

...
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including trillionlabs/Tri-70B-Intermediate-Checkpoints