Monolingual language models for 12 languages, with each model trained exclusively on the data from its respective side of a multilingual
Xiulin Yang
xiulinyang
·
AI & ML interests
Language Modeling, Interpretability, (compositional) generalization, tokenization
Recent Activity
updated
a model
6 days ago
xiulinyang/GPT2_BABYLM10MCR_50000
updated
a model
6 days ago
xiulinyang/GPT2_BABYLM10M_50000
published
a model
6 days ago
xiulinyang/GPT2_BABYLM10MCR_50000
Organizations
None yet