Sarashina2.2 Collection Large Language Models developed by SB Intuitions. Pretrained and instruction-tuned models are available in three sizes: 0.5B, 1B, and 3B. • 6 items • Updated Mar 5 • 6
Asagi-VLM Collection Asagi is a Japanese Vision & Language model, trained on a large-scale synthetic dataset. • 4 items • Updated Feb 24 • 6
TinySwallow Collection Compact Japanese models trained with "TAID: Temporally Adaptive Interpolated Distillation for Efficient Knowledge Transfer in Language Models" • 5 items • Updated Jan 30 • 17
view article Article Navigating Korean LLM Research #2: Evaluation Tools By amphora • Oct 23, 2024 • 8
LLM-jp-3 Fine-tuned Models Collection Fine-tuned models in the LLM-jp-3 model series • 25 items • Updated 30 days ago • 6
LLM-jp-3 Pre-trained Models Collection Pre-trained models in the LLM-jp-3 model series • 10 items • Updated 30 days ago • 6
view article Article How to generate text: using different decoding methods for language generation with Transformers By patrickvonplaten • Mar 1, 2020 • 218
PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency Paper • 2410.07563 • Published Oct 10, 2024 • 2
gemma-2-baku Collection The baku model series are based on the gemma-2 series and have been continually pre-trained on Japanese-specific corpora. • 4 items • Updated Mar 16 • 5
Gemma 2 JPN Release Collection A Gemma 2 2B model fine-tuned on Japanese text. It supports the Japanese language the same level of performance of EN only queries on Gemma 2. • 3 items • Updated 27 days ago • 29
Japanese SimCSE Collection Tsukagoshi et al., Japanese SimCSE Technical Report, arXiv 2023. https://arxiv.org/abs/2310.19349 • 5 items • Updated Apr 18 • 2