Introducing our first standalone model ā FluentlyLM Prinum
Introducing the first standalone model from Project Fluently LM! We worked on it for several months, used different approaches and eventually found the optimal one.
General characteristics: - Model type: Causal language models (QwenForCausalLM, LM Transformer) - Number of parameters: 32.5B - Number of parameters (not embedded): 31.0B - Number of layers: 64 - Context: 131,072 tokens - Language(s) (NLP): English, French, Spanish, Russian, Chinese, Japanese, Persian (officially supported) - License: MIT
Creation strategy: The basis of the strategy is shown in Pic. 2. We used Axolotl & Unsloth for SFT-finetuning with PEFT LoRA (rank=64, alpha=64) and Mergekit for SLERP and TIES mergers.