Aeoinum v1.1 Base 4B

A state-of-the-art language model for Russian language processing. This checkpoint contains a preliminary version of the model with 4 billion parameters.

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

tokenizer = AutoTokenizer.from_pretrained("aeonium/Aeonium-v1.1-Base-4B")
model = AutoModelForCausalLM.from_pretrained("aeonium/Aeonium-v1.1-Base-4B").cuda()

input_ids = tokenizer("Искусственный интеллект - это", return_tensors='pt').to(model.device)["input_ids"]
output = model.generate(input_ids, max_new_tokens=48, do_sample=True, temperature=0.7)
print(tokenizer.decode(output[0]))

Dataset Detail

The dataset for pre-training is collected from public data, most of which are web pages in Russian.

Training Detail

The training is performed thanks to a grant from TPU Research Cloud on a TPU v4-256 node.

Content Warning

Aeonium v1.1 is a large language model trained on a broad dataset from the internet. As such, it may generate text that contains biases, offensive language, or other disapproving content. The model outputs should not be considered factual or representative of any individual's beliefs or identity. Users should exercise caution and apply careful filtering when using Aeonium's generated text, especially for sensitive or high-stakes applications. The developers do not condone generating harmful, biased, or unethical content.

Copyright

The model is released under the Apache 2.0 license.

Downloads last month
9
Safetensors
Model size
4.04B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for aeonium/Aeonium-v1.1-Base-4B

Finetunes
1 model