Edit model card

Description

This model is a fine-tuned adaptation of Llama-2-7b-chat-hf, specifically tailored for the Javanese language. It has been developed to function as a chatbot, offering fluent and accurate responses in Javanese. The fine-tuning process utilized a dataset translated into Javanese, aiming to improve the representation of the Javanese language within language models and chatbot technologies.

Training

The model was fine-tuned on a dataset translated into Javanese using the NLLB model. This dataset includes texts from both OASST1 and OASST2, covering a wide range of conversational scenarios. The training process employed PEFT and TRL to facilitate efficient and rapid fine-tuning.

Usage

This model is intended for deployment as a conversational chatbot in Javanese. It is suitable for a variety of applications that require natural language understanding in Javanese. Users can interact with the model using the standard Hugging Face Transformers pipeline for text generation.

from peft import PeftModel, PeftConfig
from transformers import AutoModelForCausalLM

config = PeftConfig.from_pretrained("richardcsuwandi/llama2-javanese")
base_model = AutoModelForCausalLM.from_pretrained("NousResearch/Llama-2-7b-chat-hf")
model = PeftModel.from_pretrained(base_model, "richardcsuwandi/llama2-javanese")
Downloads last month
84
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for richardcsuwandi/llama2-javanese

Adapter
(349)
this model

Dataset used to train richardcsuwandi/llama2-javanese

Space using richardcsuwandi/llama2-javanese 1