Fine-tuned Qwen3-14B Model

Model Description

This is a fine-tuned version of the Qwen3-14B model optimized with Unsloth and 4-bit quantization for efficient inference. The model was fine-tuned on a curated historical text corpus focused on early 20th-century diplomatic events and treaties, enabling nuanced understanding and generation of complex political and historical narratives.

Key Features

  • 2x faster training using Unsloth optimization

  • 4-bit quantization for reduced memory footprint and efficient inference

  • Enhanced understanding of historical and diplomatic discourse

  • Compatible with Hugging Face’s TRL library

  • Supports LoRA fine-tuning for flexible adaptation

Model Details

  • Base Model: unsloth/qwen3-14b-unsloth-bnb-4bit

  • License: Apache 2.0

  • Fine-tuned by: momererkoc

  • Languages: Primarily Turkish, with English capabilities

  • Quantization: 4-bit via BitsAndBytes

Training Data

  • Extensive historical text collection focusing on early 20th-century diplomacy and treaty analysis

  • Texts rich in complex political context and multi-layered narrative

Usage Example

from unsloth import FastLanguageModel
model, tokenizer = FastLanguageModel.from_pretrained(
    model_name = "momererkoc/qwen3-14b-lozan", # YOUR MODEL YOU USED FOR TRAINING
    max_seq_length = 2048,
    load_in_4bit = True,
    token="hf_...",
)

messages = [
    {"role" : "user", "content" : "Lozan anlaşmasının sonuçları nelerdir?"}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize = False,
    add_generation_prompt = True, # Must add for generation
    enable_thinking = False, # Disable thinking
)

from transformers import TextStreamer
_ = model.generate(
    **tokenizer(text, return_tensors = "pt").to("cuda"),
    max_new_tokens = 256, # Increase for longer outputs!
    temperature = 0.7, top_p = 0.8, top_k = 20, # For non thinking
    streamer = TextStreamer(tokenizer, skip_prompt = True),
)

Performance

  • 2x faster fine-tuning compared to standard methods

  • Reduced GPU memory usage due to 4-bit quantization

  • Strong contextual understanding of historical and political texts

  • Supports generation of coherent, context-aware historical narratives

Optimization Details

  • Uses Unsloth for accelerated training

  • 4-bit quantization via BitsAndBytes

  • LoRA for parameter-efficient fine-tuning

Acknowledgments

  • Unsloth team for optimization tools

  • Hugging Face for the transformers ecosystem

  • Dataset creators for curated historical texts

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for momererkoc/qwen3-14b-lozan

Finetuned
Qwen/Qwen3-14B
Finetuned
unsloth/Qwen3-14B
Finetuned
(39)
this model

Collection including momererkoc/qwen3-14b-lozan