qwen-4b-hyperswitch-v1

qwen

Model Details

  • Base Model: Qwen/Qwen3-4B
  • Training Data: Hyperswitch repository code
  • Final Training Loss: 0.4966
  • Training Progress: 92% complete (excellent results)

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model = AutoModelForCausalLM.from_pretrained(
    "archit11/qwen-4b-hyperswitch-v1",
    torch_dtype=torch.float16,
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("archit11/qwen-4b-hyperswitch-v1")

# Generate code
prompt = "use crate::"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=150, temperature=0.3)
result = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(result)

Training Details

This model was fine-tuned on the Hyperswitch codebase to improve code completion and generation for Rust payment processing systems.

Downloads last month
56
Safetensors
Model size
4.02B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for archit11/qwen-4b-hyperswitch-v1

Base model

Qwen/Qwen3-4B-Base
Finetuned
Qwen/Qwen3-4B
Finetuned
(278)
this model