YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

library_name: transformers

license: apache-2.0

base_model: Qwen/Qwen2.5-0.5B-Instruct

tags:

  • generated_from_trainer

language:

  • zho

  • eng

  • fra

  • spa

  • por

  • deu

  • ita

  • rus

  • jpn

  • kor

  • vie

  • tha

  • ara

model-index:

  • name: careerbot_PG6_Qwen_Qwen2.5-0.5B-Instruct_model_small_2

    results: []


careerbot_PG6_Qwen_Qwen2.5-0.5B-Instruct_model_small_2

This model is a fine-tuned version of Qwen/Qwen2.5-0.5B-Instruct on the None dataset.

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05

  • train_batch_size: 16

  • eval_batch_size: 16

  • seed: 42

  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments

  • lr_scheduler_type: linear

  • lr_scheduler_warmup_steps: 100

  • num_epochs: 1

  • mixed_precision_training: Native AMP

Training results

| Training Loss | Epoch | Step | Validation Loss |

|:-------------:|:-----:|:----:|:---------------:|

| No log | 1.0 | 225 | 1.0349 |

Framework versions

  • Transformers 4.46.1

  • Pytorch 2.5.0+cu124

  • Datasets 2.19.0

  • Tokenizers 0.20.1

Downloads last month
24
Safetensors
Model size
494M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support