Safetensors
English
olmo2

You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Model Card for traclm-v5-13b-instruct

An Army domain finetune of allenai/OLMo-2-1124-13B created by finetuning on a merger of domain-specific, general-purpose instruction-following, and tool-calling datasets.

Model Details

Model Description

This model is a research project aimed at exploring whether a pretrained LLM can acquire and express tangible domain-specific knowledge about the Army domain.

  • Developed by: The Research and Analysis Center, Army Futures Command, U.S. Army
  • License: Apache 2.0
  • Model Type: Olmo2ForCausalLM
  • Finetuned from model: allenai/OLMo-2-1124-13B

Available Quantizations (for running on low-resource hardware):

  • TBP.

Downstream Use

This model is instruction-tuned, and is thus more capable of following user instructions than its corresponding base version. Furthermore, this is the first traclm model finetuned for tool-calling, thus enabling enhanced usage in agentic workflows and frameworks. However, this model is still capable of significant hallucination, so all outputs should be verified by end users.

Out-of-Scope Use

The creation of this model constitutes academic research in partnership with the Naval Postgraduate School. The purpose of this research is to inform future DoD experimentation regarding the development and application of domain-specific large language models. Experiments involving direct application of this model to downstream military tasks are encouraged, but extreme caution should be exercised before productionalization.

Prompt Format

This model was fine-tuned with a modified chat template adapted from the Qwen3 chat template. This chat template was adapted for its modernity and handling of tool-calling tokens. It is highly recommended that you use the same template for any interactions with the model. Failure to do so will significantly degrade performance.

The traclm-v5 chat template can easily be applied to text you plan to process with the model using the chat_template included in the tokenizer. Read here for additional information. Note that incoming text must be in the following common format:

messages = [{"role": "system", "content": <content>}, # system message optional
            {"role": "user", "content": "What is the capital of France?"},
            {"role": "tools", "content": <tools>}, # tools optional
]

Training Details

Training Data

This model was trained on a shuffled & filtered merger of the following datasets:

Training Procedure

The model was trained using HuggingFace's TRL framework in combination with HuggingFace's Accelerate and Microsoft's DeepSpeed framework for model/data parallelism.

Training Hardware

Training was conducted on a single compute node 4x NVIDIA A100 GPUs.

Model Card Contact

MAJ Daniel C. Ruiz ([email protected])

Downloads last month
1
Safetensors
Model size
13.7B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for TRAC-MTRY/traclm-v5-13b-instruct

Finetuned
(3)
this model

Datasets used to train TRAC-MTRY/traclm-v5-13b-instruct

Collection including TRAC-MTRY/traclm-v5-13b-instruct