Astral-1

Powerd by Upstage's SOLAR 10.7B

Model Description

Astral-1 is a 10-billion parameter model fine-tuned by Odyssey Labs. Built on the foundation of SOLAR-10.7B-Instruct-v1.0, Astral-1 is designed to excel in instruction-following tasks and demonstrates significant improvements in natural language understanding and generation.

Astral-1 is suitable for a wide variety of applications, including:

  • Question answering
  • Text summarisation
  • Dialogue systems
  • Creative writing

The model is fine-tuned with a focus on enhancing usability for non-commercial research and educational purposes.


Intended Use

  • General text generation
  • Instruction following
  • Single-turn conversations
  • Analysis and reasoning tasks

Training Details

  • Built upon: SOLAR's 10.7B instruction-tuned model
  • Inherits: SOLAR's robust training methodology
  • Maintains: Original model's performance characteristics

Technical Specifications

  • Context Length: 4096 tokens
  • Architecture Details:
    • Uses SOLAR's Depth Up-Scaling (DUS)
    • Optimized for efficiency and performance
  • Input Format: Supports both plain text and structured prompts
  • Output Format: Generated text following input instructions

Hardware Requirements

Specification Details
Minimum GPU Memory 12GB
Recommended GPU Memory 16GB for optimal performance
Quantization Options Support varies by implementation

Generation Parameters

  • Recommended to follow base SOLAR model parameters
  • Supports standard transformer generation configurations

Usage

To use Astral-1, follow the steps below:

Installation

Ensure you have the transformers library installed:

pip install transformers

Code Example

from transformers import AutoTokenizer, AutoModelForCausalLM

# Load the model and tokenizer
model_name = "odyssey-labs/Astral-1-10B"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Define your input prompt
prompt = "Explain the significance of artificial intelligence in healthcare."
inputs = tokenizer(prompt, return_tensors="pt")

# Generate a response
outputs = model.generate(**inputs, max_new_tokens=150)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

Limitations

While Astral-1 is a powerful model, it has limitations:

  • It may generate incorrect or biased outputs.
  • Not suitable for real-time or safety-critical applications.
  • Commercial use is prohibited under the Astral Research License (ARL-1.0).

Users are encouraged to validate model outputs for their specific use cases.


License

Astral-1 is released under the Astral Research License (ARL-1.0). This license is based on the CC-BY-NC 4.0 License, as the model is derived from SOLAR-10.7B-Instruct-v1.0. Key terms include:

  • Attribution to both Odyssey Labs and Upstage.
  • Non-commercial use only.
  • Distribution and derivatives must comply with ARL-1.0.

For full details, please review the license terms.


Acknowledgments

We would like to extend our gratitude to the Upstage AI team for providing the exceptional SOLAR base model, which made this work possible. Special thanks to the AI research community for their valuable contributions.

Citations

If you use Astral-1 in your research, please cite:

@misc{astral-1,
  author = {Odyssey Labs},
  title = {Astral-1: A step forward in GenAI research},
  year = {2025},
  howpublished = {\url{https://huggingface.co/odyssey-labs/Astral-1-10B}},
  note = {Based on SOLAR-10.7B-Instruct-v1.0 by Upstage.}
}
@misc{kim2023solar,
      title={SOLAR 10.7B: Scaling Large Language Models with Simple yet Effective Depth Up-Scaling}, 
      author={Dahyun Kim and Chanjun Park and Sanghoon Kim and Wonsung Lee and Wonho Song and Yunsu Kim and Hyeonwoo Kim and Yungi Kim and Hyeonju Lee and Jihoo Kim and Changbae Ahn and Seonghoon Yang and Sukyung Lee and Hyunbyung Park and Gyoungjin Gim and Mikyoung Cha and Hwalsuk Lee and Sunghun Kim},
      year={2023},
      eprint={2312.15166},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
@misc{kim2024sdpo,
      title={sDPO: Don't Use Your Data All at Once}, 
      author={Dahyun Kim and Yungi Kim and Wonho Song and Hyeonwoo Kim and Yunsu Kim and Sanghoon Kim and Chanjun Park},
      year={2024},
      eprint={2403.19270},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

Contact

For questions, feedback, or licensing inquiries, please contact Odyssey Labs:

Downloads last month
22
Safetensors
Model size
10.7B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for odyssey-labs/Astral-1-10B

Finetuned
(8)
this model

Collection including odyssey-labs/Astral-1-10B