NovaSky-AI/Sky-T1-32B-Preview cover image

NovaSky-AI/Sky-T1-32B-Preview

This is a 32B reasoning model trained from Qwen2.5-32B-Instruct with 17K data. The performance is on par with o1-preview model on both math and coding.

This is a 32B reasoning model trained from Qwen2.5-32B-Instruct with 17K data. The performance is on par with o1-preview model on both math and coding.

Public
$0.12/$0.18 in/out Mtoken
fp16
32,768
ProjectPaperLicense
NovaSky-AI/Sky-T1-32B-Preview cover image

Sky-T1-32B-Preview

Ask me anything

0.00s

Model Description

This is a 32B reasoning model trained from Qwen2.5-32B-Instruct with 17K data. The performance is on par with o1-preview model on both math and coding. Please see our blog post for more details.

  • Developed by: NovaSky Team from Sky Computing Lab at UC Berkeley.

Training Details

Training Data

17K verified correct responses from Qwen/QwQ-32B-Preview on coding, math. In addition, we add the science portion from the Still-2 paper.

Training Procedure

We perform supervised fine tuning on the data, with a batch size of 96.

Speeds

We use Llama-Factory for training. On 8 H100, the training takes 19 hours with DeepSpeed Zero-3 Offload.

Evaluation

Sky-T1-32B-PreviewQwen-2.5-32B-InstructQwQo1-preview
Math50082.476.285.481.4
AIME202443.316.750.040.0
LiveCodeBench-Easy86.384.690.792.9
LiveCodeBench-Medium56.840.856.354.9
LiveCodeBench-Hard17.99.817.116.3
GPQA-Diamond56.845.552.575.2

Acknowledgement

We would like to thanks the compute resources from Lambda Lab and AnyScale. We would like to thanks the academic feedback and support from the Still-2 Team, and Junyang Lin from the Qwen Team.

Citation

Please considering citing our blog post if you found it useful for your research. Thank you!

@misc{sky_t1_2025,
  author       = {NovaSky Team},
  title        = {Sky-T1: Fully open-source reasoning model with o1-preview performance in $450 budget},
  howpublished = {https://novasky-ai.github.io/posts/sky-t1},
  note         = {Accessed: 2025-01-09},
  year         = {2025}
}