Models

Open-source models from AMD GenAI on Hugging Face

SAND-Math-Qwen2.5-32B

State-of-the-art 32B reasoning model fine-tuned on 14K synthetic math samples, outperforming models trained on 5-50x larger datasets

reasoning math 32B
SAND-MathScience-DeepSeek-Qwen32B

32B reasoning model achieving 78.33% on AIME25, surpassing Qwen3-32B with only 27K synthetic training samples

reasoning math science 32B
Instella-3B-Math-SFT

SFT version of Instella-3B optimized for mathematical reasoning tasks

math 3B SFT
Instella-3B-Math

AMD's first fully open reasoning model trained with long chain-of-thought RL on MI300X GPUs

reasoning math 3B RL
Instella-3B-Long-Instruct

Instruction-tuned 3B model supporting 128K context length for long-context tasks

long-context 3B 128K instruct
Instella-3B-Instruct

Instruction-tuned version of Instella-3B for chat and instruction following

language-model 3B instruct
Instella-3B-SFT

Supervised fine-tuned version of Instella-3B base model

language-model 3B SFT
Instella-3B

Fully open 3B language model trained from scratch on MI300X GPUs

language-model 3B
Instella-3B-Stage1

Stage 1 pre-training checkpoint of Instella-3B model

language-model 3B pretrain
AMD-OLMo-1B-SFT-DPO

DPO-aligned version of AMD-OLMo-1B for improved helpfulness and safety

language-model 1B DPO
AMD-OLMo-1B-SFT

Supervised fine-tuned version of AMD-OLMo-1B

language-model 1B SFT
AMD-OLMo-1B

Fully open 1B language model trained on MI250 GPUs with 1.3T tokens

language-model 1B