ARM—Adaptive Reasoning Model, a reasoning model capable of adaptively selecting appropriate reasoning formats based on the task at hand.
- 2025/05/27: Thrilled to release ARM: A reasoning model capable of adaptively selecting reasoning formats based on the task, achieving a better trade-off between effectiveness and efficiency!
You can download our dataset and model from 🤗HuggingFace.
This repository contains the codebase for SFT and RL based on LLaMA-Factory and VeRL. We use two separate conda environments for each stage:
# SFT
conda env create -f environment/llama_factory_env.yaml
conda activate arm_llama_factory
# RL
conda env create -f environment/verl_env.yaml
conda activate arm_verl
pip3 install --force-reinstall torch==2.4.0 --index-url https://download.pytorch.org/whl/cu124
pip3 install flash-attn --no-build-isolationconda activate arm_llama_factory
cd LLaMA-FactoryMake sure to specify the correct model path in the .yaml file.
CUDA_VISIBLE_DEVICES=0,1,2,3 llamafactory-cli train stage1_scripts/qwen2.5_7b/train.yamlllamafactory-cli export stage1_scripts/qwen2.5_7b/merge.yamlconda activate arm_verl
cd verlMake sure to specify the correct model and data path in the .sh file.
# The training data is located in arm/verl/data/parquet.
# Alternatively, you can prepare your own training data, e.g.:
python3 stage2_scripts/data_preprocess/gsm8k.py
# You can also prepare data for the instruction-guided mode used in evaluation, e.g.:
python3 stage2_scripts/data_preprocess/instruction_guided/gsm8k.pybash stage2_scripts/trainer/run.sh# Adaptive Mode
bash stage2_scripts/generation/adaptive_run.sh
# Instruction-Guided Mode. Specify the reasoning format in the .sh file:
bash stage2_scripts/generation/instruction_guided_run.shbash stage2_scripts/evaluation/run.sh[Work in Progress] Stay tuned!
If you have any problems, please contact Siye Wu, Jian Xie.
If our paper or related resources prove valuable to your research, we kindly ask for a citation.
@article{wu2025arm,
title={ARM: Adaptive Reasoning Model},
author={Wu, Siye and Xie, Jian and Zhang, Yikai and Chen, Aili and Zhang, Kai and Su, Yu and Xiao, Yanghua},
journal={arXiv preprint arXiv:2505.20258},
year={2025}
}

