Quinn777's picture
Update README.md
07296ff verified
---
datasets:
- Quinn777/AMATH-SFT
base_model:
- liuhaotian/llava-v1.5-7b
---
# Model Card for AtomThink-LlamaV
The model is post-trained based on llava-v1.5-7b and the AtomThink framework, and can be used to solve complex multimodal mathematical problems.
# Citation
If you use this dataset in your research, please cite:
```text
@article{xiang2025can,
title={Can Atomic Step Decomposition Enhance the Self-structured Reasoning of Multimodal Large Models?},
author={Xiang, Kun and Liu, Zhili and Jiang, Zihao and Nie, Yunshuang and Cai, Kaixin and Yin, Yiyang and Huang, Runhui and Fan, Haoxiang and Li, Hanhui and Huang, Weiran and others},
journal={arXiv preprint arXiv:2503.06252},
year={2025}
}
@article{xiang2024atomthink,
title={AtomThink: A Slow Thinking Framework for Multimodal Mathematical Reasoning},
author={Xiang, Kun and Liu, Zhili and Jiang, Zihao and Nie, Yunshuang and Huang, Runhui and Fan, Haoxiang and Li, Hanhui and Huang, Weiran and Zeng, Yihan and Han, Jianhua and others},
journal={arXiv preprint arXiv:2411.11930},
year={2024}
}
```
# License
The checkpoint is released under the Apache 2.0 license. Please ensure proper attribution when using this checkpoint.