- 👨🦰 I’m currently a Master of Science candidate of Peking University (PKU).
- 👦 Before that, I received the Honours Bachelor, Huazhong University of Science and Technology (HUST).
- ❤️🔥 Now, I am intersted in Multi-modal Learning especially MLLM.
- 💥 In 2023 summer, I take part in OSPP(Open Source Promotion Plan) Summer Camp
, with the honor of contributing for MMPretrain to build prompt-based classifier.
- 💥 2023.11 - 2024.5:
MPP-Qwen-Next
is released! All training is conducted on 3090/4090 GPUs. To prevent poverty (24GB of VRAM) from limiting imagination, I implemented an MLLM version based on deepspeed Pipeline Parallel. The Repo supports {video/image/multi-image} {single/multi-turn} conversations. Let's have a try! - 💥 2024.9: We release
ChartMoE
, a multimodal large language model with Mixture-of-Expert connector, for advanced chart 1)understanding, 2)replot, 3)editing, 4)highlighting and 5)transformation. - 💥💥💥 2024.10: I am really fortunate to be involved in the development of
Aria
.Aria
is a Naive Multimodal MoE model, with best-in-class performance across multimodal, language, and coding tasks! - 🎉🎉🎉 2025.1:
ChartMoE
is accepted by ICLR2025! - 🎉🎉🎉 2025.2:
ChartMoE
is selected as ICLR2025 Oral(1.8%)!