Bowen Qu
Bowen Qu Brian

About Me

👋Hi, I’m Bowen(Brian) Qu. My research interests include Vision-Language Model and MLLM Reasoning. The logo of this website is my lovely cat - Baka (巴卡 in Chinese)!

Interests
  • Vision-Language Model(VLM)
  • MLLM Reasoning
  • Al-Generated Image/Video Quality Assesment
Education
  • Master of Science

    Peking University (PKU)

  • Bachelor of Engineering

    Huazhong University of Science and Technology (HUST)

🔥 News
  • 2025.02: 🎉🎉🎉 ChartMoE is selected as ICLR2025 Oral(1.8%)!
  • 2025.01: 🎉🎉 ChartMoE is accepted by ICLR2025!
  • 2024.10: 🎉🎉 We release Aria, a native LMM that excels on text, code, image, video, PDF and more!
  • 2024.09: 💥💥 We release ChartMoE, a MLLM with MoE connector, for advanced chart 1️⃣understanding, 2️⃣replot, 3️⃣editing, 4️⃣highlighting and 5️⃣transformation.
  • 2023.12: 💥 MPP-Qwen-Next is released! Prevent poverty (24GB of VRAM) from limiting imagination. All 7B/14B llava-like training is conducted on RTX3090 GPUs by Pipeline Parallel.
Selected Outputs

🌟 is me. * Equal Contribution (i.e.: Co-First Author). 📧 Corresponding Author.

Core-Authored Publications

🌟 is me. * Equal Contribution (i.e.: Co-First Author). 📧 Corresponding Author.

(2025). ChartMoE: Mixture of Diversely Aligned Expert Connector for Chart Understanding. ICLR2025 Oral (1.8%).
(2024). Aria: An Open Multimodal Native Mixture-of-Experts Model. Technical Report.
Selected Projects

😺 I enjoy open-sourcing. Here are a selection of projects that I’ve led or served as the core contributor.