Alert button
Picture for Haoran Wei

Haoran Wei

Alert button

OneChart: Purify the Chart Structural Extraction via One Auxiliary Token

Add code
Bookmark button
Alert button
Apr 15, 2024
Jinyue Chen, Lingyu Kong, Haoran Wei, Chenglong Liu, Zheng Ge, Liang Zhao, Jianjian Sun, Chunrui Han, Xiangyu Zhang

Viaarxiv icon

MegaScale: Scaling Large Language Model Training to More Than 10,000 GPUs

Add code
Bookmark button
Alert button
Feb 23, 2024
Ziheng Jiang, Haibin Lin, Yinmin Zhong, Qi Huang, Yangrui Chen, Zhi Zhang, Yanghua Peng, Xiang Li, Cong Xie, Shibiao Nong, Yulu Jia, Sun He, Hongmin Chen, Zhihao Bai, Qi Hou, Shipeng Yan, Ding Zhou, Yiyao Sheng, Zhuo Jiang, Haohan Xu, Haoran Wei, Zhang Zhang, Pengfei Nie, Leqi Zou, Sida Zhao, Liang Xiang, Zherui Liu, Zhe Li, Xiaoying Jia, Jianxi Ye, Xin Jin, Xin Liu

Viaarxiv icon

Small Language Model Meets with Reinforced Vision Vocabulary

Add code
Bookmark button
Alert button
Jan 23, 2024
Haoran Wei, Lingyu Kong, Jinyue Chen, Liang Zhao, Zheng Ge, En Yu, Jianjian Sun, Chunrui Han, Xiangyu Zhang

Viaarxiv icon

Vary: Scaling up the Vision Vocabulary for Large Vision-Language Models

Add code
Bookmark button
Alert button
Dec 11, 2023
Haoran Wei, Lingyu Kong, Jinyue Chen, Liang Zhao, Zheng Ge, Jinrong Yang, Jianjian Sun, Chunrui Han, Xiangyu Zhang

Viaarxiv icon

Merlin:Empowering Multimodal LLMs with Foresight Minds

Add code
Bookmark button
Alert button
Nov 30, 2023
En Yu, Liang Zhao, Yana Wei, Jinrong Yang, Dongming Wu, Lingyu Kong, Haoran Wei, Tiancai Wang, Zheng Ge, Xiangyu Zhang, Wenbing Tao

Viaarxiv icon

Autoencoder with Group-based Decoder and Multi-task Optimization for Anomalous Sound Detection

Add code
Bookmark button
Alert button
Nov 15, 2023
Yifan Zhou, Dongxing Xu, Haoran Wei, Yanhua Long

Viaarxiv icon

DreamLLM: Synergistic Multimodal Comprehension and Creation

Add code
Bookmark button
Alert button
Sep 20, 2023
Runpei Dong, Chunrui Han, Yuang Peng, Zekun Qi, Zheng Ge, Jinrong Yang, Liang Zhao, Jianjian Sun, Hongyu Zhou, Haoran Wei, Xiangwen Kong, Xiangyu Zhang, Kaisheng Ma, Li Yi

Figure 1 for DreamLLM: Synergistic Multimodal Comprehension and Creation
Figure 2 for DreamLLM: Synergistic Multimodal Comprehension and Creation
Figure 3 for DreamLLM: Synergistic Multimodal Comprehension and Creation
Figure 4 for DreamLLM: Synergistic Multimodal Comprehension and Creation
Viaarxiv icon

ChatSpot: Bootstrapping Multimodal LLMs via Precise Referring Instruction Tuning

Add code
Bookmark button
Alert button
Jul 18, 2023
Liang Zhao, En Yu, Zheng Ge, Jinrong Yang, Haoran Wei, Hongyu Zhou, Jianjian Sun, Yuang Peng, Runpei Dong, Chunrui Han, Xiangyu Zhang

Figure 1 for ChatSpot: Bootstrapping Multimodal LLMs via Precise Referring Instruction Tuning
Figure 2 for ChatSpot: Bootstrapping Multimodal LLMs via Precise Referring Instruction Tuning
Figure 3 for ChatSpot: Bootstrapping Multimodal LLMs via Precise Referring Instruction Tuning
Figure 4 for ChatSpot: Bootstrapping Multimodal LLMs via Precise Referring Instruction Tuning
Viaarxiv icon

PolyLM: An Open Source Polyglot Large Language Model

Add code
Bookmark button
Alert button
Jul 12, 2023
Xiangpeng Wei, Haoran Wei, Huan Lin, Tianhao Li, Pei Zhang, Xingzhang Ren, Mei Li, Yu Wan, Zhiwei Cao, Binbin Xie, Tianxiang Hu, Shangjie Li, Binyuan Hui, Bowen Yu, Dayiheng Liu, Baosong Yang, Fei Huang, Jun Xie

Figure 1 for PolyLM: An Open Source Polyglot Large Language Model
Figure 2 for PolyLM: An Open Source Polyglot Large Language Model
Figure 3 for PolyLM: An Open Source Polyglot Large Language Model
Figure 4 for PolyLM: An Open Source Polyglot Large Language Model
Viaarxiv icon

Multi-pass Training and Cross-information Fusion for Low-resource End-to-end Accented Speech Recognition

Add code
Bookmark button
Alert button
Jun 20, 2023
Xuefei Wang, Yanhua Long, Yijie Li, Haoran Wei

Figure 1 for Multi-pass Training and Cross-information Fusion for Low-resource End-to-end Accented Speech Recognition
Figure 2 for Multi-pass Training and Cross-information Fusion for Low-resource End-to-end Accented Speech Recognition
Figure 3 for Multi-pass Training and Cross-information Fusion for Low-resource End-to-end Accented Speech Recognition
Figure 4 for Multi-pass Training and Cross-information Fusion for Low-resource End-to-end Accented Speech Recognition
Viaarxiv icon