WanJuan: A Comprehensive Multimodal Dataset for Advancing English and Chinese Large Models
He, Conghui, Jin, Zhenjiang, Xu, Chao, Qiu, Jiantao, Wang, Bin, Li, Wei, Yan, Hang, Wang, Jiaqi, Lin, Dahua
–arXiv.org Artificial Intelligence
The rise in popularity of ChatGPT and GPT-4 has significantly accelerated the development of large models, leading to the creation of numerous impressive large language models(LLMs) and multimodal large language models (MLLMs). These cutting-edge models owe their remarkable performance to high-quality data. However, the details of the training data used in leading paradigms are often kept confidential. This lack of transparency, coupled with the scarcity of open-source data, impedes further developments within the community. As a response, this paper presents "Wan Juan", a large-scale multimodal dataset composed of both Chinese and English data, collected from a wide range of web sources. The dataset incorporates text, image-text, and video modalities, with a total volume exceeding 2TB. It was utilized in the training of InternLM, a model that demonstrated significant advantages in multi-dimensional evaluations when compared to models of a similar scale.
arXiv.org Artificial Intelligence
Sep-15-2023