A Data Collection and Details about the Tokenizers
–Neural Information Processing Systems
We collected about 30 million text-image pairs from multiple channels, and built a 2.5TB new dataset (after tokenization, the size becomes about 250GB). The sources of data are basically classified into the following categories: (1) Professional image websites (both English and Chinese). The images in the websites are usually with captions. We have already introduced tokenizers in section 2.2, and here are some details. Colored grids are all the tokens attended to by the token marked "O".
Neural Information Processing Systems
Nov-15-2025, 08:43:14 GMT
- Technology: