--- language: - en tags: - dataset - ocr - multimodal - vision - image-text-to-text datasets: - custom license: apache-2.0 --- # BLIP3-OCR-200M Dataset ## Overview The **BLIP3-OCR-200M** dataset is designed to address the limitations of current Vision-Language Models (VLMs) in processing and interpreting text-rich images, such as documents and charts. Traditional image-text datasets often struggle to capture nuanced textual information, which is crucial for tasks requiring complex text comprehension and reasoning. Art ### Key Features - **OCR Integration**: The dataset incorporates Optical Character Recognition (OCR) data during the pre-training phase of VLMs. This integration enhances vision-language alignment by providing detailed textual information alongside visual data. - **Text-Rich Content**: The dataset focuses on text-rich images and includes OCR-specific annotations, enabling more accurate handling of content like documents, charts, and other text-heavy visuals. - **Parquet Format**: The dataset is stored in Parquet format, facilitating efficient storage, processing, and retrieval of OCR metadata and images. This format is well-suited for handling large-scale datasets and can be easily integrated into data processing pipelines. ### Purpose: The primary goal of the BLIP3-OCR-200M dataset is to improve the cross-modality reasoning capabilities of VLMs by enriching the pre-training datasets with detailed textual information. This approach significantly enhances the performance of VLMs on tasks involving complex text-rich images, advancing the field of vision-language understanding. ## Key Statistics - **Total Files**: 50 Parquet files. - **Data Format**: Parquet (each file containing OCR-related metadata). - **Sample Count**: Approximately 2 million. - **Purpose**: This dataset is intended for enhancing vision-language model performance on text-rich content. ## Data Structure ### File Structure The dataset is organized into multiple Parquet files, each corresponding to a different folder from the original dataset. Each Parquet file contains flattened and cleaned data for efficient storage and retrieval. ```plaintext ./ ├── 001.parquet ├── 002.parquet ├── 003.parquet └── ... (up to 050.parquet) ``` ### Data Schema Each Parquet file contains a tabular structure with the following columns: - **`uid`**: Unique identifier for each image. - **`url`**: URL where the image was originally obtained (if available). - **`key`**: Unique key associated with each sample. - **`face_bboxes`**: Flattened JSON string containing face bounding box coordinates (if available). - **`captions`**: A flattened JSON string that contains OCR-based captions. Each caption object includes the following key entries: - **text**: The OCR-extracted text from the image. - **granularity**: An integer indicating the level of detail in the OCR annotation. We have 12 different granularities in total described later. - **include_datacomp_raw_cap**: A boolean flag indicating whether the original raw caption from the DataComp dataset is included. - **`metadata`**: A flattened JSON string that holds intermediate metadata related to the OCR process, with the following key entries: - **length**: The total count of the OCR-extracted text. - **entries**: An array of objects, each representing a token with its corresponding bounding box (`bbox`), OCR text (`text`), and confidence score (`confidence`). - **uid**: A unique identifier for the OCR data associated with each image. - **ocr_num_token_larger_than_confidence_threshold**: The number of OCR tokens that exceed a specified confidence threshold(0.9 in this case). ### Downloading the Original DataComp Images If you want to download the original images for each sample based on the `url` entry in the dataset, you can use the [`img2dataset`](https://github.com/rom1504/img2dataset) tool. This tool efficiently downloads images from URLs and stores them in a specified format. You can refer to the script provided by the DataComp project for downloading images. The script is available [here](https://github.com/mlfoundations/datacomp/blob/main/download_upstream.py). ### Example of Loading and Processing the Data You can simply access the data by: ```python from datasets import load_dataset datasets = load_dataset("Salesforce/blip3-ocr-200m") ``` Alternatively, you can load the Parquet files directly using Pandas: ```python # download the Salesforce/blip3-ocr-200m datasets import os os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1" from huggingface_hub import HfApi, logging # Specify the custom download directory custom_download_dir = "./blip3-ocr-200m" hf = HfApi() hf.snapshot_download("Salesforce/blip3-ocr-200m", repo_type="dataset", local_dir=custom_download_dir, local_dir_use_symlinks=False) import pandas as pd # Load a Parquet file into a DataFrame df = pd.read_parquet("./blip3-ocr-200m/001.parquet") # Display the first few rows print(df.head()) # Count the number of rows in the DataFrame print(f"Number of samples: {len(df)}") # Optionally, parse JSON strings if necessary df['captions'] = df['captions'].apply(json.loads) df['metadata'] = df['metadata'].apply(json.loads) # Access specific fields print(df[['uid', 'url', 'captions']].head()) ``` ### Sample JSON Metadata Below is an example of what the JSON metadata might look like for each image: ```json { "uid": "ff2dcbbdc58426e1dcebac49786b6660", "url": "https://img.ev01.to/xxrz/250x400/183/bc/57/bc5793c416cd6d457410be3a39ebb717/bc5793c416cd6d457410be3a39ebb717.jpg", "key": "000000001133", "face_bboxes": [], "captions": [ { "text": "The image contains the text \"THE\", the text \"PATRiCK\", the text \"STAR\", the text \"SHOW\"", "granularity": 0, "include_datacomp_raw_cap": false }, { "text": "The image contains the text \"THE\" located above the center, the text \"PATRiCK\" located above the center, the text \"STAR\" located above the center, the text \"SHOW\" located above the center", "granularity": 1, "include_datacomp_raw_cap": false }, { "text": "The image contains the text \"THE\" located at 1/14 of the image from top to bottom and 1/2 of the image from left to right, the text \"PATRiCK\" located at 1/7 of the image from top to bottom and 10/19 of the image from left to right, the text \"STAR\" located at 1/4 of the image from top to bottom and 10/19 of the image from left to right, the text \"SHOW\" located at 6/19 of the image from top to bottom and 1/2 of the image from left to right", "granularity": 2, "include_datacomp_raw_cap": false }, { "text": "The image contains the text \"THE\" approximately starts at [0.4, 0.0] and extends upto [0.6, 0.1], the text \"PATRiCK\" approximately starts at [0.2, 0.1] and extends upto [0.8, 0.2], the text \"STAR\" approximately starts at [0.3, 0.2] and extends upto [0.7, 0.3], the text \"SHOW\" approximately starts at [0.4, 0.3] and extends upto [0.6, 0.4]", "granularity": 3, "include_datacomp_raw_cap": false }, { "text": "The image contains the text \"THE\" starts at [0.42, 0.045] and extends upto [0.585, 0.101], the text \"PATRiCK\" starts at [0.237, 0.089] and extends upto [0.799, 0.202], the text \"STAR\" starts at [0.308, 0.182] and extends upto [0.728, 0.31], the text \"SHOW\" starts at [0.371, 0.289] and extends upto [0.647, 0.357]", "granularity": 4, "include_datacomp_raw_cap": false }, { "text": "The image contains the \"THE[0.42, 0.045][0.585, 0.101], the \"PATRiCK[0.237, 0.089][0.799, 0.202], the \"STAR[0.308, 0.182][0.728, 0.31], the \"SHOW[0.371, 0.289][0.647, 0.357]", "granularity": 5, "include_datacomp_raw_cap": false }, { "text": "The Patrick Star Show. The image contains the text \"THE\", the text \"PATRiCK\", the text \"STAR\", the text \"SHOW\"", "granularity": 0, "include_datacomp_raw_cap": true }, { "text": "The Patrick Star Show. The image contains the text \"THE\" located above the center, the text \"PATRiCK\" located above the center, the text \"STAR\" located above the center, the text \"SHOW\" located above the center", "granularity": 1, "include_datacomp_raw_cap": true }, { "text": "The Patrick Star Show. The image contains the text \"THE\" located at 1/14 of the image from top to bottom and 1/2 of the image from left to right, the text \"PATRiCK\" located at 1/7 of the image from top to bottom and 10/19 of the image from left to right, the text \"STAR\" located at 1/4 of the image from top to bottom and 10/19 of the image from left to right, the text \"SHOW\" located at 6/19 of the image from top to bottom and 1/2 of the image from left to right", "granularity": 2, "include_datacomp_raw_cap": true }, { "text": "The Patrick Star Show. The image contains the text \"THE\" approximately starts at [0.4, 0.0] and extends upto [0.6, 0.1], the text \"PATRiCK\" approximately starts at [0.2, 0.1] and extends upto [0.8, 0.2], the text \"STAR\" approximately starts at [0.3, 0.2] and extends upto [0.7, 0.3], the text \"SHOW\" approximately starts at [0.4, 0.3] and extends upto [0.6, 0.4]", "granularity": 3, "include_datacomp_raw_cap": true }, { "text": "The Patrick Star Show. The image contains the text \"THE\" starts at [0.42, 0.045] and extends upto [0.585, 0.101], the text \"PATRiCK\" starts at [0.237, 0.089] and extends upto [0.799, 0.202], the text \"STAR\" starts at [0.308, 0.182] and extends upto [0.728, 0.31], the text \"SHOW\" starts at [0.371, 0.289] and extends upto [0.647, 0.357]", "granularity": 4, "include_datacomp_raw_cap": true }, { "text": "The Patrick Star Show. The image contains the \"THE[0.42, 0.045][0.585, 0.101], the \"PATRiCK[0.237, 0.089][0.799, 0.202], the \"STAR[0.308, 0.182][0.728, 0.31], the \"SHOW[0.371, 0.289][0.647, 0.357]", "granularity": 5, "include_datacomp_raw_cap": true } ], "metadata": { "length": 4, "entries": [ { "bbox": [ [0.41964285714285715, 0.044642857142857144], [0.5848214285714286, 0.044642857142857144], [0.5848214285714286, 0.10119047619047619], [0.41964285714285715, 0.10119047619047619] ], "text": "THE", "confidence": 0.9961821436882019 }, { "bbox": [ [0.23660714285714285, 0.08928571428571429], [0.7991071428571429, 0.08928571428571429], [0.7991071428571429, 0.20238095238095238], [0.23660714285714285, 0.20238095238095238] ], "text": "PATRiCK", "confidence": 0.9370164275169373 }, { "bbox": [ [0.3080357142857143, 0.18154761904761904], [0.7366071428571429, 0.19940476190476192], [0.7276785714285714, 0.30952380952380953], [0.29910714285714285, 0.29464285714285715] ], "text": "STAR", "confidence": 0.9967861771583557 }, { "bbox": [ [0.3705357142857143, 0.28869047619047616], [0.6473214285714286, 0.28869047619047616], [0.6473214285714286, 0.35714285714285715], [0.3705357142857143, 0.35714285714285715] ], "text": "SHOW", "confidence": 0.9861563444137573 } ], "uid": "datacomp_ocr_ff2dcbbdc58426e1dcebac49786b6660", "ocr_num_token_larger_than_confidence_threshold": 4 } } ``` The corresponding image with its OCR-related caption is shown below, Art ## OCR Annotation Levels The BLIP3-OCR-200M dataset provides OCR annotations at 12 different levels of granularity, leveraging the off-the-shelf OCR engine [PaddleOCR](https://github.com/PaddlePaddle/PaddleOCR). The bounding boxes for the detected text are represented by the four corners of the text box, as detailed in [PaddleOCR’s documentation](https://github.com/PaddlePaddle/PaddleOCR/blob/e9fca96d4f66f59d620470143d6d666d58ef5b81/docs/datasets/ocr_datasets.en.md). These levels are designed to capture varying degrees of detail in the text content extracted from images, allowing researchers to explore how different levels of textual granularity and the inclusion of the original raw caption affect the performance of Vision-Language Models (VLMs). ### Levels 1-6: Different OCR Granularities The first six levels of annotation focus solely on the OCR text extracted from the images, providing varying levels of detail: 1. **Level 1 (Granularity 0)**: - **Basic Text Extraction**: This level provides a straightforward extraction of the main text present in the image. It focuses on the most prominent textual content without specifying its location or context within the image. - **Use Case**: Useful for tasks where only the presence of specific text is important, regardless of its position or structure. 2. **Level 2 (Granularity 1)**: - **Text with General Location**: This level adds a general indication of where the text is located within the image, such as "top-right corner" or "bottom-left corner." - **Use Case**: Beneficial for applications where understanding the approximate position of text within the image adds context to its meaning. 3. **Level 3 (Granularity 2)**: - **Text with Detailed Positioning**: At this level, the text is annotated with more precise positional information, describing specific areas within the image (e.g., "7/10 from top to bottom, 2/5 from left to right"). - **Use Case**: Ideal for scenarios where the exact location of text within the image plays a crucial role, such as in document layout analysis. 4. **Level 4 (Granularity 3)**: - **Text with Bounding Box Coordinates**: This level provides bounding box coordinates that outline the area where the text is located, offering a more spatially precise annotation. - **Use Case**: Useful for tasks involving text region detection, where knowing the exact spatial boundaries of text is necessary. 5. **Level 5 (Granularity 4)**: - **Text with Precise Bounding Boxes**: Similar to Level 4 but with even more precise bounding box information, capturing subtle details of text placement within the image. - **Use Case**: Suitable for fine-grained text detection tasks, such as identifying specific words or characters within a tightly constrained area. 6. **Level 6 (Granularity 5)**: - **Text with OCR Tags and Bounding Boxes**: This level includes OCR tags combined with precise bounding box coordinates, marking both the text and its spatial characteristics within the image. - **Use Case**: Best for detailed image-text alignment tasks where both the text content and its exact spatial metadata are required for accurate interpretation. ### Levels 7-12: Different OCR Granularities with Raw Captions The next six levels combine the OCR annotations described above with the original raw captions from the dataset. This allows researchers to explore how the inclusion of the raw caption, alongside the OCR data, impacts model performance. 7. **Level 7 (Granularity 0 + Raw Caption)**: - **Basic Text Extraction with Raw Caption**: Combines the basic OCR text with the original raw caption extracted from the image. - **Use Case**: Enables comparison between raw text content and the basic OCR output, providing insights into how models handle raw versus processed text. 8. **Level 8 (Granularity 1 + Raw Caption)**: - **Text with General Location and Raw Caption**: This level integrates the OCR text with general location information and the raw caption. - **Use Case**: Useful for understanding how location context and raw text influence model interpretations and performance. 9. **Level 9 (Granularity 2 + Raw Caption)**: - **Text with Detailed Positioning and Raw Caption**: Provides detailed positional information along with the raw caption, combining structured OCR data with the unprocessed text. - **Use Case**: Ideal for evaluating models that require both detailed text placement and raw text content for tasks like document processing. 10. **Level 10 (Granularity 3 + Raw Caption)**: - **Text with Bounding Box Coordinates and Raw Caption**: Merges the bounding box annotations with the raw caption, offering spatial context alongside the original text. - **Use Case**: Suitable for tasks where spatial data and raw text content are both critical, such as in multi-modal information retrieval. 11. **Level 11 (Granularity 4 + Raw Caption)**: - **Text with Precise Bounding Boxes and Raw Caption**: Enhances the precise bounding box annotations with the raw caption, providing detailed spatial-textual information. - **Use Case**: Best for high-precision tasks where exact text boundaries and raw text are both important, such as in OCR refinement. 12. **Level 12 (Granularity 5 + Raw Caption)**: - **Text with OCR Tags, Bounding Boxes, and Raw Caption**: The most detailed level, combining OCR tags, precise bounding boxes, and the raw caption. - **Use Case**: Ideal for comprehensive multi-modal tasks that require complete text and spatial context, integrating all available data for model training or evaluation. ### OCR Caption Data Schema This includes up to 12 different levels of granularity, with the following structure: ```plaintext captions: ├── 'caption_1' │ ├── 'text': 'image contains text X, text Y and text Z' │ ├── 'granularity': 0 │ ├── 'include_datacomp_raw_cap': False ├── 'caption_2' │ ├── 'text': 'image contains text X located at the top-right corner, text Y ...' │ ├── 'granularity': 1 │ ├── 'include_datacomp_raw_cap': False ├── 'caption_3' │ ├── 'text': 'image contains text X located 7/10 from the top to bottom and 2/5 from left to right ...' │ ├── 'granularity': 2 │ ├── 'include_datacomp_raw_cap': False ├── ... ├── 'caption_7' │ ├── 'text': + 'image contains text X, text Y and text Z' │ ├── 'granularity': 0 │ ├── 'include_datacomp_raw_cap': True ├── ... ├── 'caption_12' │ ├── 'text': + 'image contains X[0.113,0.222],[0.112,0.411]' │ ├── 'granularity': 5 │ ├── 'include_datacomp_raw_cap': True ``` In the previous example, the `` is 'The Patrick Star Show.' ### Exploring the Annotations Researchers are encouraged to explore the performance of Vision-Language Models (VLMs) across these different levels of granularity. By experimenting with various combinations of OCR detail and the inclusion of raw captions, it is possible to gain insights into how different types of textual information influence model learning and decision-making processes. This structured approach allows for fine-tuned analysis of VLMs' capabilities in handling text-rich images, providing a valuable resource for advancing research in the field of multimodal machine learning. ## License We release BLIP3-OCR-200M under an Apache2.0 license, designating it primarily as a research artifact. . This dataset is being released for research purposes only. This repository includes the extracted original text in the underlying images. It is the responsibility of the user to check and/or obtain the proper copyrights to use any of the images of the original dataset. ## Citation ```bibtex @article{blip3-xgenmm, author = {Le Xue, Manli Shu, Anas Awadalla, Jun Wang, An Yan, Senthil Purushwalkam, Honglu Zhou, Viraj Prabhu, Yutong Dai, Michael S Ryoo, Shrikant Kendre, Jieyu Zhang, Can Qin, Shu Zhang, Chia-Chih Chen, Ning Yu, Juntao Tan, Tulika Manoj Awalgaonkar, Shelby Heinecke, Huan Wang, Yejin Choi, Ludwig Schmidt, Zeyuan Chen, Silvio Savarese, Juan Carlos Niebles, Caiming Xiong, Ran Xu}, title = {xGen-MM (BLIP-3): A Family of Open Large Multimodal Models}, journal = {arXiv preprint}, month = {August}, year = {2024}, eprint = {2408.08872}, archivePrefix = {arXiv}, primaryClass = {cs.CV}, url = {https://arxiv.org/abs/2408.08872}, } ```