Datasets:
Tasks:
Image-to-Text
Formats:
parquet
Sub-tasks:
image-captioning
Languages:
English
Size:
1M - 10M
License:
| # coding=utf-8 | |
| # Copyright 2020 HuggingFace Datasets Authors. | |
| # | |
| # Licensed under the Apache License, Version 2.0 (the "License"); | |
| # you may not use this file except in compliance with the License. | |
| # You may obtain a copy of the License at | |
| # | |
| # http://www.apache.org/licenses/LICENSE-2.0 | |
| # | |
| # Unless required by applicable law or agreed to in writing, software | |
| # distributed under the License is distributed on an "AS IS" BASIS, | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
| # See the License for the specific language governing permissions and | |
| # limitations under the License. | |
| # Lint as: python3 | |
| """Conceptual Captions dataset.""" | |
| import csv | |
| import textwrap | |
| import datasets | |
| _DESCRIPTION = """\ | |
| Google's Conceptual Captions dataset has more than 3 million images, paired with natural-language captions. | |
| In contrast with the curated style of the MS-COCO images, Conceptual Captions images and their raw descriptions are harvested from the web, | |
| and therefore represent a wider variety of styles. The raw descriptions are harvested from the Alt-text HTML attribute associated with web images. | |
| The authors developed an automatic pipeline that extracts, filters, and transforms candidate image/caption pairs, with the goal of achieving a balance of cleanliness, | |
| informativeness, fluency, and learnability of the resulting captions. | |
| """ | |
| _HOMEPAGE = "http://data.statmt.org/cc-100/" | |
| _LICENSE = """\ | |
| The dataset may be freely used for any purpose, although acknowledgement of | |
| Google LLC ("Google") as the data source would be appreciated. The dataset is | |
| provided "AS IS" without any warranty, express or implied. Google disclaims all | |
| liability for any damages, direct or indirect, resulting from the use of the | |
| dataset. | |
| """ | |
| _CITATION = """\ | |
| @inproceedings{sharma2018conceptual, | |
| title = {Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset For Automatic Image Captioning}, | |
| author = {Sharma, Piyush and Ding, Nan and Goodman, Sebastian and Soricut, Radu}, | |
| booktitle = {Proceedings of ACL}, | |
| year = {2018}, | |
| } | |
| """ | |
| _URLS = { | |
| "unlabeled": { | |
| "train": "https://storage.googleapis.com/gcc-data/Train/GCC-training.tsv?_ga=2.191230122.-1896153081.1529438250", | |
| "validation": "https://storage.googleapis.com/gcc-data/Validation/GCC-1.1.0-Validation.tsv?_ga=2.141047602.-1896153081.1529438250", | |
| }, | |
| "labeled": { | |
| "train": "https://storage.googleapis.com/conceptual-captions-v1-1-labels/Image_Labels_Subset_Train_GCC-Labels-training.tsv?_ga=2.234395421.-20118413.1607637118", | |
| }, | |
| } | |
| _DESCRIPTIONS = { | |
| "unlabeled": textwrap.dedent( | |
| """\ | |
| The basic version of the dataset split into Training, Validation, and Test splits. | |
| The Training split consists of 3,318,333 image-URL/caption pairs, with a total number of 51,201 total token types in the captions (i.e., total vocabulary). | |
| The average number of tokens per captions is 10.3 (standard deviation of 4.5), while the median is 9.0 tokens per caption. | |
| The Validation split consists of 15,840 image-URL/caption pairs, with similar statistics. | |
| """ | |
| ), | |
| "labeled": textwrap.dedent( | |
| """\ | |
| A subset of 2,007,090 image-URL/caption pairs from the training set with machine-generated image labels. | |
| The image labels are obtained using the Google Cloud Vision API. | |
| Each image label has a machine-generated identifier (MID) corresponding to the label's Google Knowledge Graph entry and a confidence score for its presence in the image. | |
| Note: 2,007,528 is the number of image-URL/caption pairs specified by the authors, but some rows are missing labels, so they are not included. | |
| """ | |
| ), | |
| } | |
| class ConceptualCaptions(datasets.GeneratorBasedBuilder): | |
| """Builder for Conceptual Captions dataset.""" | |
| VERSION = datasets.Version("1.0.0") | |
| BUILDER_CONFIGS = [ | |
| datasets.BuilderConfig("unlabeled", version=VERSION, description=_DESCRIPTIONS["unlabeled"]), | |
| datasets.BuilderConfig("labeled", version=VERSION, description=_DESCRIPTIONS["labeled"]), | |
| ] | |
| DEFAULT_CONFIG_NAME = "unlabeled" | |
| def _info(self): | |
| features = datasets.Features( | |
| { | |
| "image_url": datasets.Value("string"), | |
| "caption": datasets.Value("string"), | |
| }, | |
| ) | |
| if self.config.name == "labeled": | |
| features.update( | |
| { | |
| "labels": datasets.Sequence(datasets.Value("string")), | |
| "MIDs": datasets.Sequence(datasets.Value("string")), | |
| "confidence_scores": datasets.Sequence(datasets.Value("float64")), | |
| } | |
| ) | |
| return datasets.DatasetInfo( | |
| description=_DESCRIPTION, | |
| features=features, | |
| supervised_keys=None, | |
| homepage=_HOMEPAGE, | |
| license=_LICENSE, | |
| citation=_CITATION, | |
| ) | |
| def _split_generators(self, dl_manager): | |
| downloaded_data = dl_manager.download(_URLS[self.config.name]) | |
| splits = [ | |
| datasets.SplitGenerator( | |
| name=datasets.Split.TRAIN, | |
| gen_kwargs={"annotations_file": downloaded_data["train"]}, | |
| ), | |
| ] | |
| if self.config.name == "unlabeled": | |
| splits += [ | |
| datasets.SplitGenerator( | |
| name=datasets.Split.VALIDATION, | |
| gen_kwargs={"annotations_file": downloaded_data["validation"]}, | |
| ), | |
| ] | |
| return splits | |
| def _generate_examples(self, annotations_file): | |
| if self.config.name == "unlabeled": | |
| with open(annotations_file, encoding="utf-8") as f: | |
| for i, row in enumerate(csv.reader(f, delimiter="\t")): | |
| # Sanity check | |
| assert len(row) == 2 | |
| caption, image_url = row | |
| yield i, { | |
| "image_url": image_url, | |
| "caption": caption, | |
| }, | |
| else: | |
| with open(annotations_file, encoding="utf-8") as f: | |
| for i, row in enumerate(csv.reader(f, delimiter="\t")): | |
| caption, image_url, labels, MIDs, confidence_scores = row | |
| if not labels: | |
| continue | |
| yield i, { | |
| "image_url": image_url, | |
| "caption": caption, | |
| "labels": labels.split(","), | |
| "MIDs": MIDs.split(","), | |
| "confidence_scores": [float(x) for x in confidence_scores.split(",")], | |
| }, | |