File size: 6,577 Bytes
15a0ce2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 |
# coding=utf-8
import json
import os.path
import datasets
from seacrowd.utils import schemas
from seacrowd.utils.configs import SEACrowdConfig
from seacrowd.utils.constants import Licenses, Tasks
_DATASETNAME = "uit_viic"
_CITATION = """\
@InProceedings{10.1007/978-3-030-63007-2_57,
author="Lam, Quan Hoang
and Le, Quang Duy
and Nguyen, Van Kiet
and Nguyen, Ngan Luu-Thuy",
editor="Nguyen, Ngoc Thanh
and Hoang, Bao Hung
and Huynh, Cong Phap
and Hwang, Dosam
and Trawi{\'{n}}ski, Bogdan
and Vossen, Gottfried",
title="UIT-ViIC: A Dataset for the First Evaluation on Vietnamese Image Captioning",
booktitle="Computational Collective Intelligence",
year="2020",
publisher="Springer International Publishing",
address="Cham",
pages="730--742",
abstract="Image Captioning (IC), the task of automatic generation of image captions, has attracted
attentions from researchers in many fields of computer science, being computer vision, natural language
processing and machine learning in recent years. This paper contributes to research on Image Captioning
task in terms of extending dataset to a different language - Vietnamese. So far, there has been no existed
Image Captioning dataset for Vietnamese language, so this is the foremost fundamental step for developing
Vietnamese Image Captioning. In this scope, we first built a dataset which contains manually written
captions for images from Microsoft COCO dataset relating to sports played with balls, we called this dataset
UIT-ViIC (University Of Information Technology - Vietnamese Image Captions). UIT-ViIC consists of 19,250
Vietnamese captions for 3,850 images. Following that, we evaluated our dataset on deep neural network models
and did comparisons with English dataset and two Vietnamese datasets built by different methods. UIT-ViIC
is published on our lab website (https://sites.google.com/uit.edu.vn/uit-nlp/) for research purposes.",
isbn="978-3-030-63007-2"
}
"""
_DESCRIPTION = """
UIT-ViIC contains manually written captions for images from Microsoft COCO dataset relating to sports
played with ball. UIT-ViIC consists of 19,250 Vietnamese captions for 3,850 images. For each image,
UIT-ViIC provides five Vietnamese captions annotated by five annotators.
"""
_HOMEPAGE = "https://drive.google.com/file/d/1YexKrE6o0UiJhFWpE8M5LKoe6-k3AiM4"
_PAPER_URL = "https://arxiv.org/abs/2002.00175"
_LICENSE = Licenses.UNKNOWN.value
_HF_URL = ""
_LANGUAGES = ["vi"]
_LOCAL = False
_SUPPORTED_TASKS = [Tasks.IMAGE_CAPTIONING]
_SOURCE_VERSION = "1.0.0"
_SEACROWD_VERSION = "2024.06.20"
_URLS = "https://drive.google.com/uc?export=download&id=1YexKrE6o0UiJhFWpE8M5LKoe6-k3AiM4"
_Split_Path = {
"train": "UIT-ViIC/uitviic_captions_train2017.json",
"validation": "UIT-ViIC/uitviic_captions_val2017.json",
"test": "UIT-ViIC/uitviic_captions_test2017.json",
}
class UITViICDataset(datasets.GeneratorBasedBuilder):
BUILDER_CONFIGS = [
SEACrowdConfig(name=f"{_DATASETNAME}_source", version=datasets.Version(_SOURCE_VERSION), description=_DESCRIPTION, subset_id=f"{_DATASETNAME}", schema="source"),
SEACrowdConfig(name=f"{_DATASETNAME}_seacrowd_imtext", version=datasets.Version(_SEACROWD_VERSION), description=_DESCRIPTION, subset_id=f"{_DATASETNAME}", schema="seacrowd_imtext"),
]
def _info(self):
if self.config.schema == "source":
features = datasets.Features(
{
"license": datasets.Value("int32"),
"file_name": datasets.Value("string"),
"coco_url": datasets.Value("string"),
"flickr_url": datasets.Value("string"),
"height": datasets.Value("int32"),
"width": datasets.Value("int32"),
"date_captured": datasets.Value("string"),
"image_id": datasets.Value("int32"),
"caption": datasets.Value("string"),
"cap_id": datasets.Value("int32"),
}
)
elif self.config.schema == "seacrowd_imtext":
features = schemas.image_text_features()
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
license=_LICENSE,
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
file_paths = dl_manager.download_and_extract(_URLS)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={"filepath": os.path.join(file_paths, _Split_Path["train"])},
),
datasets.SplitGenerator(
name=datasets.Split.VALIDATION,
gen_kwargs={"filepath": os.path.join(file_paths, _Split_Path["validation"])},
),
datasets.SplitGenerator(
name=datasets.Split.TEST,
gen_kwargs={"filepath": os.path.join(file_paths, _Split_Path["test"])},
),
]
def _generate_examples(self, filepath):
"""Yields examples."""
with open(filepath, encoding="utf-8") as f:
json_dict = json.load(f)
images = {itm["id"]: itm for itm in json_dict["images"]}
captns = json_dict["annotations"]
for idx, capt in enumerate(captns):
image_id = capt["image_id"]
if self.config.schema == "source":
yield idx, {
"license": images[image_id]["license"],
"file_name": images[image_id]["file_name"],
"coco_url": images[image_id]["coco_url"],
"flickr_url": images[image_id]["flickr_url"],
"height": images[image_id]["height"],
"width": images[image_id]["width"],
"date_captured": images[image_id]["date_captured"],
"image_id": capt["image_id"],
"caption": capt["caption"],
"cap_id": capt["id"],
}
elif self.config.schema == "seacrowd_imtext":
yield idx, {
"id": capt["id"],
"image_paths": [images[image_id]["coco_url"], images[image_id]["flickr_url"]],
"texts": capt["caption"],
"metadata": {
"context": "",
"labels": ["Yes"],
},
}
|