pranavSIT's picture
added pali inference
74e8f2f
# Copyright 2024 Big Vision Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Evaluator for VQAV2 dataset.
"""
import functools
import re
import big_vision.evaluators.common as c
import big_vision.pp.tokenizer
import big_vision.utils as u
import numpy as np
# Temporary global flag to facilitate backwards compatability. Will be removed
# by the end of year 2023.
API = "jit"
class Evaluator:
"""VQAv2 evaluator."""
def __init__(
self, predict_fn, tokenizer, outfile="{workdir}/{split}.json",
*, data, devices, **kw):
self.get_data_iter, self.steps = c.eval_input_pipeline(
keep_on_cpu={"answers", "answer_type", "question_type", "question_id"},
data=data, devices=devices, **kw)
self.outfile = c.resolve_outfile(outfile, split=data.get("split"))
# We'll need the tokenizer to detokenize the model outputs later.
self.tok = big_vision.pp.tokenizer.get_tokenizer(tokenizer)
self.decode = functools.partial(
predict_fn, devices=devices, eos_token=self.tok.eos_token)
def run(self, train_state):
"""Does one evaluation run, yields metrics."""
accuracies_by_type = {"yes/no": [], "number": [], "other": []}
json_out = []
for _, batch in zip(range(self.steps), self.get_data_iter()):
# (batch, seqlen) array of decoded (generated) token sequences suffixes.
tokens = self.decode(train_state, batch)
# (local_batch,) that indicates padding examples (0) vs real examples (1).
tokens = u.get_local_slice_from_fsarray(tokens)
ex_masks = u.get_local_slice_from_fsarray(batch["_mask"])
# Turn predictions into texts and then scores, one by one.
for i in range(len(tokens)):
if ex_masks[i] == 0: # Skip last-batch padding examples
continue
# Extract the suffix/answer from the generated string, skip bos.
answer = self.tok.to_str(tokens[i], stop_at_eos=True)
json = {"question_id": batch["question_id"][i].item(), "answer": answer}
# The rest is computation of VQA-score which compares to multiple GTs.
# This is described better here: https://visualqa.org/evaluation.html
if (gt_answers := batch["answers"][i]).size:
# Always need to do light space-processing:
gt_answers = [stripspace_vqav2(a) for a in gt_answers]
answer = stripspace_vqav2(answer)
# Only post-process if not all agree. Supposedly avoids postproc OCR:
# https://github.com/GT-Vision-Lab/VQA/issues/14#issuecomment-1334695361
if len(set(gt_answers)) > 1:
answer = postprocess_vqav2_text(answer)
gt_answers = [postprocess_vqav2_text(a) for a in gt_answers]
# Accuracy is avg over all ten leave-one-out GT's.
# https://github.com/GT-Vision-Lab/VQA/issues/1#issuecomment-199921352
# An answer is counted 100% correct as soon as 3 GT's agree with it.
matches = answer == np.array(gt_answers)
acc = np.mean([
np.clip(np.sum(np.delete(matches, i_leave_out)) / 3, 0, 1)
for i_leave_out in range(10)
])
accuracies_by_type[batch["answer_type"][i]].append(acc)
# Update json with fully post-processed answer and gt:
json["answer_raw"] = json["answer"]
json["answer"] = answer
json["gts"] = gt_answers
json_out.append(json)
# At this point `accuracies` is a list of per-example scores. However,
# remember that each host holds a different subset of the examples! So if
# we were to just return the mean accuracy here, we would effectively only
# have evaluated on the main host's (who writes metrics) subset!
# So now, we need to compute global means.
# There is one more caveat: `process_sum` needs the summands on each host
# to have the same size. So we either need to include dummy values for
# the padding examples (last batch, annoying), or we only sum scalars as in
# sufficient statistics, which we do here.
sum_accs = c.process_sum({k: sum(v) for k, v in accuracies_by_type.items()})
num_accs = c.process_sum({k: len(v) for k, v in accuracies_by_type.items()})
num = c.process_sum(len(json_out))
# Yielding metric_name, value means logging the metric.
if n := sum(num_accs.values()):
yield "acc", sum(sum_accs.values()) / n
if n := num_accs["yes/no"]:
yield "acc/yesno", sum_accs["yes/no"] / n
yield "num/yesno", n
if n := num_accs["number"]:
yield "acc/number", sum_accs["number"] / n
yield "num/number", n
if n := num_accs["other"]:
yield "acc/other", sum_accs["other"] / n
yield "num/other", n
yield "num", num # Just for sanity checks.
c.multiprocess_write_json(self.outfile, json_out)
# Post-processing required is described at https://visualqa.org/evaluation.html
def stripspace_vqav2(txt):
return txt.replace("\n", " ").replace("\t", " ").strip()
def postprocess_vqav2_text(txt):
"""Cleanup string according to VQA."""
has_digit_comma = re.search(r"(\d)(\,)(\d)", txt) is not None
out = txt
for p in PUNCT:
# NOTE: digit_comma here looks like a bug in official code, so we follow it.
if has_digit_comma or f"{p} " in txt or f" {p}" in txt:
out = out.replace(p, "")
else:
out = out.replace(p, " ")
# Remove full-stops that aren't part of a number.
out = re.sub(r"(?!<=\d)(\.)(?!\d)", "", out, flags=re.UNICODE)
words = []
for word in out.lower().split():
if word not in ARTICLES:
words.append(REPLACEMENTS.get(word, word))
return " ".join(words)
# pylint: disable=line-too-long
REPLACEMENTS = {
# CONTRACTIONS
"aint": "ain't", "arent": "aren't", "cant": "can't", "couldve": "could've", "couldnt": "couldn't",
"couldn'tve": "couldn't've", "couldnt've": "couldn't've", "didnt": "didn't", "doesnt": "doesn't", "dont": "don't", "hadnt": "hadn't",
"hadnt've": "hadn't've", "hadn'tve": "hadn't've", "hasnt": "hasn't", "havent": "haven't", "hed": "he'd", "hed've": "he'd've",
"he'dve": "he'd've", "hes": "he's", "howd": "how'd", "howll": "how'll", "hows": "how's", "Id've": "I'd've", "I'dve": "I'd've",
"Im": "I'm", "Ive": "I've", "isnt": "isn't", "itd": "it'd", "itd've": "it'd've", "it'dve": "it'd've", "itll": "it'll", "let's": "let's",
"maam": "ma'am", "mightnt": "mightn't", "mightnt've": "mightn't've", "mightn'tve": "mightn't've", "mightve": "might've",
"mustnt": "mustn't", "mustve": "must've", "neednt": "needn't", "notve": "not've", "oclock": "o'clock", "oughtnt": "oughtn't",
"ow's'at": "'ow's'at", "'ows'at": "'ow's'at", "'ow'sat": "'ow's'at", "shant": "shan't", "shed've": "she'd've", "she'dve": "she'd've",
"she's": "she's", "shouldve": "should've", "shouldnt": "shouldn't", "shouldnt've": "shouldn't've", "shouldn'tve": "shouldn't've",
"somebody'd": "somebodyd", "somebodyd've": "somebody'd've", "somebody'dve": "somebody'd've", "somebodyll": "somebody'll",
"somebodys": "somebody's", "someoned": "someone'd", "someoned've": "someone'd've", "someone'dve": "someone'd've",
"someonell": "someone'll", "someones": "someone's", "somethingd": "something'd", "somethingd've": "something'd've",
"something'dve": "something'd've", "somethingll": "something'll", "thats": "that's", "thered": "there'd", "thered've": "there'd've",
"there'dve": "there'd've", "therere": "there're", "theres": "there's", "theyd": "they'd", "theyd've": "they'd've",
"they'dve": "they'd've", "theyll": "they'll", "theyre": "they're", "theyve": "they've", "twas": "'twas", "wasnt": "wasn't",
"wed've": "we'd've", "we'dve": "we'd've", "weve": "we've", "werent": "weren't", "whatll": "what'll", "whatre": "what're",
"whats": "what's", "whatve": "what've", "whens": "when's", "whered": "where'd", "wheres": "where's", "whereve": "where've",
"whod": "who'd", "whod've": "who'd've", "who'dve": "who'd've", "wholl": "who'll", "whos": "who's", "whove": "who've", "whyll": "why'll",
"whyre": "why're", "whys": "why's", "wont": "won't", "wouldve": "would've", "wouldnt": "wouldn't", "wouldnt've": "wouldn't've",
"wouldn'tve": "wouldn't've", "yall": "y'all", "yall'll": "y'all'll", "y'allll": "y'all'll", "yall'd've": "y'all'd've",
"y'alld've": "y'all'd've", "y'all'dve": "y'all'd've", "youd": "you'd", "youd've": "you'd've", "you'dve": "you'd've",
"youll": "you'll", "youre": "you're", "youve": "you've",
# NUMBERS
"none": "0", "zero": "0", "one": "1", "two": "2",
"three": "3", "four": "4", "five": "5", "six": "6",
"seven": "7", "eight": "8", "nine": "9", "ten": "10",
}
# pylint: enable=line-too-long
PUNCT = [
";", "/", "[", "]", "\"", "{", "}",
"(", ")", "=", "+", "\\", "_", "-",
">", "<", "@", "`", ",", "?", "!"
]
ARTICLES = {"a", "an", "the"}