| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import argparse |
| | import json |
| | import random |
| | import string |
| | import sys |
| | import time |
| | |
| |
|
| | from pycocoevalcap.tokenizer.ptbtokenizer import PTBTokenizer |
| | from pycocoevalcap.meteor.meteor import Meteor |
| | from pycocoevalcap.cider.cider import Cider |
| | from pycocoevalcap.bleu.bleu import Bleu |
| | from pycocoevalcap.rouge.rouge import Rouge |
| |
|
| | Set = set |
| | import numpy as np |
| |
|
| |
|
| | def random_string(string_length): |
| | letters = string.ascii_lowercase |
| | return ''.join(random.choice(letters) for i in range(string_length)) |
| |
|
| |
|
| | def remove_nonascii(text): |
| | return ''.join([i if ord(i) < 128 else ' ' for i in text]) |
| |
|
| |
|
| | class ANETcaptions(object): |
| | PREDICTION_FIELDS = ['results', 'version', 'external_data'] |
| |
|
| | def __init__(self, ground_truth_filenames=None, prediction_filename=None, |
| | tious=None, distances=[1, 3, 5, 10, 30, 60], max_proposals=1000, |
| | prediction_fields=PREDICTION_FIELDS, verbose=False, no_lang_eval=False): |
| | |
| | if len(tious) == 0: |
| | raise IOError('Please input a valid tIoU.') |
| | if not ground_truth_filenames: |
| | raise IOError('Please input a valid ground truth file.') |
| | if not prediction_filename: |
| | raise IOError('Please input a valid prediction file.') |
| |
|
| | self.verbose = verbose |
| | self.no_lang_eval = no_lang_eval |
| | self.tious = tious |
| | self.distances = distances |
| | self.max_proposals = max_proposals |
| | self.pred_fields = prediction_fields |
| | self.ground_truths = self.import_ground_truths(ground_truth_filenames) |
| | self.prediction = self.import_prediction(prediction_filename) |
| | self.ground_truths_keys = [vid for gt in self.ground_truths for vid in gt] |
| | print('available video number', len(set(self.ground_truths_keys) & set(self.prediction.keys()))) |
| |
|
| | |
| | if not self.no_lang_eval: |
| | self.tokenizer = PTBTokenizer() |
| | self.scorers = [ |
| | (Meteor(), "METEOR"), |
| | (Cider(), "CIDEr"), |
| | (Rouge(), "Rouge-L"), |
| | (Bleu(4), ["Bleu_1", "Bleu_2", "Bleu_3", "Bleu_4"]), |
| | ] |
| |
|
| | def import_prediction(self, prediction_filename): |
| | if self.verbose: |
| | print("| Loading submission...") |
| | if isinstance(prediction_filename, dict): |
| | submission = prediction_filename |
| | else: |
| | submission = json.load(open(prediction_filename)) |
| | |
| | |
| | |
| | results = {} |
| | for vid_id in submission['results']: |
| | results[vid_id] = submission['results'][vid_id][:self.max_proposals] |
| | return results |
| |
|
| | def import_ground_truths(self, filenames): |
| | gts = [] |
| | self.n_ref_vids = Set() |
| | for filename in filenames: |
| | if isinstance(filename, dict): |
| | gt = filename |
| | else: |
| | gt = json.load(open(filename)) |
| | self.n_ref_vids.update(gt.keys()) |
| | gts.append(gt) |
| | if self.verbose: |
| | print("| Loading GT. #files: %d, #videos: %d" % (len(filenames), len(self.n_ref_vids))) |
| | return gts |
| |
|
| | def iou(self, interval_1, interval_2): |
| | start_i, end_i = interval_1[0], interval_1[1] |
| | start, end = interval_2[0], interval_2[1] |
| | intersection = max(0, min(end, end_i) - max(start, start_i)) |
| | union = min(max(end, end_i) - min(start, start_i), end - start + end_i - start_i) |
| | iou = float(intersection) / (union + 1e-8) |
| | return iou |
| |
|
| | def check_gt_exists(self, vid_id): |
| | for gt in self.ground_truths: |
| | if vid_id in gt: |
| | return True |
| | return False |
| |
|
| | def get_gt_vid_ids(self): |
| | vid_ids = set([]) |
| | for gt in self.ground_truths: |
| | vid_ids |= set(gt.keys()) |
| | return list(vid_ids) |
| |
|
| | def evaluate(self): |
| | aggregator = {} |
| | self.scores = {} |
| | if not self.no_lang_eval: |
| | for tiou in self.tious: |
| | scores = self.evaluate_tiou(tiou) |
| | for metric, score in scores.items(): |
| | if metric not in self.scores: |
| | self.scores[metric] = [] |
| | self.scores[metric].append(score) |
| | if True: |
| | |
| | self.scores['Recall'] = [] |
| | self.scores['Precision'] = [] |
| | self.scores['F1'] = [] |
| | for tiou in self.tious: |
| | precision, recall = self.evaluate_detection(tiou) |
| | self.scores['Recall'].append(recall) |
| | self.scores['Precision'].append(precision) |
| | self.scores['F1'].append(2 * recall * precision / (recall + precision) if recall + precision else 0.) |
| | for tiou in self.distances: |
| | precision, recall = self.evaluate_navigation(tiou) |
| | self.scores['Recall'].append(recall) |
| | self.scores['Precision'].append(precision) |
| | self.scores['F1'].append(2 * recall * precision / (recall + precision) if recall + precision else 0.) |
| |
|
| | def evaluate_detection(self, tiou): |
| | gt_vid_ids = self.get_gt_vid_ids() |
| | |
| | |
| | recall = [] |
| | precision = [] |
| | for vid_i, vid_id in enumerate(gt_vid_ids): |
| | if vid_id not in self.prediction: |
| | continue |
| | best_recall = 0 |
| | best_precision = 0 |
| | for gt in self.ground_truths: |
| | if vid_id not in gt: |
| | continue |
| | refs = gt[vid_id] |
| | ref_set_covered = set([]) |
| | pred_set_covered = set([]) |
| | num_gt = 0 |
| | num_pred = 0 |
| | if vid_id in self.prediction: |
| | for pred_i, pred in enumerate(self.prediction[vid_id]): |
| | pred_timestamp = pred['timestamp'] |
| | for ref_i, ref_timestamp in enumerate(refs['timestamps']): |
| | if self.iou(pred_timestamp, ref_timestamp) > tiou: |
| | ref_set_covered.add(ref_i) |
| | pred_set_covered.add(pred_i) |
| |
|
| | new_precision = float(len(pred_set_covered)) / max(len(self.prediction[vid_id]), 1) |
| | best_precision = max(best_precision, new_precision) |
| | new_recall = float(len(ref_set_covered)) / len(refs['timestamps']) |
| | best_recall = max(best_recall, new_recall) |
| | recall.append(best_recall) |
| | precision.append(best_precision) |
| | return sum(precision) / len(precision), sum(recall) / len(recall) |
| |
|
| | def evaluate_navigation(self, tiou): |
| | gt_vid_ids = self.get_gt_vid_ids() |
| | |
| | |
| | recall = [] |
| | precision = [] |
| | for vid_i, vid_id in enumerate(gt_vid_ids): |
| | if vid_id not in self.prediction: |
| | continue |
| | best_recall = 0 |
| | best_precision = 0 |
| | for gt in self.ground_truths: |
| | if vid_id not in gt: |
| | continue |
| | refs = gt[vid_id] |
| | ref_set_covered = set([]) |
| | pred_set_covered = set([]) |
| | num_gt = 0 |
| | num_pred = 0 |
| | if vid_id in self.prediction: |
| | for pred_i, pred in enumerate(self.prediction[vid_id]): |
| | pred_timestamp = pred['timestamp'] |
| | for ref_i, ref_timestamp in enumerate(refs['timestamps']): |
| | if abs(pred_timestamp[0] - ref_timestamp[0]) < tiou: |
| | ref_set_covered.add(ref_i) |
| | pred_set_covered.add(pred_i) |
| |
|
| | new_precision = float(len(pred_set_covered)) / max(len(self.prediction[vid_id]), 1) |
| | best_precision = max(best_precision, new_precision) |
| | new_recall = float(len(ref_set_covered)) / len(refs['timestamps']) |
| | best_recall = max(best_recall, new_recall) |
| | recall.append(best_recall) |
| | precision.append(best_precision) |
| | return sum(precision) / len(precision), sum(recall) / len(recall) |
| |
|
| | def evaluate_tiou(self, tiou): |
| | |
| | res = {} |
| | gts = {} |
| | gt_vid_ids = self.get_gt_vid_ids() |
| |
|
| | unique_index = 0 |
| |
|
| | |
| | vid2capid = {} |
| |
|
| | cur_res = {} |
| | cur_gts = {} |
| |
|
| | for vid_id in gt_vid_ids: |
| |
|
| | |
| | |
| | if vid_id not in self.prediction: |
| | continue |
| |
|
| | |
| | |
| | else: |
| | vid2capid[vid_id] = [] |
| | |
| | for pred in self.prediction[vid_id]: |
| | has_added = False |
| | for gt in self.ground_truths: |
| | if vid_id not in gt: |
| | continue |
| | gt_captions = gt[vid_id] |
| | for caption_idx, caption_timestamp in enumerate(gt_captions['timestamps']): |
| | if self.iou(pred['timestamp'], caption_timestamp) >= tiou: |
| | cur_res[unique_index] = [{'caption': remove_nonascii(pred['sentence'])}] |
| | cur_gts[unique_index] = [ |
| | {'caption': remove_nonascii(gt_captions['sentences'][caption_idx])}] |
| | vid2capid[vid_id].append(unique_index) |
| | unique_index += 1 |
| | has_added = True |
| |
|
| | |
| | |
| | if not has_added: |
| | cur_res[unique_index] = [{'caption': remove_nonascii(pred['sentence'])}] |
| | cur_gts[unique_index] = [{'caption': random_string(random.randint(10, 20))}] |
| | vid2capid[vid_id].append(unique_index) |
| | unique_index += 1 |
| |
|
| | |
| | output = {} |
| | for scorer, method in self.scorers: |
| | if self.verbose: |
| | print('computing %s score...' % (scorer.method())) |
| |
|
| | |
| | all_scores = {} |
| |
|
| | |
| | tokenize_res = self.tokenizer.tokenize(cur_res) |
| | tokenize_gts = self.tokenizer.tokenize(cur_gts) |
| |
|
| | |
| | for vid in vid2capid.keys(): |
| | res[vid] = {index: tokenize_res[index] for index in vid2capid[vid]} |
| | gts[vid] = {index: tokenize_gts[index] for index in vid2capid[vid]} |
| |
|
| | for vid_id in gt_vid_ids: |
| |
|
| | if vid_id not in self.prediction: |
| | continue |
| |
|
| | if len(res[vid_id]) == 0 or len(gts[vid_id]) == 0: |
| | if type(method) == list: |
| | score = [0] * len(method) |
| | else: |
| | score = 0 |
| | else: |
| | score, scores = scorer.compute_score(gts[vid_id], res[vid_id]) |
| | all_scores[vid_id] = score |
| | |
| |
|
| | |
| | if type(method) == list: |
| | scores = np.mean(list(all_scores.values()), axis=0) |
| | for m in range(len(method)): |
| | output[method[m]] = scores[m] |
| | if self.verbose: |
| | print("Calculated tIoU: %1.1f, %s: %0.3f" % (tiou, method[m], output[method[m]])) |
| | else: |
| | output[method] = np.mean(list(all_scores.values())) |
| | if self.verbose: |
| | print("Calculated tIoU: %1.1f, %s: %0.3f" % (tiou, method, output[method])) |
| | return output |
| |
|
| |
|
| | def eval_dvc(submission, references, tious=[0.3, 0.5, 0.7, 0.9], distances=[1, 3, 5, 10, 30, 60], max_proposals_per_video=1000, verbose=False, no_lang_eval=False): |
| | |
| | evaluator = ANETcaptions(ground_truth_filenames=references, |
| | prediction_filename=submission, |
| | tious=tious, |
| | distances=distances, |
| | max_proposals=max_proposals_per_video, |
| | verbose=verbose, no_lang_eval=no_lang_eval) |
| | evaluator.evaluate() |
| | score = evaluator.scores |
| | |
| | loc_score = {} |
| | for i, x in enumerate(tious): |
| | for y in ["Recall", "Precision", "F1"]: |
| | loc_score[y + "@" + str(x)] = score[y][i] |
| | for y in ["Recall", "Precision", "F1"]: |
| | loc_score[y] = np.array([score[y][i] for i in range(len(tious))]).mean() |
| | if distances: |
| | for i, x in enumerate(distances): |
| | for y in ["Recall", "Precision", "F1"]: |
| | loc_score[y + "@" + str(x) + "s"] = score[y][len(tious) + i] |
| | avg_eval_score = {key: np.array(value).mean() for key, value in score.items() if key not in ["Recall", "Precision", "F1"]} |
| | avg_eval_score.update(loc_score) |
| | return avg_eval_score |
| |
|
| | if __name__ == '__main__': |
| | eval_dvc(pred_path, references, |
| | tious=[0.3, 0.5, 0.7, 0.9], |
| | max_proposals_per_video=1000, |
| | verbose=False, |
| | no_lang_eval=False) |
| | eval_soda(pred_path, references, verbose=False) |