| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798 |
- import json
- import tempfile
- import torch
- from pycocotools.cocoeval import COCOeval
- from dataset.coco import COCODataset
- from utils.box_ops import rescale_bboxes
- class COCOAPIEvaluator():
- def __init__(self, cfg, data_dir, device, transform=None):
- # ----------------- Basic parameters -----------------
- self.image_set = 'val2017'
- self.transform = transform
- self.device = device
- # ----------------- Metrics -----------------
- self.map = 0.
- self.ap50_95 = 0.
- self.ap50 = 0.
- # ----------------- Dataset -----------------
- self.dataset = COCODataset(cfg=cfg, data_dir=data_dir, image_set=self.image_set, transform=None, is_train=False)
- @torch.no_grad()
- def evaluate(self, model):
- model.eval()
- ids = []
- data_dict = []
- num_images = len(self.dataset)
- print('total number of images: %d' % (num_images))
- # start testing
- for index in range(num_images): # all the data in val2017
- if index % 500 == 0:
- print('[Eval: %d / %d]'%(index, num_images))
- # load an image
- img, id_ = self.dataset.pull_image(index)
- orig_h, orig_w, _ = img.shape
- orig_size = [orig_w, orig_h]
- # preprocess
- x, _, ratio = self.transform(img)
- x = x.unsqueeze(0).to(self.device)
-
- id_ = int(id_)
- ids.append(id_)
- # inference
- outputs = model(x)
- scores = outputs['scores']
- labels = outputs['labels']
- bboxes = outputs['bboxes']
- # rescale bboxes
- bboxes = rescale_bboxes(bboxes, orig_size, ratio)
- # process outputs
- for i, box in enumerate(bboxes):
- x1 = float(box[0])
- y1 = float(box[1])
- x2 = float(box[2])
- y2 = float(box[3])
- label = self.dataset.class_ids[int(labels[i])]
-
- bbox = [x1, y1, x2 - x1, y2 - y1]
- score = float(scores[i]) # object score * class score
- A = {"image_id": id_, "category_id": label, "bbox": bbox,
- "score": score} # COCO json format
- data_dict.append(A)
- annType = ['segm', 'bbox', 'keypoints']
- # Evaluate the Dt (detection) json comparing with the ground truth
- if len(data_dict) > 0:
- print('evaluating ......')
- cocoGt = self.dataset.coco
- # workaround: temporarily write data to json file because pycocotools can't process dict in py36.
- _, tmp = tempfile.mkstemp()
- json.dump(data_dict, open(tmp, 'w'))
- cocoDt = cocoGt.loadRes(tmp)
- cocoEval = COCOeval(self.dataset.coco, cocoDt, annType[1])
- cocoEval.params.imgIds = ids
- cocoEval.evaluate()
- cocoEval.accumulate()
- cocoEval.summarize()
- ap50_95, ap50 = cocoEval.stats[0], cocoEval.stats[1]
- print('ap50_95 : ', ap50_95)
- print('ap50 : ', ap50)
- self.map = ap50_95
- self.ap50_95 = ap50_95
- self.ap50 = ap50
- return ap50, ap50_95
- else:
- return 0, 0
|