mmocr/tests/test_evaluation/test_metrics/test_f_metric.py

158 lines
5.9 KiB
Python
Raw Normal View History

2022-07-07 16:33:09 +08:00
# Copyright (c) OpenMMLab. All rights reserved.
import unittest
import torch
from mmengine.structures import InstanceData
2022-07-07 16:33:09 +08:00
from mmocr.evaluation import F1Metric
from mmocr.structures import KIEDataSample
2022-07-07 16:33:09 +08:00
class TestF1Metric(unittest.TestCase):
def test_init(self):
with self.assertRaises(AssertionError):
F1Metric(num_classes='3')
with self.assertRaises(AssertionError):
F1Metric(num_classes=3, ignored_classes=[1], cared_classes=[0])
with self.assertRaises(AssertionError):
F1Metric(num_classes=3, ignored_classes=1)
with self.assertRaises(AssertionError):
F1Metric(num_classes=2, mode=1)
with self.assertRaises(AssertionError):
F1Metric(num_classes=1, mode='1')
def test_macro_f1(self):
mode = 'macro'
preds_cases = [
2022-07-07 16:33:09 +08:00
[
KIEDataSample(
pred_instances=InstanceData(
labels=torch.LongTensor([0, 1, 2])),
gt_instances=InstanceData(
labels=torch.LongTensor([0, 1, 4])))
2022-07-07 16:33:09 +08:00
],
[
KIEDataSample(
gt_instances=InstanceData(labels=torch.LongTensor([0, 1])),
pred_instances=InstanceData(
labels=torch.LongTensor([0, 1]))),
KIEDataSample(
gt_instances=InstanceData(labels=torch.LongTensor([4])),
pred_instances=InstanceData(labels=torch.LongTensor([2])))
]
2022-07-07 16:33:09 +08:00
]
# num_classes < the maximum label index
metric = F1Metric(num_classes=3, ignored_classes=[1])
metric.process(None, preds_cases[0])
2022-07-07 16:33:09 +08:00
with self.assertRaises(AssertionError):
metric.evaluate(size=1)
for preds in preds_cases:
2022-07-07 16:33:09 +08:00
metric = F1Metric(num_classes=5, mode=mode)
metric.process(None, preds)
result = metric.evaluate(size=len(preds))
2022-07-07 16:33:09 +08:00
self.assertAlmostEqual(result['kie/macro_f1'], 0.4)
# Test ignored_classes
metric = F1Metric(num_classes=5, ignored_classes=[1], mode=mode)
metric.process(None, preds)
result = metric.evaluate(size=len(preds))
2022-07-07 16:33:09 +08:00
self.assertAlmostEqual(result['kie/macro_f1'], 0.25)
# Test cared_classes
metric = F1Metric(
num_classes=5, cared_classes=[0, 2, 3, 4], mode=mode)
metric.process(None, preds)
result = metric.evaluate(size=len(preds))
2022-07-07 16:33:09 +08:00
self.assertAlmostEqual(result['kie/macro_f1'], 0.25)
def test_micro_f1(self):
mode = 'micro'
preds_cases = [[
KIEDataSample(
gt_instances=InstanceData(
labels=torch.LongTensor([0, 1, 0, 1, 2])),
2022-07-07 16:33:09 +08:00
pred_instances=InstanceData(
labels=torch.LongTensor([0, 1, 2, 2, 0])))
],
[
KIEDataSample(
gt_instances=InstanceData(
labels=torch.LongTensor([0, 1, 2])),
2022-07-07 16:33:09 +08:00
pred_instances=InstanceData(
labels=torch.LongTensor([0, 1, 0]))),
KIEDataSample(
gt_instances=InstanceData(
labels=torch.LongTensor([0, 1])),
2022-07-07 16:33:09 +08:00
pred_instances=InstanceData(
labels=torch.LongTensor([2, 2])))
]]
# num_classes < the maximum label index
metric = F1Metric(num_classes=1, ignored_classes=[0], mode=mode)
metric.process(None, preds_cases[0])
2022-07-07 16:33:09 +08:00
with self.assertRaises(AssertionError):
metric.evaluate(size=1)
for preds in preds_cases:
2022-07-07 16:33:09 +08:00
# class 0: tp: 1, fp: 1, fn: 1
# class 1: tp: 1, fp: 1, fn: 0
# class 2: tp: 0, fp: 1, fn: 2
# overall: tp: 2, fp: 3, fn: 3
# f1: 0.4
metric = F1Metric(num_classes=3, mode=mode)
metric.process(None, preds)
result = metric.evaluate(size=len(preds))
2022-07-07 16:33:09 +08:00
self.assertAlmostEqual(result['kie/micro_f1'], 0.4, delta=0.01)
metric = F1Metric(num_classes=5, mode=mode)
metric.process(None, preds)
result = metric.evaluate(size=len(preds))
2022-07-07 16:33:09 +08:00
self.assertAlmostEqual(result['kie/micro_f1'], 0.4, delta=0.01)
# class 0: tp: 1, fp: 1, fn: 1
# class 2: tp: 0, fp: 1, fn: 2
# overall: tp: 1, fp: 2, fn: 3
# f1: 0.285
metric = F1Metric(num_classes=5, ignored_classes=[1], mode=mode)
metric.process(None, preds)
result = metric.evaluate(size=len(preds))
2022-07-07 16:33:09 +08:00
self.assertAlmostEqual(result['kie/micro_f1'], 0.285, delta=0.001)
metric = F1Metric(
num_classes=5, cared_classes=[0, 2, 3, 4], mode=mode)
metric.process(None, preds)
result = metric.evaluate(size=len(preds))
2022-07-07 16:33:09 +08:00
self.assertAlmostEqual(result['kie/micro_f1'], 0.285, delta=0.001)
def test_arguments(self):
mode = ['micro', 'macro']
preds = [
KIEDataSample(
gt_instances=InstanceData(
test_labels=torch.LongTensor([0, 1, 0, 1, 2])),
2022-07-07 16:33:09 +08:00
pred_instances=InstanceData(
test_labels=torch.LongTensor([0, 1, 2, 2, 0])))
]
# class 0: tp: 1, fp: 1, fn: 1
# class 1: tp: 1, fp: 1, fn: 0
# class 2: tp: 0, fp: 1, fn: 2
# overall: tp: 2, fp: 3, fn: 3
# micro_f1: 0.4
# macro_f1:
metric = F1Metric(num_classes=3, mode=mode, key='test_labels')
metric.process(None, preds)
2022-07-07 16:33:09 +08:00
result = metric.evaluate(size=1)
self.assertAlmostEqual(result['kie/micro_f1'], 0.4, delta=0.01)
self.assertAlmostEqual(result['kie/macro_f1'], 0.39, delta=0.01)