kie_metric.py 2.3 KB
Newer Older
L
add kie  
LDOUBLEV 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13
# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
L
fix  
LDOUBLEV 已提交
14
# The code is refer from: https://github.com/open-mmlab/mmocr/blob/main/mmocr/core/evaluation/kie_metric.py
L
add kie  
LDOUBLEV 已提交
15 16 17 18 19 20

from __future__ import absolute_import
from __future__ import division
from __future__ import print_function

import numpy as np
L
debug  
LDOUBLEV 已提交
21
import paddle
L
add kie  
LDOUBLEV 已提交
22 23 24 25 26 27 28 29

__all__ = ['KIEMetric']


class KIEMetric(object):
    def __init__(self, main_indicator='hmean', **kwargs):
        self.main_indicator = main_indicator
        self.reset()
L
debug  
LDOUBLEV 已提交
30 31
        self.node = []
        self.gt = []
L
add kie  
LDOUBLEV 已提交
32 33 34 35 36

    def __call__(self, preds, batch, **kwargs):
        nodes, _ = preds
        gts, tag = batch[4].squeeze(0), batch[5].tolist()[0]
        gts = gts[:tag[0], :1].reshape([-1])
L
debug  
LDOUBLEV 已提交
37 38
        self.node.append(nodes.numpy())
        self.gt.append(gts)
L
add kie  
LDOUBLEV 已提交
39 40 41 42 43 44 45 46 47 48 49 50 51 52 53

    def compute_f1_score(self, preds, gts):
        ignores = [0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 25]
        C = preds.shape[1]
        classes = np.array(sorted(set(range(C)) - set(ignores)))
        hist = np.bincount(
            (gts * C).astype('int64') + preds.argmax(1), minlength=C
            **2).reshape([C, C]).astype('float32')
        diag = np.diag(hist)
        recalls = diag / hist.sum(1).clip(min=1)
        precisions = diag / hist.sum(0).clip(min=1)
        f1 = 2 * recalls * precisions / (recalls + precisions).clip(min=1e-8)
        return f1[classes]

    def combine_results(self, results):
L
debug  
LDOUBLEV 已提交
54 55 56 57
        node = np.concatenate(self.node, 0)
        gts = np.concatenate(self.gt, 0)
        results = self.compute_f1_score(node, gts)
        data = {'hmean': results.mean()}
L
add kie  
LDOUBLEV 已提交
58 59 60
        return data

    def get_metric(self):
L
debug  
LDOUBLEV 已提交
61

62
        metrics = self.combine_results(self.results)
L
add kie  
LDOUBLEV 已提交
63
        self.reset()
64
        return metrics
L
add kie  
LDOUBLEV 已提交
65 66 67

    def reset(self):
        self.results = []  # clear results
L
debug  
LDOUBLEV 已提交
68 69
        self.node = []
        self.gt = []