Novel regularization method for the class imbalance problem

2021 
Abstract In neural network models, obtaining a high-quality dataset is critical because they are generally reliant on training data. A common problem that occurs is class imbalance, in which models tend to be biased to the majority class when the training data is not balanced. To overcome this problem, we propose a novel regularization method that provides a penalty to the loss function, using two facets of the distribution of the model’s output p ( y | x ) : (1) skewed mean and (2) variance divergence between p ( y | x ∈ D + ) and p ( y | x ∈ D − ) . The experimental results demonstrate that our methods consistently improve the performance on imbalanced datasets. Moreover, the combination of two regularization methods provides a substantial performance improvement on five sentence classification datasets and also an image classification dataset; notably, state-of-the-art performances are achieved on the WikiQA and SelQA datasets.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    30
    References
    0
    Citations
    NaN
    KQI
    []