A Deep Fully Convolution Neural Network for Semantic Segmentation Based on Adaptive Feature Fusion

2018 
Fully convolutional neural network is a special deep neural networks based on convolutional neural networks and are often used for semantic segmentation. This paper proposes an improved fully convolutional neural network which fuses the feature maps of deeper layers and shallower layers to improve the performance of image segmentation. In the process of feature fusion, adaptive parameters are introduced to enable different layers to participate in feature fusion as different proportion. The deep layers of neural network mainly extract the abstract information of the object, and the shallow layers of neural network extracts the refined features of objects, such as edge information and precise shape. Adaptive parameters can speed up the training speed and improve the prediction accuracy. In the early stages of training, the feature maps of shallow layers have a larger fusion coefficient, which allows the neural network to learn the feature of object's location and shape quickly. As the training progresses, gradually weakening the fusion coefficient of shallow layers and increasing the fusion coefficient of deep layers which can enhance the network's ability of predicting the details of the objects. This paper uses Scene Parsing Challenge 2016 dataset presented by MIT for training. Experiments show that the method proposed in this paper speeds up the training and improves the pixel prediction accuracy.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    4
    References
    4
    Citations
    NaN
    KQI
    []