Skip to content
forked from lxtGH/GALD-DGCNet

Source code and models of GALD net 🎉

Notifications You must be signed in to change notification settings

ytian8/GALD-Net

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

34 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

High Performance Road Scene Semantic Segmentaion 🎉

I will release the all the state-of-the-art models and code trained on Cityscape dataset including Deeplabv3, Deeplabv3+, PSPnet, DAnet, GloreNet, EMANet as soon as possible.

There is also a co-current repo for Fast Road Scene Semantic Segmentation:Fast_Seg ⚡ and thanks for your attention 😃

Some Advices on Training

Please see the Common.md for the details.

Training & validation

Please see the train_distribute.py for the details.

Test & Demo

Please see the demo.py for the details.

Trained Model

GALD-Net (BMVC 2019,arxiv)

We propose Global Aggregation then Local Distribution (GALD) scheme to distribute global information to each position adaptively according to the local information around the position. GALD net achieves top performance on Cityscapes dataset. Both source code and models will be available soon. The work was done at DeepMotion AI Research

GFF-Net (AAAI 2020,arxiv)

We proposed Gated Fully Fusion (GFF) to fuse features from multiple levels through gates in a fully connected way. Specifically, features at each level are enhanced by higher-level features with stronger semantics and lower-level features with more details, and gates are used to control the pass of useful information which significantly reducing noise propagation during fusion. (Joint work: Key Laboratory of Machine Perception, School of EECS @Peking University and DeepMotion AI Research )

DGCNet (BMVC 2019,arxiv)

We propose Dual Graph Convolutional Network (DGCNet) models the global context of the input feature by modelling two orthogonal graphs in a single framework. (Joint work: University of Oxford, Peking University and DeepMotion AI Research)

Comparisons with state-of-the-art models on Cityscapes dataset

Method Conference Backbone mIoU(%)
RefineNet CVPR2017 ResNet-101 73.6
SAC ICCV2017 ResNet-101 78.1
PSPNet CVPR2017 ResNet-101 78.4
DUC-HDC WACV2018 ResNet-101 77.6
AAF ECCV2018 ResNet-101 77.1
BiSeNet ECCV2018 ResNet-101 78.9
PSANet ECCV2018 ResNet-101 80.1
DFN CVPR2018 ResNet-101 79.3
DSSPN CVPR2018 ResNet-101 77.8
DenseASPP CVPR2018 DenseNet-161 80.6
OCNet - ResNet-101 81.7
CCNet ICCV2019 ResNet-101 81.4
GALD-Net BMVC2019 ResNet50 80.8
GALD-Net BMVC2019 ResNet101 81.8
GFF-Net AAAI2020 ResNet101 82.3
DGCN-Net BMVC2019 ResNet101 82.0
GALD-Net(use coarse data) BMVC2019 ResNet101 82.9
GALD-Net(use Mapillary) BMVC2019 ResNet101 83.3

Detailed Results are shown

GALD-Net: here
GFF-Net:here
Both are (Single Model Result)

Citation

Please read our paper for model details. If you find the codebase usefull, please consider cite our paper.

@inproceedings{xiangtl_gald
title={Global Aggregation then Local Distribution in Fully Convolutional Networks},
author={Li, Xiangtai and Zhang, Li and You, Ansheng and Yang, Maoke and Yang, Kuiyuan and Tong, Yunhai},
booktitle={BMVC2019},
}
@inproceedings{xiangtl_gff
  title     = {GFF: Gated Fully Fusion for semantic segmentation},
  author    = {Li, Xiangtai and  Zhao Houlong and Han Lei and Tong Yunhai and Yang Kuiyuan},
  booktitle = {AAAI2020},
  year = {2019}
}
@inproceedings{zhangli_dgcn
title={Dual Graph Convolutional Network for Semantic Segmentation},
author={Zhang, Li(*) and Li, Xiangtai(*) and Arnab, Anurag and Yang, Kuiyuan and Tong, Yunhai and Torr, Philip HS},
booktitle={BMVC2019},
}

License

MIT License

Acknowledgement

Thanks to previous open-sourced repo:
Encoding
CCNet
TorchSeg
pytorchseg

About

Source code and models of GALD net 🎉

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%