|
[1] Kaiming He et al. “Deep residual learning for image recognition”.Proceedings ofthe IEEE conference on computer vision and pattern recognition. 2016, pp. 770–778. [2] Kaiming He et al. “Identity mappings in deep residual networks”.Europeanconference on computer vision. Springer. 2016, pp. 630–645. [3] Shanghua Gao et al. “Res2net: A new multi-scale backbone architecture”.IEEEtransactions on pattern analysis and machine intelligence(2019). [4] Saining Xie et al. “Aggregated residual transformations for deep neural net-works”.Proceedings of the IEEE conference on computer vision and patternrecognition. 2017, pp. 1492–1500. [5] Christian Szegedy et al. “Going deeper with convolutions”.Proceedings of theIEEE conference on computer vision and pattern recognition. 2015, pp. 1–9. [6] Christian Szegedy et al. “Rethinking the inception architecture for computervision”.Proceedings of the IEEE conference on computer vision and patternrecognition. 2016, pp. 2818–2826. [7] Christian Szegedy et al. “Inception-v4, inception-resnet and the impact of resid-ual connections on learning”.arXiv preprint arXiv:1602.07261(2016). [8] Gao Huang et al. “Densely connected convolutional networks”.Proceedings ofthe IEEE conference on computer vision and pattern recognition. 2017, pp. 4700–4708. [9] Karen Simonyan and Andrew Zisserman. “Very deep convolutional networks forlarge-scale image recognition”.arXiv preprint arXiv:1409.1556(2014). [10] Forrest Iandola et al. “Densenet: Implementing efficient convnet descriptor pyra-mids”.arXiv preprint arXiv:1404.1869(2014).[11] Xiangyu Zhang et al. “Shufflenet: An extremely efficient convolutional neuralnetwork for mobile devices”.Proceedings of the IEEE conference on computervision and pattern recognition. 2018, pp. 6848–6856. [12] François Chollet. “Xception: Deep learning with depthwise separable convo-lutions”.Proceedings of the IEEE conference on computer vision and patternrecognition. 2017, pp. 1251–1258. [13] Mark Sandler et al. “Mobilenetv2: Inverted residuals and linear bottlenecks”.Proceedings of the IEEE conference on computer vision and pattern recognition.2018, pp. 4510–4520. [14] Huiyu Wang et al. “Elastic: Improving cnns with dynamic scaling policies”.Pro-ceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2019, pp. 2258–2267. [15] Chun-Fu Chen et al. “Big-little net: An efficient multi-scale feature representa-tion for visual and speech recognition”.arXiv preprint arXiv:1807.03848(2018). [16] Yunpeng Chen et al. “Drop an octave: Reducing spatial redundancy in con-volutional neural networks with octave convolution”.Proceedings of the IEEEInternational Conference on Computer Vision. 2019, pp. 3435–3444. [17] Jie Hu, Li Shen, and Gang Sun. “Squeeze-and-excitation networks”.Proceed-ings of the IEEE conference on computer vision and pattern recognition. 2018,pp. 7132–7141. [18] Sanghyun Woo et al. “Cbam: Convolutional block attention module”.Proceed-ings of the European conference on computer vision (ECCV). 2018, pp. 3–19. [19] Xiang Li et al. “Selective kernel networks”.Proceedings of the IEEE conferenceon computer vision and pattern recognition. 2019, pp. 510–519. [20] Qilong Wang et al. “ECA-net: Efficient channel attention for deep convolutionalneural networks”.Proceedings of the IEEE/CVF Conference on Computer Vi-sion and Pattern Recognition. 2020, pp. 11534–11542. [21] Hengshuang Zhao et al. “Pyramid scene parsing network”.Proceedings of theIEEE conference on computer vision and pattern recognition. 2017, pp. 2881–2890. [22] Tsung-Yi Lin et al. “Feature pyramid networks for object detection”.Proceed-ings of the IEEE conference on computer vision and pattern recognition. 2017,pp. 2117–2125. [23] Ke Sun et al. “Deep high-resolution representation learning for human poseestimation”.Proceedings of the IEEE conference on computer vision and patternrecognition. 2019, pp. 5693–5703. [24] Gao Huang et al. “Multi-scale dense networks for resource efficient image clas-sification”.arXiv preprint arXiv:1703.09844(2017). [25] Tsung-Wei Ke, Michael Maire, and Stella X Yu. “Multigrid neural architec-tures”.Proceedings of the IEEE Conference on Computer Vision and PatternRecognition. 2017, pp. 6665–6673. [26] Sergey Ioffe and Christian Szegedy. “Batch normalization: Accelerating deepnetwork training by reducing internal covariate shift”.arXiv preprint arXiv:1502.03167(2015). [27] Kelvin Xu et al. “Show, attend and tell: Neural image caption generation withvisual attention”.International conference on machine learning. 2015, pp. 2048–2057. [28] Tony Lindeberg.Scale-space theory in computer vision. Vol. 256. Springer Sci-ence & Business Media, 2013. [29] Jia Deng et al. “Imagenet: A large-scale hierarchical image database”.2009 IEEEconference on computer vision and pattern recognition. Ieee. 2009, pp. 248–255. [30] Vinod Nair and Geoffrey E Hinton. “Rectified linear units improve restrictedboltzmann machines”.ICML. 2010. |