|
1. O. Mees, A. Eitel, W. Burgard. “Choosing Smartly: Adaptive Multimodal Fusion for Object Detection in Changing Environments,” IROS, 2016. 2. W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.- Y. Fu, and A. C. Berg. “Ssd: Single shot multibox detector,” ECCV, 2016. 3. J. Dai, Y. Li, K. He, and J. Sun. “R-fcn: Object detection via region-based fully convolutional networks,” arXiv preprint arXiv:1605.06409, 2016. 4. S. Ren, K. He, R. Girshick, and J. Sun. “Faster R-CNN: Towards real-time object detection with region proposal networks,” NIPS, 2015. 5. J. Long, E. Shelhamer, T. Darrell. “Fully Convolutional Networks for Semantic Segmentation,” CVPR, 2015. 6. J. Redmon, S. Divvala, R. Girshick, and A. Farhadi. “You only look once: Unified, real-time object detection,” arXiv preprint arXiv:1506.02640, 2015. 7. A. Eitel, J. T. Springenberg, L. Spinello, M. Riedmiller, and W. Burgard. “Multimodal deep learning for robust RGB-D object recognition,” IROS, 2015. 8. S. Gupta, R. Girshick, P. Arbeláez, and J. Malik. “Learning rich features from rgb-d images for object detection and segmentation,” ECCV, 2014. 9. J. Schlosser, C. K. Chow, and Z. Kira. “Fusing lidar and images for pedestrian detection using convolutional neural networks,” ICRA, 2016. 10. J. Wagner, V. Fischer, M. Herman, and S. Behnke. “Multispectral pedestrian detection using deep fusion convolutional neural networks,” ESANN, 2016. 11. X. Chen, H. Ma, J. Wang, B. Li, T. Xia. “Multi-View 3D Object Detection Network for Autonomous Driving,” CVPR, 2017. 12. A. Geiger, P. Lenz, and R. Urtasun. “Are we ready for autonomous driving? the kitti vision benchmark suite,” CVPR, 2012. 13. S. D. Jain, B. Xiong, K. Grauman. “FusionSeg: Learning to combine motion and appearance for fully automatic segmentation of generic objects in videos,” CVPR, 2017. 14. H. Wang, Y. Wang, Q. Zhang, S. Xiang, and C. Pan. “Locality-Sensitive Deconvolution Networks with Gated Fusion for RGB-D Indoor Semantic Segmentation,” CVPR, 2017 15. S. J. Park, K. S. Hong, S. Lee. “RDFNet: RGB-D Multi-level Residual Feature Fusion for Indoor Semantic Segmentation,” ECCV, 2017. 16. A. Valada, J. Vertens, A. Dhall, W. Burgard. “AdapNet: Adaptive Semantic Segmentation in Adverse Environmental Conditions,” ICRA, 2017. 17. K. He, X. Zhang, S. Ren, and J. Sun. “Deep residual learning for image recognition,” arXiv preprint arXiv:1512.03385, 2015. 18. T. Zahavy, S. Mannor, A. Magnani, A. Krishnan. “IS A PICTURE WORTH A THOUSAND WORDS? A DEEP MULTI-MODAL FUSION ARCHITECTURE FOR PRODUCT CLASSIFICATION IN E-COMMERCE,” ICLR, 2017. 19. A. Frome, G. S Corrado, J. Shlens, S. Bengio, J. Dean, T. Mikolov, et al. “Devise: A deep visual-semantic embedding model,” Advances in neural information processing systems, 2013. 20. S. Poria, E. Cambria, N. Howard, G. B. Huang, and Amir Hussain. “Fusing audio, visual and textual clues for sentiment analysis from multimodal content,” Neurocomputing 174:50–59, 2016. 21. A. Krizhevsky, I. Sutskever, and G. E Hinton. “Imagenet classification with deep convolutional neural networks,” Advances in neural information processing systems, 2012. 22. J. Tompson, R. Goroshin, A. Jain, Y. LeCun, and C. Bregler. “Efficient object localization using convolutional networks,” CVPR, 2015, pp. 648–656. 23. R. A. Jacobs, M. I. Jordan, S. J. Nowlan, and G. E. Hinton. “Adaptive mixtures of local experts,” Neural computation, 1991.
|