|
[1]D.LopezPazandM.Ranzato, “Gradient episodic memory for continual learning,” in Advances in Neural Information Processing Systems, pp. 6467–6476, 2017.1,6,17 [2]A. Chaudhry, M. Ranzato, M. Rohrbach, and M. Elhoseiny, “Efficient lifelong learning with A-GEM,” in International Conference on Learning Representations,2019.1,6,16,17,18,20 [3]R. Aljundi, M. Lin, B. Goujaud, and Y. Bengio, “Gradient-based sample selection for online continual learning,” in Advances in Neural Information Processing Systems 32(H. Wallach, H. Larochelle, A. Beygelzimer, F. d'AlchéBuc, E. Fox, and R. Garnett, eds.), pp. 11816–11825, Curran Associates, Inc., 2019.1,6,7,16,17,18,19,20 [4]R. Aljundi, E. Belilovsky, T. Tuytelaars, L. Charlin, M. Caccia, M. Lin, and L. PageCaccia, “Online continual learning with maximal interfered retrieval,” in Advances in Neural Information Processing Systems 32(H. Wallach, H. Larochelle, A. Beygelzimer, F. d'AlchéBuc, E. Fox, and R. Garnett, eds.),pp. 11849–11860, Curran Associates, Inc., 2019.1,6,7 [5]J. Kirkpatrick, R. Pascanu, N. Rabinowitz, J. Veness, G. Desjardins, A. A. Rusu,K.Milan, J.Quan, T.Ramalho, A.GrabskaBarwinska, et al., “Overcoming catastrophic forgetting in neural networks,” Proceedings of the national academy of sciences, vol. 114, no. 13, pp. 3521–3526, 2017.1,5,12,17,18,19,20 [6]M. Riemer, I. Cases, R. Ajemian, M. Liu, I. Rish, Y. Tu, and G. Tesauro, “Learning to learn without forgetting by maximizing transfer and minimizing interference,” in International Conference on Learning Representations, 2019.2,6,7 [7]A.C. Cheng, C. H. Lin, D.C. Juan, W. Wei, and M. Sun, “Instanas: Instanceaware neural architecture search,”arXiv preprint arXiv:1811.10201, 2018.2,7,9 [8]Z.Wu, T.Nagarajan, A.Kumar, S.Rennie, L.S.Davis, K.Grauman, and R.Feris, “Blockdrop: Dynamic inference paths in residual networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8817–8826,2018.2 [9]A.VeitandS.Belongie, “Convolutional networks with adaptive inference graphs,” in Proceedings of the European Conference on Computer Vision (ECCV), pp. 3–18, 2018.227 [10]T. L. Hayes and C. Kanan, “Lifelong machine learning with deep streaming linear discriminant analysis,” in Proceedings of the IEEE/CVF Conference on ComputerVision and Pattern Recognition Workshops, pp. 220–221, 2020.5,6 [11]X. Li, Y. Zhou, T. Wu, R. Socher, and C. Xiong, “Learn to grow: A continual structure learning framework for overcoming catastrophic forgetting,” in International Conference on Machine Learning, pp. 3925–3934, 2019.5,12 [12]A. Mallya and S. Lazebnik, “Packnet: Adding multiple tasks to a single network by iterative pruning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7765–7773, 2018.5,10 [13]J. Serra, D. Suris, M. Miron, and A. Karatzoglou, “Overcoming catastrophic forgetting with hard attention to the task,” in Proceedings of the 35th International Conference on Machine Learning(J. Dy and A. Krause, eds.), vol. 80 ofProceedings of Machine Learning Research, (Stockholmsmässan, Stockholm Sweden), pp. 4548–4557, PMLR, 10–15 Jul 2018.5,10,17,18,19,20 [14]C.Y.Hung, C.H.Tu, C.E.Wu, C.H.Chen, Y.M.Chan, andC.S.Chen, “Compacting, picking and growing for unforgetting continual learning,” in Advances in Neural Information Processing Systems, pp. 13647–13657, 2019.5 [15]R.Aljundi, P.Chakravarty,andT.Tuytelaars, “Expertgate: Lifelong learning with a network of experts,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3366–3375, 2017.5,6 [16]J. Rajasegaran, M. Hayat, S. Khan, F. S. Khan, and L. Shao, “Random path selection for incremental learning,” Advances in Neural Information Processing Systems, 2019.5,6,16,17,18 [17]T. L. Hayes, N. D. Cahill, and C. Kanan, “Memory efficient experience replay for streaming learning,” in 2019 International Conference on Robotics and Automation (ICRA), pp. 9769–9776, IEEE, 2019.6 [18]T. L. Hayes, K. Kafle, R. Shrestha, M. Acharya, and C. Kanan, “Remind your neural network to prevent catastrophic forgetting,” in Proceedings of the European Conference on Computer Vision (ECCV), 2020.6 [19]Y.C.Hsu, Y.C.Liu, A.Ramasamy, and Z.Kira, “Re evaluating continual learning scenarios: A categorization and case for strong baselines,” in NeurIPS Continual learning Workshop, 2018.6,7 [20]G.M. van de Ven and A.S.Tolias, “Three scenarios for continual learning,”arXivpreprint arXiv:1904.07734, 2019.6,7 [21]S. Yan, J. Xie, and X. He, “Der: Dynamically expandable representation for class incremental learning,”arXiv preprint arXiv:2103.16788, 2021.6 [22]G. Gupta, K. Yadav, and L. Paull, “Lamaml: Lookahead meta-learning for continual learning,”arXiv preprint arXiv:2007.13904, 2020.628 [23]J.Zhang, J.Zhang, S.Ghosh, D.Li, S.Tasci, L.Heck, H.Zhang,andC.C.J.Kuo, “Class incremental learning via deep model consolidation,” inThe IEEE Winter Conference on Applications of Computer Vision, pp. 1131–1140, 2020.7 [24]G. Bender, P.J. Kindermans, B. Zoph, V. Vasudevan, and Q. Le, “Understanding and simplifying one shot architecture search,” in International Conference on Machine Learning, pp. 549–558, 2018.7,12 [25]E. Bengio, P.L. Bacon, J. Pineau, and D. Precup, “Conditional computation in neural networks for faster models,”arXiv preprint arXiv:1511.06297, 2015.7 [26]R. M. French, “Using semidistributed representations to overcome catastrophic forgetting in connectionist networks,” in Proceedings of the 13th annual cognitive science society conference, vol. 1, pp. 173–178, 1991.7 [27]M. Lin, J. Fu, and Y. Bengio, “Conditional computation for continual learning,”arXiv preprint arXiv:1906.06635, 2019.7 [28]R.K.Srivastava, J.Masci, S.Kazerounian, F.Gomez, andJ.Schmidhuber, “Compete to compute,” in Advances in neural information processing systems, pp.2310–2318, 2013.7 [29]M.Mundt, Y.W.Hong, I.Pliushch, and V.Ramesh, “A wholistic view of continual learning with deep neural networks: Forgotten lessons and the bridge to active and open-world learning,”arXiv preprint arXiv:2009.01797, 2020.8 [30]B. Liu, “Learning on the job: Online lifelong and continual learning,” inProceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 13544–13549,2020.8 [31]K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016.9,16 [32]D. Stamoulis, R. Ding, D. Wang, D. Lymberopoulos, B. Priyantha, J. Liu, and D. Marculescu, “Singlepath nas: Designing hardware efficient convnets in less than 4 hours,” inarXiv preprint arXiv:1904.02877, 2019.10 [33]S. J. Rennie, E. Marcheret, Y. Mroueh, J. Ross, and V. Goel, “Selfcritical sequence training for image captioning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7008–7024, 2017.11 [34]S. Kornblith, M. Norouzi, H. Lee, and G. E. Hinton, “Similarity of neural network representations revisited,” in Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 915 June 2019, Long Beach, California, USA(K. Chaudhuri and R. Salakhutdinov, eds.), vol. 97 of Proceedings of Machine Learning Research, pp. 3519–3529, PMLR, 2019.13 [35]H. Tang, R. Houthooft, D. Foote, A. Stooke, O. X. Chen, Y. Duan, J. Schulman,F. DeTurck, and P. Abbeel, “# exploration: A study of countbased exploration for deep reinforcement learning,” in Advances in neural information processing systems, pp. 2753–2762, 2017.1329 [36]R. Aljundi, K. Kelchtermans, and T. Tuytelaars, “Task free continual learning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 11254–11263, 2019.15,19 [37]H.Pham, M.Y.Guan, B.Zoph, Q.V.Le, and J.Dean, “Efficient neural architecture search via parameter sharing,” in Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July1015, 2018, pp. 4092–4101, 2018.17,18 [38]T. DeVries and G. W. Taylor, “Improved regularization of convolutional neural networks with cutout,”arXiv preprint arXiv:1708.04552, 2017.24
|