|
[1] B. W. Turney, “Anatomy in a modern medical curriculum,” Annals of the Royal College of Surgeons of England, 2007. [2] R. Zhang, P. Isola, and A. A. Efros, “Colorful image colorization,” in European Conference on Computer Vision (ECCV), 2016. [3] P. Isola, J.-Y. Zhu, T. Zhou, and A.A.Efros, “Image-to-image translation with conditional adversarial networks,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017. [4] J. -Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired imagetoimage translation using cycleconsistent adversarial networks,” in IEEE International Conference on Computer Vision (ICCV), 2017. [5] K. Nazeri, E. Ng, T. Joseph, F. Z. Qureshi, and M. Ebrahimi, “Edgeconnect: Generative image inpainting with adversarial edge learning,” arXiv preprint arXiv:1901.00212, 2019. [6] M. J. Ackerman, “The visible human project,” Proceedings of the IEEE, 1998. [7] I. Goodfellow, J. PougetAbadie, M. Mirza, B. Xu, D. WardeFarley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in Advances in Neural In formation Processing Systems (NeurIPS), 2014. [8] J. Gui, Z. Sun, Y. Wen, D. Tao, and J. Ye, “A review on generative adversarial networks: Algorithms, theory, and applications,” arXiv preprint arXiv:2001.06937, 2020. [9] L.A.Gatys, A.S.Ecker, and.Bethge, “Imagestyletransferusingconvolutionalneural networks,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. [10] J. Johnson, A. Alahi, and L. FeiFei, “Perceptual losses for realtime style transfer and superresolution,” in European Conference on Computer Vision (ECCV), 2016. [11] J. M. Wolterink, A. M. Dinkla, M. H. Savenije, P. R. Seevinck, C. A. van den Berg, and I. Išgum, “Deep mr to ct synthesis using unpaired data,” in International workshop on simulation and synthesis in medical imaging, 2017. [12] C.B. Jin, H. Kim, M. Liu, W. Jung, S. Joo, E. Park, Y. S. Ahn, I. H. Han, J. I. Lee, and X. Cui, “Deep ct to mr synthesis using paired and unpaired data,” Sensors, 2019. [13] M.U.G.Khan, Y.Gotoh, and N.Nida, “Medicalimagecolorizationforbettervisualization and segmentation,” in Medical Image Understanding and Analysis, 2017. [14] X. Zeng, S. Tong, Y. Lu, L. Xu, and Z. Huang, “Adaptive medical image deep color perception algorithm,” IEEE Access, 2020. [15] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” arXiv preprint arXiv:1411.1784, 2014. [16] N. Kanopoulos, N. Vasanthavada, and R. L. Baker, “Design of an image edge detection filter using the sobel operator,” IEEE Journal of SolidState Circuits, 1988. [17] C. Ledig, L. Theis, F. Huszár, J. Caballero, A. Cunningham, A. Acosta, A. Aitken, A. Tejani, J. Totz, Z. Wang, et al., “Photorealistic single image superresolution using a generative adversarial network,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017. [18] K.SimonyanandA.Zisserman, “Verydeepconvolutionalnetworksforlargescaleimage recognition,” arXiv preprint arXiv:1409.1556, 2014. [19] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. [20] J. Deng, W. Dong, R. Socher, L.J. Li, K. Li, and L. FeiFei, “ImageNet: A Large Scale Hierarchical Image Database,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2009. [21] V. Spitzer, M. J. Ackerman, A. L. Scherzinger, and D. Whitlock, “The Visible Human Male: A Technical Report,” Journal of the American Medical Informatics Association, 1996. [22] X.Mao,Q.Li,H.Xie,R.Y.Lau,Z.Wang,andS.PaulSmolley,“Leastsquaresgenerative adversarial networks,” in IEEE International Conference on Computer Vision (ICCV), 2017. [23] Zhou Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Transactions on Image Processing, 2004. [24] R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness of deep features as a perceptual metric,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018. [25] D. Pathak, P. Krahenbuhl, J. Donahue, T. Darrell, and A. A. Efros, “Context encoders: Feature learning by inpainting,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. [26] K. Yan, X. Wang, L. Lu, and R. M. Summers, “Deeplesion: automated mining of large scale lesion annotations and universal lesion detection with deep learning,” Journal of medical imaging, 2018. [27] X.Ying, “Anoverviewofoverfittinganditssolutions,”inJournalofPhysics: Conference Series, 2019. [28] Visible Body, “Human anatomy atlas.” https://www.visiblebody.com/ anatomy-and-physiology-apps/human-anatomy-atlas, 2021. [29] M.Y. Liu, X. Huang, A. Mallya, T. Karras, T. Aila, J. Lehtinen, and J. Kautz, “Fewshot unsupervised imagetoimage translation,” in IEEE International Conference on Com puter Vision (ICCV), 2019. [30] M. Wang and W. Deng, “Deep visual domain adaptation: A survey,” Neurocomputing, 2018. [31] Y. Li, N. Wang, J. Liu, and X. Hou, “Demystifying neural style transfer,” arXiv preprint arXiv:1701.01036, 2017. [32] Z. Murez, S. Kolouri, D. Kriegman, R. Ramamoorthi, and K. Kim, “Image to image translation for domain adaptation,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
|