|
[1] S. Peng, J. Dong, Q. Wang, S. Zhang, Q. Shuai, X. Zhou, and H. Bao, “Animatable neural radiance fields for modeling dynamic human bodies,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 14314–14323, 2021. [2] S.Peng, Y.Zhang, Y.Xu,Q.Wang,Q.Shuai,H.Bao,andX.Zhou,“Neuralbody: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9054–9063, 2021. [3] C.-Y. Weng, B. Curless, P. P. Srinivasan, J. T. Barron, and I. Kemelmacher-Shlizerman, “Humannerf: Free-viewpoint rendering of moving people from monocular video,” in Proceedings of the IEEE/CVF conference on computer vision and pattern Recognition, pp. 16210–16220, 2022. [4] M. Loper, N. Mahmood, J. Romero, G. Pons-Moll, and M. J. Black, “Smpl: A skinned multi-person linear model, inSeminalGraphicsPapers: PushingtheBoundaries, Volume 2, pp. 851–866, 2023. [5] G. Pavlakos, V. Choutas, N. Ghorbani, T. Bolkart, A. A. Osman, D. Tzionas, and M. J. Black, “Expressive body capture: 3d hands, face, and body from a single image,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10975–10985, 2019. [6] B.Mildenhall, P.P.Srinivasan, M.Tancik, J.T.Barron, R.Ramamoorthi, andR.Ng,“Nerf: Representing scenes as neural radiance fields for view synthesis,” Communications of the ACM, vol. 65, no. 1, pp. 99–106, 2021. [7] K. Park, U. Sinha, J. T. Barron, S. Bouaziz, D. B. Goldman, S. M. Seitz, and R. Martin Brualla, “Nerfies: Deformable neural radiance fields,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 5865–5874, 2021. [8] A. Pumarola, E. Corona, G. Pons-Moll, and F. Moreno-Noguer, “D-nerf: Neural radiance fields for dynamic scenes,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10318–10327, 2021. [9] E. Tretschk, A. Tewari, V. Golyanik, M. Zollhöfer, C. Lassner, and C. Theobalt, “Non rigid neural radiance fields: Reconstruction and novel view synthesis of a dynamic scene from monocular video,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 12959–12970, 2021. [10] Y. Kwon, D. Kim, D.Ceylan, and H. Fuchs, “Neural human performer: Learning generalizable radiance fields for human performance rendering,” Advances in Neural Information Processing Systems, vol. 34, pp. 24741–24752, 2021. [11] F. Zhao, W. Yang, J. Zhang, P. Lin, Y. Zhang, J. Yu, and L. Xu, “Humannerf: Efficiently generated humanradiancefieldfromsparseinputs,” inProceedingsoftheIEEE/CVFConference on Computer Vision and Pattern Recognition, pp. 7743–7753, 2022. [12] X.Gao,J.Yang,J.Kim,S.Peng,Z.Liu, andX.Tong,“Mps-nerf: Generalizable 3d human rendering from multiview images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022. [13] S. Hu, F. Hong, L. Pan, H. Mei, L. Yang, and Z. Liu, “Sherf: Generalizable human nerf from a single image,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9352–9364, 2023. [14] Z. Cao, T. Simon, S.-E. Wei, and Y. Sheikh, “Realtime multi-person 2d pose estimation using part affinity fields,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7291–7299, 2017. [15] K.-E. Lin, Y.-C. Lin, W.-S. Lai, T.-Y. Lin, Y.-C. Shih, and R. Ramamoorthi, “Vision transformer for nerf-based view synthesis from a single input image,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 806–815, 2023. [16] A.Nichol,H.Jun,P.Dhariwal,P.Mishkin,andM.Chen,“Point-e: Asystemforgenerating 3d point clouds from complex prompts,” arXiv preprint arXiv:2212.08751, 2022. [17] A. Yu, V. Ye, M. Tancik, and A. Kanazawa, “pixelnerf: Neural radiance fields from one or few images,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4578–4587, 2021. [18] C. Gao, A. Saraf, J. Kopf, and J.-B. Huang, “Dynamic view synthesis from dynamic monocular video,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 5712–5721, 2021. [19] K.Park, U.Sinha, P.Hedman, J.T.Barron, S.Bouaziz, D.B.Goldman, R.Martin-Brualla, and S. M. Seitz, “Hypernerf: A higher-dimensional representation for topologically varying neural radiance fields,” arXiv preprint arXiv:2106.13228, 2021. [20] W. Xian, J.-B. Huang, J. Kopf, and C. Kim, “Space-time neural irradiance fields for free viewpoint video,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9421–9431, 2021. [21] J. L. Schonberger and J.-M. Frahm, “Structure-from-motion revisited,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4104–4113, 2016. [22] Y.-L. Liu, C. Gao, A. Meuleman, H.-Y. Tseng, A. Saraf, C. Kim, Y.-Y. Chuang, J. Kopf, andJ.-B. Huang, “Robustdynamicradiancefields,” in Proceedings of the IEEE/CVFConference on Computer Vision and Pattern Recognition, pp. 13–23, 2023. [23] S. Wang, K. Schwarz, A. Geiger, and S. Tang, “Arah: Animatable volume rendering of articulated human sdfs,” in European conference on computer vision, pp. 1–19, Springer, 2022. [24] S.-Y. Su, F. Yu, M. Zollhöfer, and H. Rhodin, “A-nerf: Articulated neural radiance fields for learning human shape, appearance, and pose,” Advances in Neural Information Processing Systems, vol. 34, pp. 12278–12291, 2021. [25] M. Mihajlovic, A. Bansal, M. Zollhoefer, S. Tang, and S. Saito, “Keypointnerf: Generalizing image-based volumetric avatars using relative spatial encoding of keypoints,” in European conference on computer vision, pp. 179–197, Springer, 2022. [26] “Renderpeople,” 2018. https://renderpeople.com/. [27] Z. Cai, D. Ren, A. Zeng, Z. Lin, T. Yu, W. Wang, X. Fan, Y. Gao, Y. Yu, L. Pan, et al., “Humman: Multi-modal 4d human dataset for versatile sensing and modeling,” in European Conference on Computer Vision, pp. 557–577, Springer, 2022. [28] Z. Zheng, T. Yu, Y. Wei, Q. Dai, and Y. Liu, “Deephuman: 3d human reconstruction from a single image,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 7739–7749, 2019. [29] E.R.Chan, C.Z.Lin, M.A.Chan,K.Nagano,B.Pan,S.DeMello,O.Gallo, L.J.Guibas, J. Tremblay, S. Khamis, et al., “Efficient geometry-aware 3d generative adversarial networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16123–16133, 2022. [30] N. Max, “Optical models for direct volume rendering,” IEEE Transactions on Visualization and Computer Graphics, vol. 1, no. 2, pp. 99–108, 1995. [31] A. Maćkiewicz and W. Ratajczak, “Principal components analysis (pca),” Computers & Geosciences, vol. 19, no. 3, pp. 303–342, 1993. [32] R.Ranftl, K.Lasinger, D.Hafner, K.Schindler, andV.Koltun “Towardsrobustmonocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer,” IEEE transactions on pattern analysis and machine intelligence, vol. 44, no. 3, pp. 1623–1637, 2020. [33] Z. Yu, W. Cheng, X. Liu, W. Wu, andK.-Y. Lin, “Monohuman: Animatable human neural field from monocular video,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16943–16953, 2023. [34] J. T. Kajiya and B. P. Von Herzen, “Ray tracing volume densities,” ACM SIGGRAPH computer graphics, vol. 18, no. 3, pp. 165–174, 1984. [35] U. Sara, M. Akter, and M. S. Uddin, “Image quality assessment through fsim, ssim, mse and psnr—a comparative study,” Journal of Computer and Communications, vol. 7, no. 3, pp. 8–18, 2019. [36] Z.Wang,A.C.Bovik,H.R.Sheikh,andE.P.Simoncelli “Imagequalityassessment: from error visibility to structural similarity,” IEEE transactions on image processing, vol. 13, no. 4, pp. 600–612, 2004. [37] R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness of deep features as a perceptual metric,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 586–595, 2018. [38] A.Mittal, R.Soundararajan, andA.C.Bovik, “Makinga“completelyblind"imagequality analyzer,” IEEE Signal processing letters, vol. 20, no. 3, pp. 209–212, 2012. |