为了使得视觉与学习领域相关从业者快速及时地了解领域的最新发展动态和前沿技术进展,VALSE最新推出了《论文速览》栏目,将在每周发布一至两篇顶会顶刊论文的录制视频,对单个前沿工作进行细致讲解。本期VALSE论文速览选取了来自浙江大学的3D视图合成方面的工作,该视频由论文第一作者浙江大学彭思达同学录制。
论文题目:Neural Body: Implicit Neural Representations with Structured Latent Codes for Novel View Synthesis of Dynamic Humans
作者列表:彭思达 (浙江大学),张远青 (浙江大学),徐英豪 (香港中文大学),王倩倩 (康奈尔大学),帅青 (浙江大学),鲍虎军 (浙江大学),周晓巍 (浙江大学)
B站观看网址:
https://www.bilibili.com/video/BV15F411p7H7/
复制链接到浏览器打开或点击阅读原文即可跳转至观看页面。
论文摘要:
This paper addresses the challenge of novel view synthesis for a human performer from a very sparse set of camera views. Some recent works have shown that learning implicit neural representations of 3D scenes achieves remarkable view synthesis quality given dense input views. However, the representation learning will be ill-posed if the views are highly sparse. To solve this ill-posed problem, our key idea is to integrate observations over video frames. To this end, we propose Neural Body, a new human body representation which assumes that the learned neural representations at different frames share the same set of latent codes anchored to a deformable mesh, so that the observations across frames can be naturally integrated. The deformable mesh also provides geometric guidance for the network to learn 3D representations more efficiently. To evaluate our approach, we create a multi-view dataset named ZJU-MoCap that captures performers with complex motions. Experiments on ZJU-MoCap show that our approach outperforms prior works by a large margin in terms of novel view synthesis quality. We also demonstrate the capability of our approach to reconstruct a moving person from a monocular video on the People-Snapshot dataset.
论文信息:
[1] Sida Peng, Yuanqing Zhang, Yinghao Xu, Qianqian Wang, Qing Shuai, Hujun Bao, Xiaowei Zhou, Neural Body: Implicit Neural Representations with Structured Latent Codes for Novel View Synthesis of Dynamic Humans. In CVPR, 2021.
论文链接:
[https://openaccess.thecvf.com/content/CVPR2021/papers/Peng_Neural_Body_Implicit_Neural_Representations_With_Structured_Latent_Codes_for_CVPR_2021_paper.pdf]
代码链接:
[https://github.com/zju3dv/neuralbody]
视频讲者简介:
彭思达是浙江大学CAD&CG国家重点实验室四年级博士研究生,导师为周晓巍研究员。研究方向为三维视觉,主要研究三维重建与视角合成。博士至今以一作身份在TPAMI、CVPR、ICCV等会议或期刊发表5篇论文,论文引用超过500次。在2020年获得CCF-CV学术新锐奖,全国仅评选3人。在2021年一作论文入围CVPR Best Paper Candidates。发表论文均已开源,在GitHub上Star数超过2000次。
特别鸣谢本次论文速览主要组织者:
月度轮值AC:董宣毅 (Amazon)、谢凌曦 (华为数字技术有限公司)
季度责任AC:许永超 (武汉大学)
活动参与方式
1、VALSE每周举行的Webinar活动依托B站直播平台进行,欢迎在B站搜索VALSE_Webinar关注我们!
直播地址:
https://live.bilibili.com/22300737;
历史视频观看地址:
https://space.bilibili.com/562085182/
2、VALSE Webinar活动通常每周三晚上20:00进行,但偶尔会因为讲者时区问题略有调整,为方便您参加活动,请关注VALSE微信公众号:valse_wechat 或加入VALSE QQ R群,群号:137634472);
*注:申请加入VALSE QQ群时需验证姓名、单位和身份,缺一不可。入群后,请实名,姓名身份单位。身份:学校及科研单位人员T;企业研发I;博士D;硕士M。
3、VALSE微信公众号一般会在每周四发布下一周Webinar报告的通知。
4、您也可以通过访问VALSE主页:http://valser.org/ 直接查看Webinar活动信息。Webinar报告的PPT(经讲者允许后),会在VALSE官网每期报告通知的最下方更新。
看到这了,确定不关注下吗?关注下再走吧~