|
Qi Zhang (张琦)
I am currently a Researcher in the Visual Computing Center@Tencent AI Lab. I do research in 3D computer
vision and computational photography, where my interests focus on neural radiance fields (NeRF) and
neural rendering, 3D reconstruction and modeling, AI generated content (AIGC), image rectifcation,
and light field imaging.
Before joining Tencent AI Lab in Jun. 2021, I received my Ph.D. degree from the School of Computer
Science of Northwestern Polytechnical University in 2021, I was supervised by Prof. Qing Wang.
I was a visiting student at the Australian National University (ANU) between Jul. 2019 to Aug. 2020,
which was supervised by Prof. Hongdong Li.
I received the Outstanding Doctoral
Dissertation Award Nominee from China Computer Federation (CCF) in 2021. I also won
the 2021 ACM Xi'an Doctoral Dissertation Award and the 2023 NWPU Doctoral Dissertation Award.
At Tencent, I've worked on 4D Content
Generation and
3D Chat on Tencetn Meeting If you
are interested in
the internship about neural rendering (e.g. NeRF, SDF, inverse rendering),
digital avatars, and 3D generation, please feel free to contact me via e-mail and wechat.
Email  / 
CV  / 
Bio  / 
Google Scholar  / 
WeChat  / 
Github
|
News
🎉🎉2024.03: 5 papers accetped to CVPR 2024!
🎉🎉2024.02: DINER extended to TPAMI!
🎉🎉2023.12: NeIF accetped to AAAI 2023
🎉🎉2023.08: LoD-NeuS accetped to SIGGRAPH Asia 2023
🎉🎉2023.07: Pyramid NeRF accetped to IJCV 2023
🎉🎉2023.03: 7 papers (with 1 highlight paper) accetped to CVPR 2023!
🎉🎉2022.08: 1 paper (journal track) accetped to SIGGRAPH Asia 2022
🎉🎉2022.03: 4 papers accetped to CVPR 2022
🎉🎉2021.12: CCF Outstanding Doctoral Dissertation Award Nominee
|
Research
Please find below a complete list of my publications with representative papers highlighted. The IEEE TPAMI, IJCV are top journals in the field of
computer vision and computational photography. The CVPR is the premier conference in Computer Vision
research community.
|
arXiv 2024
|
Analytic-Splatting Anti-Aliased 3D Gaussian Splatting via Analytic Integration
Zhihao Liang*,
Qi Zhang,
Wenbo Hu,
Lei Zhu,
Ying Feng,
Kui Jia
arXiv, 2024
Project Page
/
arXiv
/
Code
/
Viewer
We propose GS-IR, a novel inverse rendering approach based on 3D Gaussian Splatting (GS)
that leverages forward mapping volume rendering to achieve photorealistic novel view
synthesis and relighting results.
|
arXiv 2024
|
Advances in 3D Generation: A Survey
Xiaoyu Li*,
Qi Zhang*,
Di Kang,
Weihao Cheng,
Yiming Gao,
Jingbo Zhang,
Zhihao Liang,
Jing Liao,
Yanpei Cao,
Ying Shan
arXiv, 2024
Project Page
/
arXiv
/
Code
In this survey, we aim to introduce the fundamental methodologies of 3D generation methods and establish a structured roadmap,
encompassing 3D representation, generation methods, datasets, and corresponding applications.
|
arXiv 2024
|
UV Gaussians: Joint Learning of Mesh Deformation and Gaussian Textures for Human Avatar Modeling
Yujiao Jiang,
Qingmin Liao,
Xiaoyu Li,
Li Ma,
Qi Zhang,
Chaopeng Zhang,
Zongqing Lu,
Ying Shan
arXiv, 2024
Project Page
/
arXiv
/
Code
In this paper, we introduce a texture-consistent back view synthesis module that could transfer the reference image
content to the back view through depth and text-guided attention injection with the help of stable diffusion model.
|
CVPR 2024
|
GS-IR: 3D Gaussian Splatting for Inverse Rendering
Zhihao Liang*,
Qi Zhang*,
Ying Feng,
Ying Shan,
Kui Jia
CVPR, 2024
Project Page
/
arXiv
/
Code
We propose GS-IR, a novel inverse rendering approach based on 3D Gaussian Splatting (GS)
that leverages forward mapping volume rendering to achieve photorealistic novel view
synthesis and relighting results.
|
CVPR 2024
|
HumanNorm: Learning Normal Diffusion Model for High-quality and Realistic 3D Human Generation
Xin Huang*,
Ruizhi Shao*,
Qi Zhang,
Hongwen Zhang,
Ying Feng,
Yebin Liu,
Qing Wang
CVPR, 2024
Project Page
/
arXiv
/
Code
We propose HumanNorm, a novel approach for high-quality and realistic 3D human generation by learning the normal diffusion model
including a normal-adapted diffusion model and a normal-aligned diffusion model.
|
CVPR 2024
|
FINER: Flexible spectral-bias tuning in Implicit NEural Representation by Variable-periodic Activation Functions
Zhen Liu*,
Hao Zhu*,
Qi Zhang,
Jingde Fu,
Weibing Deng
Zhan Ma,
Yanwen Guo
Xun Cao
CVPR, 2024
Project Page
/
PDF
/
Code
We have identified that this frequency-related problem can be greatly alleviated by introducing variable-periodic activation functions, for which we propose FINER.
|
CVPR 2024
|
HumanRef: Single Image to 3D Human Generation via Reference-Guided Diffusion
Jingbo Zhang,
Xiaoyu Li,
Qi Zhang,
Yanpei Cao
Ying Shan,
Jing Liao
CVPR, 2024
Project Page
/
arXiv
/
Code
HumanRef, a reference-guided 3D human generation framework, is capable of generating 3D clothed human with realistic,
view-consistent texture and geometry from a single image input with the help of stable diffusion model.
|
CVPR 2024
|
ConTex-Human: Free-View Rendering of Human from a Single Image with Texture-Consistent Synthesis
Xiangjun Gao,
Xiaoyu Li,
Chaopeng Zhang,
Qi Zhang,
Yanpei Cao,
Ying Shan,
Long Quan
CVPR, 2024
Project Page
/
arXiv
/
Code
In this paper, we introduce a texture-consistent back view synthesis module that could transfer the reference image
content to the back view through depth and text-guided attention injection with the help of stable diffusion model.
|
|
NeIF: A Pre-convolved Representation for Plug-and-Play Neural Illumination Fields
Yiyu Zhuang*,
Qi Zhang*,
Xuan Wang,
Hao Zhu,
Ying Feng,
Xiaoyu Li,
Ying Shan,
Xun Cao
AAAI, 2024
Project Page
/
arXiv
/
Code
We propose a fully differentiable framework named neural ambient illumination (NeAI) that uses
Neural Radiance Fields (NeRF) as a lighting model to handle complex lighting in a physically based
way.
|
TPAMI 2024
|
Disorder-invariant Implicit Neural Representation
Hao Zhu*,
Shaowen Xie*,
Zhen Liu*,
Fengyi Liu
Qi Zhang,
You Zhou,
Yi Lin,
Zhan Ma,
Xun Cao,
IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2024
Project Page
/
arXiv
/
Code
In this paper, we find that such a frequency-related problem could be largely solved by re-arranging
the coordinates of the input signal, for which we propose the disorder-invariant implicit neural
representation (DINER) by augmenting a hash-table to a traditional INR backbone.
|
SIGGRAPH Asia 2023
|
Anti-Aliased Neural Implicit Surfaces with Encoding Level of Detail
Yiyu Zhuang*,
Qi Zhang*,
Ying Feng,
Hao Zhu,
Yao Yao,
Xiaoyu Li,
Yanpei Cao,
Ying Shan,
Xun Cao
SIGGRAPH Asia, 2023
Project Page
/
arXiv
/
Code
Our method, called LoD-NeuS, adaptively encodes Level of Detail (LoD) features derived from
the multi-scale and multi-convoluted tri-plane representation. By optimizing a neural Signal Distance
Field (SDF), our method is capable of reconstructing high-fidelity geometry
|
IJCV 2023
|
Pyramid NeRF: Frequency Guided Fast Radiance Field Optimization
Junyu Zhu*,
Hao Zhu*,
Qi Zhang,
Fang Zhu
Zhan Ma,
Xun Cao
International Journal of Computer Vision (IJCV), 2023
Project Page
/
PDF
/
Code
In this paper, we propose the Pyramid NeRF, which guides the NeRF training in a 'low-frequency
first, high-frequency second' style using the image pyramids and could improve the training and
inference speed at 15x and 805x, respectively.
|
CVPR 2023
CVPR 2023
|
Wide-angle Rectification via Content-aware Conformal Mapping
Qi Zhang,
Hongdong Li,
Qing Wang
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023
Project Page
/
arXiv
We propose a new content-aware optimization framework to preserve both local conformal shape (e.g.
face or salient regions) and global linear structures (straight lines).
|
|
Inverting the Imaging Process by Learning an Implicit Camera Model
Xin Huang,
Qi Zhang,
Ying Feng,
Hongdong Li,
Qing Wang
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023
Project Page
/
arXiv
/
Code
this paper proposes a novel implicit camera model which represents the physical imaging process of a
camera as a deep neural network. We demonstrate the power of this new implicit camera model on two
inverse imaging tasks: i) generating all-in-focus photos, and ii) HDR imaging.
|
|
Local Implicit Ray Function for Generalizable Radiance Field Representation
Xin Huang,
Qi Zhang,
Ying Feng,
Xiaoyu Li,
Xuan Wang,
Qing Wang
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023
Project Page
/
arXiv
/
Code
We propose LIRF (Local Implicit Ray Function), a generalizable neural rendering approach for novel
view rendering. Given 3D positions within conical frustums, LIRF takes 3D coordinates and the
features of conical frustums as inputs and predicts a local volumetric radiance field.
|
CVPR 2023 (Highlight)
|
DINER: Disorder-Invariant Implicit Neural Representation
Shaowen Xie*,
Hao Zhu*,
Zhen Liu*,
Qi Zhang,
You Zhou,
Xun Cao,
Zhan Ma
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023
Project Page
/
arXiv
/
Code
In this paper, we find that such a frequency-related problem could be largely solved by re-arranging
the coordinates of the input signal, for which we propose the disorder-invariant implicit neural
representation (DINER) by augmenting a hash-table to a traditional INR backbone.
|
|
Local-to-Global Registration for Bundle-Adjusting Neural Radiance Fields
Yue Chen
Xingyu Chen,
Xuan Wang,
Qi Zhang,
Yu Guo,
Ying Shan,
Fei Wang
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023
Project Page
/
arXiv
/
Code
We propose L2G-NeRF, a Local-to-Global registration method for bundle-adjusting Neural Radiance
Fields, including the pixel-wise local alignment and the frame-wise global alignment.
|
|
UV Volumes for Real-time Rendering of Editable Free-view Human Performance
Yue Chen,
Xuan Wang*,
Xingyu Chen,
Qi Zhang,
Xiaoyu Li,
Yu Guo,
Jue Wang,
Fei Wang
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023
Project Page /
arXiv /
Code /
Video
We propose the UV Volumes, a new approach that can achieve real-time rendering, and editable NeRF,
decomposing a dynamic human into 3D UV Volumes and a 2D appearance texture.
|
CVPR 2023
|
Fine-Grained Face Swapping via Regional GAN Inversion
Zhian Liu*,
Maomao Li*,
Yong Zhang*,
Cairong Wang,
Qi Zhang,
Jue Wang,
Yongwei Nie
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023
Project Page
/
arXiv
/
Code
We present a novel paradigm for high-fidelity face swapping that faithfully preserves the desired
subtle geometry and texture details.
|
SIGGRAPH 2022
|
Neural Parameterization for Dynamic Human Head Editing
Li Ma,
Xiaoyu Li,
Jing Liao,
Xuan Wang,
Qi Zhang,
Jue Wang,
Pedro V. Sander
ACM Transactions on Graphics, 2022
Project Page
/
arXiv
/
Code
We try to introduce explicit parameters into implicit dynamic NeRF representations to achieve
editing of 3D human heads.
|
arXiv 2022
|
Stereo Unstructured Magnification: Multiple Homography Image for View Synthesis
Qi Zhang*,
Xin Huang*,
Ying Feng,,
Xue Wang,
Hongdong Li,
Qing Wang
arXiv, 2022
arXiv
We propose a novel multiple homography image (MHI) representation, comprising of a set of scene
planes with fixed normals and distances, for view synthesis from stereo images.
|
|
HDR-NeRF: High Dynamic Range Neural Radiance Fields
Xin Huang,
Qi Zhang,
Ying Feng,
Hongdong Li,
Xuan Wang,
Qing Wang
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022
Project Page
/
arXiv
/
Code
/
Dataset
/
video
We present High Dynamic Range Neural Radiance Fields (HDR-NeRF) to recover an HDR radiance field
from a set of low dynamic range (LDR) views with different exposures.
|
|
Hallucinated Neural Radiance Fields in the Wild
Xingyu Chen,
Qi Zhang,
Xiaoyu Li,
Yue Chen,
Ying Feng,
Xuan Wang,
Jue Wang
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022
Project Page
/
arXiv
This paper studies the problem of hallucinated NeRF: i.e. recovering a realistic NeRF at a different
time of day from a group of tourism images.
|
|
Deblur-NeRF: Neural Radiance Fields from Blurry Images
Li Ma,
Xiaoyu Li,
Jing Liao,
Qi Zhang,
Xuan Wang,
Jue Wang,
Pedro V. Sander
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022
Project Page
/
arXiv
In this paper, we propose Deblur-NeRF, the first method that can recover a sharp NeRF from blurry
input. A novel Deformable Sparse Kernel (DSK) module is presented for both camera motion blur and
defocus blur.
|
CVPR 2022
|
FENeRF: Face Editing in Neural Radiance Fields
Jingxiang Sun,
Xuan Wang,
Yong Zhang,
Xiaoyu Li,
Qi Zhang,
Yebin Liu,
Jue Wang
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022
Project Page
/
arXiv
A 3D-aware generator (FENeRF) is prorposed to produce view-consistent and locally-editable portrait
images.
|
TPAMI 2022
|
Ray-Space Epipolar Geometry for Light Field Cameras
Qi Zhang,
Qing Wang,
Hongdong Li,
Jingyi Yu
IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2022
PDF
/
bibtex
This paper fills in this gap by developing a novel ray-space epipolar geometry which intrinsically
encapsulates the complete projective relationship between two light fields.
Ray-space fundamental matrix and its properties are then derived to constrain ray-ray
correspondences for general and special motions.
|
IJCV 2021
IJCV 2021
|
3D Scene Reconstruction with an Un-calibrated Light Field Camera
Qi Zhang,
Hongdong Li,
Xue Wang,
Qing Wang
International Journal of Computer Vision (IJCV), 2021
PDF
/
bibtex
This paper is concerned with the problem of multi-view 3D reconstruction with an un-calibrated
micro-lens array based light field camera..
|
MTA 2021
|
Region-based Depth Feature Descriptor for Saliency Detection on Light Field
Xue Wang,
Yingying Dong,
Qi Zhang,
Qing Wang
Multimedia Tools and Applications, 2021
PDF
/
bibtex
By reinterpreting the usage of dark channels in estimating the amount of defocus, a novel
region-based depth feature descriptor (RDFD) defined over the focal stack is proposed.
|
|
Full View Optical Flow Estimation Leveraged From Light Field Superpixel
Hao Zhu,
Xiaoming Sun,
Qi Zhang,
Qing Wang,
Antonio Robles-Kelly,
Hongdong Li,
Shaodi You
IEEE Transactions on Computational Imaging, 2019
PDF
/
bibtex
Our method employs the structure delivered by the four-dimensional light field over multiple views
making use of superpixels for a full view optical flow estiamtion.
|
CVPR 2019
CVPR 2019
|
Ray-Space Projection Model for Light Field Camera
Qi Zhang,
Jinbo
Ling,
Qing Wang,
Jingyi Yu
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019
PDF
/
Supp
/
bibtex
In the paper, we propose a novel ray-space projection model to transform sets of rays captured by
multiple light field cameras in term of the Plucker coordinates.
|
TIP 2019
|
4D Light Field Superpixel and Segmentation
Hao Zhu,
Qi Zhang,
Qing Wang,
Hongdong Li
IEEE Transactions on Image Processing (TIP), 2019
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2017
PDF
/
bibtex
The light field superpixel (LFSP) is first defined mathematically and then a refocus-invariant
metric named LFSP self-similarity is proposed to evaluate the segmentation performance.
|
ACCV 2018
|
Common Self-polar Triangle of Concentric Conics for Light Field Camera Calibration
Qi Zhang,
Qing Wang
Asian Conference on Computer Vision, 2018
PDF
/
Supp
/
bibtex
Instead of a planar checkerboard, we propose to calibrate light field camera using a concentric
conics pattern based on the property and reconstruction of common self-polar triangle with respect
to concentric circle and ellipse..
|
TPAMI 2019
TPAMI 2019
|
A Generic Multi-Projection-Center Model and Calibration Method for Light Field Camera
Qi Zhang,
Chunping Zhang,
Jinbo Ling,
Qing Wang,
Jingyi Yu
IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2019
arXiv
/
PDF
/
Code
/
bibtex
The MPC model can generally parameterize light field in different imaging formations, including
conventional and focused light field cameras.
|
Other Publications
Qi Zhang, Qing Wang. Common self-polar triangle of separate circles for light
field
camera calibration[J]. Xibei Gongye Daxue Xuebao/Journal of Northwestern Polytechnical
University,
2021.
Yaning Li, Qi Zhang, Xue Wang, Qing Wang. Light Field SLAM based on Ray-Space
Projection Model[C]. Optoelectronic Imaging and Multimedia Technology VI, 2019.
Qi Zhang, Xue Wang, Qing Wang. Light Field Planar Homography and Its
Application[C]. Optoelectronic Imaging and Multimedia Technology VI, 2019.
Zhao Ren, Qi Zhang, Hao Zhu, Qing Wang. Extending the FOV from disparity and
color consistencies in multiview light fields[C]. IEEE International Conference on Image
Processing (ICIP), 2017
|
This template is a modification to Jon Barron's website.
Feel free to clone it for your own use while attributing the original author Jon Barron.
|
|