Visual and Geometric Perception Lab
The Vision and Geometric Perception Laboratory (VGPL), of the School of Computer Science and Technology at Donghua University, was established in 2013 by Dr. Shen Cai. It focuses on theoretical and applied research in Computer Vision, Computer Graphics, and Artificial Intelligence, particularly in 3D tasks using geometric constraints and deep learning methods. Key research areas include camera calibration, pose estimation, image-based 3D reconstruction, robot navigation, feature extraction and matching, 3D object recognition, neural implicit reconstruction, and various 3D representations. The lab also collaborates with companies on industrial vision inspection, action recognition, object detection/segmentation, AR/VR, 3D registration, and robot grasping. Currently, there are 12 graduate students and 1 intern in the lab, with 16 students and 6 interns having graduated so far.
Directions in the field of Computer Vision (CV)
Camera Calibration aims to calculate the internal and external parameters of a camera and lens distortion. Research in this direciton covers pattern design, feature extraction, homography computation based on 2D-2D point correspondences, lens distortion model selection, and the joint optimization of internal and external parameters. Our lab has conducted theoretical and applied research in many areas, such as fast and interpretable homography decomposition, deep homography estimation, feature correspondence, calibration using conic curves and hybrid primitives, rapid multi-camera calibration, robot-based calibration, depth camera calibration, and zoom calibration. We have published a number of academic papers and applied our findings to various company projects.
Pose Estimation aims to calculate the Euclidean transformation (rotation and translation) between 3D coordinate systems. It often uses 3D-2D point pairs to estimate the camera's absolute pose from a single image or uses 2D-2D point pairs to estimate relative pose from two images. Our lab has completed an in-depth research on perspective-three-point (P3P) problem.
Image-based 3D Reconstruction aims to reconstruct scenes or objects from images or RGBD data. Its research includes feature extraction, pose estimation, 3D representation, and the joint optimization of external parameters and 3D points. Our lab has explored multiple directions such as depth camera based fusion reconstruction, multi-view stereo, and robot navigation.
Directions in the field of Computer Graphics (CG)
3D Representation is a key research topic in graphics. To meet the needs of different tasks, 3D representations often need to be switched or combined. Our lab focuses on using orthogonal distance fields and spherical primitives to accomplish various 3D representation tasks. These include fast and accurate neural implicit representation based on orthogonal distance fields, concise inner(spherical) outer-ball representation, concise spherical - node - graph representation, double - layer spherical - shell representation, and hierarchical shell representation. We've published several academic papers in these areas.
3D Object Recognition and Segmentation involve classifying and segmenting known 3D models. Our lab has published multiple papers on deep classification networks based on spherical projection, classification networks based on spatial key - sphere representation, classification networks based on spherical - node graphs, and part segmentation based on key spheres.
3D Point Cloud Registration aims to quickly and accurately match a known 3D model in a scene and estimate relative pose using matched 3D-3D point pairs. Our lab has conducted research in 3D feature extraction, random sample consensus, and robotic grasping, applying these findings to various corporate projects.

0