JINRUI
profile photo

Jinrui Zhang

I am currently a postdoctoral researcher advised by Prof. Ju Ren in the Department of Computer Science and Technology, Tsinghua University. I received his B.Sc., M.Sc., Ph.D. degrees all in computer science, from Central South University, China, in 2016, 2018 and 2023, co-advised by Prof. Yaoxue Zhang and Prof. Deyu Zhang. I was a Visiting PhD Student at HCS Lab, Seoul National University in 2022-2023, working with Prof. Youngki Lee. I also joined HEX group at Microsoft Research Asia from Sept. 2019 to Mar. 2020 and Institute for AI Industry Research(AIR), Tsinghua University from Aug. 2021 to Apr. 2022 as a research intern, respectively, mentored by Prof. Yunxin Liu.

Email        Google Scholar        Github       CV(last updated: 2022.06)

News

Researchs

My research interests on mobile computing and edge computing, including AR/VR, intelligent edge/mobile systems, mobile sensing and applications of machine learning.

  • Depth Estimation using On-deivce Dual Camerasbas

  • The system overview and workflow of MobiDepth.


    Example depth maps generated by MobiDepth, AnyNet, MADNet
    with and without online adaptation (named MADNet-MAD and MADNet-No,
    respectively), and ARCore, with the person sitting, walking and standing.

    The initial idea was came up in about 2021 when dining with some friends at Beijing. While binocular depth estimation is a mature technique, it is challenging to realize the technique on commodity mobile devices due to the different focal lengths and unsynchronized frame flows of the on-device dual cameras and the heavy stereo-matching algorithm. Even the ARCore and the ARKit are the most common solution adopted by existing mobile systems, they requrie the camera to be moving and expects the target object to be stationary, which significantly restricts its usage scenarios. To this end, we propose MobiDepth, a real-time depth estimation system using the widely-available on-device dual cameras. MobiDepth resolves all the issues of the three existing solutions, i.e., it does not rely on any dedicated sensors or pre-training, and works well for target objects in motion.

  • MobiPose/MVPose : Realtime Multi-Person Pose Estimation on Mobile Devices

  • The system architecture and workflow of MVPose.


    Example results of MVPose from live videos.

    The initial verison of MobiPose was accepted by ACM SenSys 2020, which achieved over 20 frames per second pose estimation with 3 persons per frame, and significantly outperforms the state-of-the-art baseline(MobineNetV3-SSDLite+PoseNet), with a speedup of up to 4.5 and 2.8 in latency on CPU andGPU, respectively, and an improvement of 5.1% in pose-estimation model accuracy.

    The second version, MVPose took a motion-vector-based approach to fast and accurately track the human keypoints across consecutive frames, rather than running expensive human-detection model and pose-estimation model for every frame, which achieved over 30 frames per second pose estimation with 4 persons per frame.

    Publications

    Conference Papers:

    Journal Papers:

    Patents

    Services

    Experiences

    Awards

    Honors