🧍‍♂️ Biography

Ge-Peng Ji is a final-year PhD candidate at the School of Computing, Australian National University, advised by Professor Nick Barnes. Prior to this, he received his Master’s degree from the School of Computer Science at Wuhan University in 2021. He has published over 30 peer-reviewed academic papers, with more than 10,000+ citations and an h-index of 24, and holds six Chinese technical patents. He regularly serves as a reviewer for top-tier AI journals and conferences, including TPAMI, IJCV, TMI, CVPR, ICCV, and MICCAI. He was named to the Stanford/Elsevier Top 2% Scientists List in both 2024 and 2025.

👨‍💻 Research Interests

His research centers on subtle visual perception (微视觉感知), aiming to model hard-to-detect patterns – often imperceptible to human vision yet semantically meaningful – in complex environments using advanced AI techniques, including computer vision, multimodal learning, and reinforcement learning. Empowering such perceptual capabilities in intelligent systems has broad real-world implications, including 1️⃣ healthcare AI, where early identification of subtle anomalies in medical imaging can enable timely and potentially life-saving interventions; 2️⃣ camouflaged scene understanding, where objects blend into their surroundings due to low contrast and unclear boundaries; and 3️⃣ ultra-precision vision applications, where identifying ultra-tiny objects or subtle structures requires fine-grained representation and accurate localization.

🔥 News

🚨
NOTE
I am currently on the academic job market and open to postdoctoral and research positions. Please feel free to reach out if you think my background may be a good fit for your group. I am also open to interdisciplinary collaborations and welcome connections with researchers from diverse backgrounds, especially in medical applications.


  • 2026.03 We are excited to launch the Colon-X project, an open initiative aimed at advancing intelligent colonoscopy toward clinical reasoning.
  • 2026.01 Our paper “Frontiers in Intelligent Colonoscopy” is now officially available on Springer platform. Code can be found on our GitHub repository.

📝 Representative Publications

^ indicates equal contribution, and * denotes corresponding author. For a full list of publications, please visit my 🎓Google Scholar profile.

🚩 [Research Topic #1] Healthcare AI

[MICCAI 2020] PraNet: Parallel Reverse Attention Network for Polyp Segmentation

Deng-Ping Fan, Ge-Peng Ji, Tao Zhou, Geng Chen, Huazhu Fu*, Jianbing Shen*, Ling Shao, and Ali Borji
Medical Image Computing and Computer Assisted Intervention, Virtual, October 4-8, 2020

Links: 📄Paper (arXiv; Springer; 中译文) | 📦Project Page | 📦Huawei Ascend ModelZoo | 🎬Technical Video at MedicoEval'2020 Workshop

Keywords: #polyp-segmentation, #intelligent-colonoscopy

TL;DR: A pioneering work that introduced a reverse attention mechanism for early lesion detection (particularly effective for weak boundaries), now widely recognized as a standard baseline in medical image segmentation.

Impact: Early acceptance & Oral Presentation (Top 13%) | MICCAI2025 Young Scientist Publication Impact Award (see certificate & award ceremony photo) | #1 most-cited MICCAI paper (by Google Scholar Metrics 2025) | Ranked #1 in accuracy at the MediaEval 2020 Medico Task | Most Influential Application Paper Award at the Jittor Developer Conference 2021 | Featured in the Stanford AI Index Report 2022

[MICCAI 2021] Progressively Normalized Self-Attention Network for Video Polyp Segmentation

Ge-Peng Ji^, Yu-Cheng Chou^, Deng-Ping Fan*, Geng Chen, Huazhu Fu*, Debesh Jha, Ling Shao
Medical Image Computing and Computer Assisted Intervention, Strasbourg, France, September 27-October 1, 2021

Links: 📄Paper (arXiv; Springer; 中译文) | 📦Project Page | 🎬Introduction Video (YouTube)

Keywords: #video-polyp-segmentation, #intelligent-colonoscopy

TL;DR: We propose a new self-attention mechanism for video polyp segmentation -- an extensible, plug-and-play design that delivers ultra-fast inference (170+ FPS on a single RTX 2080 GPU).

Impact: Early acceptance (Top 13%) | MICCAI 2021 Student Travel Award (Top 5.8% = 95/1630)

[MIR 2022] Video Polyp Segmentation: A Deep Learning Perspective

Ge-Peng Ji^, Guobao Xiao^, Yu-Cheng Chou^, Deng-Ping Fan*, Kai Zhao, Geng Chen, and Luc Van Gool
Machine Intelligence Research, 2022, 19 (6): 531-549. (IF: 8.7; 中国最具国际影响力学术期刊)

Keywords: #video-benchmark, #video-polyp-segmentation, #intelligent-colonoscopy

Links: 📄Paper (arXiv; Springer; 中译文) | 📦Project Page | 🎬Introduction Video (~2min)

TL;DR: A comprehensive benchmark and systematic study for video polyp segmentation, providing standardized evaluation and insights for future research.

[MIR 2026] Frontiers in Intelligent Colonoscopy

Ge-Peng Ji, Jingyi Liu, Peng Xu, Nick Barnes, Fahad Shahbaz Khan, Salman Khan, Deng-Ping Fan*
Machine Intelligence Research, 2026, 23 (1), 70-114. (IF: 8.7; 中国最具国际影响力学术期刊)

Keywords: #survey, #multimodal-benchmark, #multimodal-large-language-model, #intelligent-colonoscopy

Links: 📄Paper (arXiv; Springer; 中译文) | 📦Project Page | 🗂️ColonSurvey | 🤗ColonINST | 🤗ColonGPT

TL;DR: A comprehensive survey of intelligent colonoscopy, covering the latest advances in multimodal benchmarks and large language models, and providing insights into future research directions.

[arXiv 2026] Colon-X: Advancing Intelligent Colonoscopy toward Clinical Reasoning

Ge-Peng Ji, Jingyi Liu, Deng-Ping Fan*, Huazhu Fu, Nick Barnes

Keywords: #multimodal-benchmark, #multimodal-large-language-model, #multimodal-reasoning, #intelligent-colonoscopy

Links: 📄Paper (arXiv) | 📦Project Page

TL;DR: Introducing COLON-X, a multimodal benchmark designed to advance intelligent colonoscopy toward clinical reasoning, featuring comprehensive datasets and evaluation protocols for multimodal learning and reasoning in colonoscopy.

🚩 [Research Topic #2] Camouflaged Scene Understanding

Camouflaged scene understanding focuses on the perception and analysis of objects that blend into their surroundings, making them difficult to detect due to low contrast and unclear boundaries. This research area is crucial for applications such as wildlife monitoring, search and rescue operations, and military surveillance. The core challenge lies in developing models that can effectively capture subtle visual cues and contextual information to accurately identify camouflaged objects in complex environments.

[TPAMI 2021] Concealed Object Detection (extended version of CVPR 2020)

Deng-Ping Fan, Ge-Peng Ji, Ming-Ming Cheng*, and Ling Shao
IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022, 44 (10): 6024-6042. (IF: 18.6)

Keywords: #image-benchmark, #camouflaged-object-detection, #concealed-object-detection

Links: 📄TPAMI 2021 Journal Paper (arXiv; IEEE Xplore; GitHub; Supplementary Material; 中译文) | 📄CVPR 2020 Conference Paper (CVF Open Access; Github; 中译文) | 📦Project Page | 📦Download COD10K dataset | 🎮Online Playground

TL;DR: Introducing COD10K, a large-scale benchmark for camouflaged object detection, featuring comprehensive datasets and evaluation protocols for advancing the field of camouflage-aware computer vision.

Impact: CVPR Oral Paper (Top 5.7% = 335/5865) | ESI Hot Paper (Top 0.1%) | Distinguish Paper at the Jittor Developer Conference 2021 | Covered by "New Scientist" magazine (see the media snapshot) | 推动国产Jittor生态体系建设

[MIR 2023] Deep Gradient Learning for Efficient Camouflaged Object Detection

Ge-Peng Ji, Deng-Ping Fan*, Yu-Cheng Chou, Dengxin Dai, Alexander Liniger, Luc Van Gool
Machine Intelligence Research, 2023, 20 (1): 92-108. (IF: 8.7; JCR Q1; 中国最具国际影响力学术期刊)

Keywords: #efficient-camouflaged-object-segmentation, #polyp-segmentation, #industrial-defect-segmentation

Links: 📄Paper (arXiv; Springer) | 📦Project Page (supporting PyTorch/Jittor/Ascend platforms) | 🎬Introduction video (~2min)

TL;DR: Introducing DGNet (Deep Gradient Network), a novel approach for efficient camouflaged object detection that leverages gradient-based learning to enhance detection performance while significantly reducing computational overhead.

[SCIS 2023] SAM Struggles in Concealed Scenes -- Empirical Study on Segment Anything

Ge-Peng Ji, Deng-Ping Fan, Peng Xu*, Ming-Ming Cheng, Bowen Zhou, Luc Van Gool
SCIENCE CHINA Information Sciences, 2023, 66: 226101. (IF: 7.6; CiteScore: 12.6; CCF-A; 中国最具国际影响力学术期刊)

Keywords: #image-benchmark, #promptable-segmentation, #camouflaged-object-segmentation, #polyp-segmentation, #industrial-defect-segmentation

Links: 📄Paper (arXiv; Springer) | 📰获《中国科学信息科学》官方微信公众号专题解读

TL;DR: SAM evaluation on concealed scenes, revealing significant performance gaps and highlighting the need for specialized approaches in camouflaged object detection.

🚩 [Research Topic #3] Ultra-Precision Vision Applications

In real-world vision systems such as autonomous driving (e.g., distant traffic signs and lane markings), image matting (e.g., hair-level boundary extraction), and remote sensing (e.g., tiny object detection in high-resolution imagery), even pixel-level inaccuracies can lead to critical failures. Our goal is to unlock ultra-precision representations in complex vision systems, enabling precise structure delineation and reliable tiny object localization.

[ICCV 2021] Full-Duplex Strategy for Video Object Segmentation

Ge-Peng Ji, Deng-Ping Fan*, Keren Fu, Zhe Wu, Jianbing Shen, Ling Shao
International Conference on Computer Vision, Virtual, October 11-17, 2021

Keywords: #video-object-segmentation, #video-salient-object-detection

Links: 📄ICCV 2021 Conference Paper (arXiv; IEEE Xplore; 中译文) | 📄CVMJ 2023 Journal Extension (IF: 18.3; 中国最具国际影响力学术期刊; arXiv; IEEE Xplore) | 📦Project Page

TL;DR: A pioneering work that introduces a full-duplex learning strategy to capture static–dynamic interactions, enabling precise segmentation of moving objects in dynamic scenes.

Impact: Honorable Mention Award at CVMJ 2023 (see the certificate)

[ICCV 2025] LawDIS: Language-Window-based Controllable Dichotomous Image Segmentation

Xinyu Yan, Meijun Sun, Ge-Peng Ji, Fahad Shahbaz Khan, Salman Khan, Deng-Ping Fan*
International Conference on Computer Vision, Honolulu, Hawaii, USA, October 19-23, 2025

Keywords: #promptable-segmentation, #stable-diffusion, #multimodal-model, #dischotomous-image-segmentation

Links: 📄Paper (arXiv; CVF Open Access; 中译文) | 📦Project Page | 🎬introduction video (~1min)

TL;DR: Describe, segment, and post-refine! LawDIS enables human users to precisely segment objects with simple language instructions, and interactively post-refine user-specified regions of arbitrary size.

💻 Working Experience

  • 2023.03 - 2023.09 Visiting Scholar, Machine Learning Department, Mohamed bin Zayed University of Artificial Intelligence (MBZUAI), Abu Dhabi, UAE.
  • 2022.04 - 2022.07 Research Assistant, Sensing Intelligence and Machine Learning Laboratory (SIGMA Lab), Wuhan University, Wuhan, China.
  • 2021.06 - 2022.04 Research Intern (Talent Program), Alibaba Group (ICBU Technology Department), Hangzhou, China.
  • 2020.11 - 2021.04 Research Intern, Inception Institute of Artificial Intelligence (IIAI-CV&Med Team), Abu Dhabi, UAE.

💬 Invited Talks

  • 2026.03.19 ANU Seminar Talk “Multimodal Intelligence in Colonoscopy: Perception, Understanding, and Reasoning”
  • 2023.09.23 国际图象图形学学术会议(ICIG 2023) - 多模态数据感知与学习Workshop大会报告 “Towards AI-Powered Colonoscopy” (Page1/Page2)
  • 2023.07.06 机器智能前沿论坛·第2期 “伪装场景感知及多模态应用: 一种基于梯度先验信息学习的高效伪装目标分割方法” (Page & Recorded Video)
  • 2023.03.31 中国图象图形学学会第三期学生会员分享论坛会议 “Towards AI-Powered Colonoscopy” (Page)
  • 2023.01.06 Shenzhen University Talking 2023: “Towards AI-Powered Colonoscopy”
  • 2022.12.07 Anhui University Talking 2022: “Colonoscopy in the AI Era”
  • 2022.10.10 VALSE 2021 Workshop 6 – Spotlight Talking: “Camouflaged Object Detection and its Applications” (Poster)
  • 2021.11.15 Synced-ICCV 2021 “Full-Duplex Strategy for Video Object Segmentation”
  • 2021.08.28 CSIG-ICCV 2021 “Full-Duplex Strategy for Video Object Segmentation” (Page)
  • 2021.05.15 Alibaba Group ICBU Talking 2021: “Camouflaged Object Detection in the Deep Learning Era”

🏆 Honors and Awards

  • 2025 MICCAI Young Scientist Publication Impact Award (Top ~0.2% from 2020-2024 accepted papers, Link)
  • 2025 CVPR 2025 Outstanding Reviewer Award (Top 5.6%=711/12000+, Link)
  • 2025 Stanford/Elsevier Top 2% Scientists List 2025 (Link)
  • 2024 MICCAI2024 Young Scientist Publication Impact Award Shortlist
  • 2024 Stanford/Elsevier Top 2% Scientists List 2024 (Link)
  • 2024 CVM Honorable Paper Mention Award in 2023 (Journal Updates)
  • 2024 MIR Outstanding Reviewer Award in 2023 (MIR Journal News)
  • 2023 IEEE Transactions on Medical Imaging Distinguished Reviewer (Bronze Level 2022 – 2023)
  • 2021 Jittor Developer Conference Distinguished Paper & Most Influential (Application) Paper
  • 2021 MICCAI Student Travel Award (MICCAI link)
  • 2020 Multimedia Evaluation Benchmark Workshop (Ranking Top-1 in terms of accuracy metric)
  • 2019 Sparse Representation and Intelligent Analysis Competition of Remote Sensing Images (Co-organized by Huawei Co., LTD \& Wuhan University), Object Detection Track (Ranking: Top 6%).

📃 Academic Services

Program Committee Member & Conference Reviewer

Journal Reviewer

  • IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)
  • International Journal of Computer Vision (IJCV)
  • IEEE Transactions on Image Processing (TIP)
  • IEEE Journal of Biomedical and Health Informatics (JBHI)
  • IEEE Transactions on Medical Imaging (TMI)
  • IEEE Transactions on Visualization and Computer Graphics (TVCG)
  • IEEE Transactions on Circuits and Systems for Video Technology (TCSVT)
  • IEEE Transactions on Multimedia (TMM)
  • Computational Visual Media (CVM)
  • Machine Intelligence Research (MIR)
  • Information Fusion (IF)
  • Neurocomputing (NC)
  • Computer Vision and Image Understanding (CVIU)
  • Scientific Reports (Link)
  • Expert Systems with Applications (ESWA)
  • Digital Signal Processing (DSP)
  • Sensors (Link)
  • Signal Processing: Image Communication (SPIC)
  • The Visual Computer (TVC)
  • International Journal of Imaging Systems and Technology (IJIST)
  • Diagnostics (Link)
  • Biocybernetics and Biomedical Engineering (BBE)

🔗 Useful Resources