At Adobe, I work on computer vision for Imaging products. I am the primary contributor to several features including Instruction-based Image Editing, Select Subject, Object Finder, and Select People Details. I earned my Ph.D. at Arizona State University, advised by Baoxin Li. I have contributed to 8 tech transfers to products including Photoshop, Lightroom, and Stardust.
Now recruiting summer research interns in image/video editing with MLLM.
I strive for simple yet scalable methods in image understanding and editing. Representative works are highlighted below. Full list available on Google Scholar.
CVPR 2025
OmniStyle is the first end-to-end style transfer framework based on the Diffusion Transformer (DiT) architecture, achieving high-quality 1K-resolution stylization by leveraging the large-scale, filtered OmniStyle-1M dataset. It supports both instruction- and image-guided stylization, enabling efficient and versatile style transfer across diverse styles.
CVPR 2025
A unified vision-language model for free-form mask grounding and compositional captioning.
2025 (Highlight)
pdf/
project page
Foundaitional multi-modal generative model UniReal is a universal framework for multiple image generation and editing tasks. We leverage a video model to handld image tasks by treating different numbers of input/output images as frames. We also seek universal supervisions from video data, thus generating realistic results that understand the world dynamics.
AAAI 2025
Sigstyle is a style preserved style transfer method via personalized subject editing diffusion model.
ECCV 2024
Jing Gu, Nanxuan Zhao, Wei Xiong, Qing Liu, Zhifei Zhang, He Zhang, Jianming Zhang, HyunJoon Jung, Yilin Wang★, Xin Eric Wang★
★ Co-advisor
A method for personalized subject driven image editing.
CVPR 2024
Nannan Li, Qing Liu, Krishna Kumar Singh, Yilin Wang, Jianming Zhang, Bryan A. Plummer, Zhe Lin
Human editing via diffusion.
AAAI (oral) 2024
Bowen Zhang, Qing Liu, Jianming Zhang, Yilin Wang, Akide Liu, Zhe Lin, Yifan Liu
Amodal segmentation considers mutual occlusion.
NeurIPS 2023
Jing Gu, Yilin Wang, Nanxuan Zhao, Tsu-Jui Fu, Wei Xiong, Qing Liu, Zhifei Zhang, He Zhang, Jianming Zhang, HyunJoon Jung, Xin Eric Wang
A method for personalized subject driven image editing.
CVPR 2023
Yiqun Mei, He Zhang, Xuaner Zhang, Jianming Zhang, Zhixin Shu, Yilin Wang, Zijun Wei, Yan Shi, HyunJoon Jung, Vishal M. Patel
A scribble-based relighting system that allows users to interactively manipulate portrait lighting effects with ease.
ICLR 2023
Jeya Maria Jose Valanarasu, He Zhang, Jianming Zhang, Yilin Wang, Zhe Lin, Yinglan Ma, Zijun Wei, Kalyan Sunkavalli, Vishal M. Patel
Interactive harmonization for portrait photos.
CVPR 2022
Chenglin Yang, Yilin Wang, Jianming Zhang, He Zhang, Zijun Wei, Zhe Lin, Alan Yuille
Light-weight vision transformer models for vision tasks.
ICCV 2021
Yifan Jiang, He Zhang, Jianming Zhang, Yilin Wang, Zhe Lin, Kalyan Sunkavalli, Simon Chen, Sohrab Amirghodsi, Sarah Kong, Zhangyang Wang
Image harmonization based on self-supervised learning.
CVPR 2021
Qihang Yu, Jianming Zhang, He Zhang, Yilin Wang, Zhe Lin, Ning Xu, Yutong Bai, Alan Yuille
Mask guided image matting.
CVPR 2021
Xin Yuan, Zhe Lin, Jason Kuen, Jianming Zhang, Yilin Wang, Michael Maire, Ajinkya Kale, Baldo Faieta
Intra- and inter-modal similarity preservation for multimodal representation learning.
ECCV 2020
Kenan E. Ak, Ning Xu, Zhe Lin, Yilin Wang
ECCV 2020
Shikun Liu, Zhe Lin, Yilin Wang, Jianming Zhang, Federico Perazzi, Edward Johns
ICCV 2019
Yulun Zhang, Chen Fang, Yilin Wang, Zhaowen Wang, Zhe Lin, Yun Fu, Jimei Yang