Kai Wang

I'm now working as a PostDoc in LAMP group at Computer Vision Center (CVC), UAB. Before that I graduate as a PhD under the supervision of Joost van de Weijer in 2022. I received my master degree in image processing from Jilin University in 2017 and the bachelor degree from Jilin University in 2014. I have worked on a wide variety of projects including Diffusion Models, Continual Learning and Vision Transformers. Now I am mainly working on multiple projects on diffusion models and supervising several PhD students on related topics.

Email  /  Google Scholar  /  LinkedIn  /  Twitter  /  Github  /  Awesome Diffusion

Publications
LocInv: Localization-aware Inversion for Text-Guided Image Editing
Chuanming Tang, Kai Wang*, Fei Yang, Joost van de Weijer.
CVPR AI4CC workshop, 2024
project page / arXiv

We address the cross-attention leakage problem in Text-to-Image diffusion model inversions by introducing the localization priors.

Diffusion-based network for unsupervised landmark detection
Tao Wu, Kai Wang*, Chuanming Tang, Jianlin Zhang.
Knowledge-Based Systems, 2024
Journal

We propose a novel diffusion-based network (DBN) for unsupervised landmark detection, which leverages the generation ability of the diffusion models to detect the landmark locations.

IterInv: Iterative Inversion for Pixel-Level T2I Models
Chuanming Tang, Kai Wang*, Joost van de Weijer.
ICME, 2024 & NeurIPS 2023 workshop on Diffusion Models
project page / arXiv

We develop an iterative inversion (IterInv) technique for the pixel level T2I models and verify IterInv with the open-source DeepFloyd-IF model.

Plasticity-Optimized Complementary Networks for Unsupervised Continual Learning
Alex Gomez-Villa, Bartlomiej Twardowski, Kai Wang*, Joost van de Weijer.
WACV, 2024
project page / arXiv

In this continual learning paper, we propose to train an expert network that is relieved of the duty of keeping the previous knowledge and can focus on performing optimally on the new tasks (optimizing plasticity).

Dynamic Prompt Learning: Addressing Cross-Attention Leakage for Text-Based Image Editing
Kai Wang*, Fei Yang, Shiqi Yang, Muhammad Atif Butt, Joost van de Weijer.
NeurIPS, 2023
project page / arXiv

In this paper, we propose Dynamic Prompt Learning (DPL) to force cross-attention maps to focus on correct noun words in the text prompt. By updating the dynamic tokens for nouns in the textual input with the proposed leakage repairment losses, we achieve fine-grained image editing over particular objects while preventing undesired changes to other image regions.

ScrollNet: Dynamic Weight Importance for Continual Learning
Fei Yang, Kai Wang*, Joost van de Weijer.
ICCV VCL workshop, 2023
project page / arXiv

In this paper, we propose ScrollNet as a scrolling neural network for continual learning. ScrollNet can be seen as a dynamic network that assigns the ranking of weight importance for each task before data exposure, thus achieving a more favorable stability-plasticity tradeoff during sequential task learning by reassigning this ranking for different tasks.

One ring to bring them all: Towards open-set recognition under domain shift
Shiqi Yang, Yaxing Wang, Kai Wang*, Shangling Jui, Joost van de Weijer.
Under Review, 2023
project page / arXiv

In this paper, we investigate Source-free Open-partial Domain Adaptation (SF-OPDA), which addresses the situation where there exist both domain and category shifts between source and target domains.

Positive Pair Distillation Considered Harmful: Continual Meta Metric Learning for Lifelong Object Re-Identification
Kai Wang*, Chenshen Wu, Andrew D. Bagdanov, Xialei Liu, Shiqi Yang, Joost van de Weijer.
BMVC, 2022
project page / arXiv

To address the incremental Person ReID problem, we apply continual meta metric learning to lifelong object re-identification. To prevent forgetting of previous tasks, we use knowledge distillation and explore the roles of positive and negative pairs. Based on our observation that the distillation and metric losses are antagonistic, we propose to remove positive pairs from distillation to robustify model updates.

Attention Distillation: self-supervised vision transformer students need more guidance
Kai Wang*, Fei Yang, Joost van de Weijer.
BMVC, 2022
project page / arXiv

In this paper, we study knowledge distillation of self-supervised vision transformers (ViT-SSKD). We show that directly distilling information from the crucial attention mechanism from teacher to student can significantly narrow the performance gap between both. .

Attracting and Dispersing: A Simple Approach for Source-free Domain Adaptation
Shiqi Yang, Kai Wang*, Yaxing Wang, Shangling Rui, Joost van de Weijer.
NeurIPS (Spotlight), 2022
project page / arXiv

We propose a simple but effective source-free domain adaptation (SFDA) method. Treating SFDA as an unsupervised clustering problem and following the intuition that local neighbors in feature space should have more similar predictions than other features, we propose to optimize an objective of prediction consistency. This objective encourages local neighborhood features in feature space to have similar predictions while features farther away in feature space have dissimilar predictions, leading to efficient feature clustering and cluster assignment simultaneously.

HCV: Hierarchy-Consistency Verification for Incremental Implicitly-Refined Classification
Kai Wang*, Xialei Liu, Luis Herranz, Joost van de Weijer.
BMVC, 2021
project page / arXiv

Current incremental learning methods lack the ability to build a concept hierarchy by associating new concepts to old ones. A more realistic setting tackling this problem is referred to as Incremental Implicitly-Refined Classification (IIRC), which simulates the recognition process from coarse-grained categories to fine-grained categories. To overcome forgetting in this benchmark, we propose Hierarchy-Consistency Verification (HCV) as an enhancement to existing continual learning methods.

ACAE-REMIND for Online Continual Learning with Compressed Feature Replay
Kai Wang*, Luis Herranz, Joost van de Weijer.
Pattern Recognition Letters, 2021
arXiv

We propose an auxiliary classifier auto-encoder (ACAE) module for feature replay at intermediate layers with high compression rates. The reduced memory footprint per image allows us to save more exemplars for replay.

On Implicit Attribute Localization for Generalized Zero-Shot Learning
Shiqi Yang, Kai Wang*, Luis Herranz, Joost van de Weijer.
IEEE Signal Processing Letters, 2021
arXiv

In this paper, we show that common ZSL backbones (without explicit attention nor part detection) can implicitly localize attributes, yet this property is not exploited.

Semantic Drift Compensation for Class-Incremental Learning
Lu Yu, Bartlomiej Twardowski, Xialei Liu, Luis Herranz, Kai Wang*, Yongmei Cheng, Shangling Rui, Joost van de Weijer.
CVPR, 2020
project page / arXiv

Embedding networks have the advantage that new classes can be naturally included into the network without adding new weights. Therefore, we study incremental learning for embedding networks. In addition, we propose a new method to estimate the drift, called semantic drift, of features and compensate for it without the need of any exemplars. We approximate the drift of previous tasks based on the drift that is experienced by current task data.

Education
Universitat Autònoma de Barcelona
Postdoctoral Researcher
Supervised by Prof. Joost van de Weijer
2022-2024
Universitat Autònoma de Barcelona
Ph.D. student
Supervised by Prof. Joost van de Weijer
2017-2022
Jilin University
Undergraduate student and Master student
2010-2017