I am currently a Ph.D. student in the Department of Computer Science at City University of Hong Kong supervised by Prof. Shiqi Wang. From Mar. 2024 to Aug. 2024, I was a visiting scholar at The University of Tokyo, under the supervision of Prof. Lei Ma. Before that, I received my B.E. degree in School of Computer Science and Technology from Shandong University with first class honours in 2020.
Research: I have broad interests in Trustworthy machine learning, Theoretical interpretation of deep learning models, and Out-of-distribution detection and generalization. I am enthusiastic about understanding the internal workings of machine learning algorithms and designing tools to make them explainable and robust.
Mis: I love backpacking and adventuring. In July 2024, I felt deeply depressed about my research. Each day, I’d wake up, head to the university, and go through a monotonous routine, finding it harder and harder to appreciate the wonderful things around me. Seeking a change, I set out on my first solo trip, traveling across Japan – from the southern warmth of Okinawa and Kyushu to the northern charm of Hokkaido, with stops in Kansai and Kanto along the way. I met incredible people from different countries and walks of life. They lifted my spirits in ways I hadn’t expected and helped me find myself. I’m grateful for this experience.
news
Nov 4, 2024 | I am seeking job opportunity starting in 2025 Fall. If you think I’m a good fit, please contact me via lyibing112@gmail.com. |
Jun 1, 2024 | One paper gets accepted in TIP 2024. This paper dicusses the feature alignment problem of the contrastive learning, and presents a high-level concept contrast approach. |
Jan 16, 2023 | One paper gets accepted in ICLR 2024 with Spotlight presentation (Top 5%). We present neuron activation coverage (NAC) that works for both OOD detection and generalization problems. |
(*) denotes equal contribution
-
Neuron Activation Coverage: Rethinking Out-of-distribution Detection and Generalization
Yibing Liu, Chris Xing Tian, Haoliang Li, Lei Ma, and Shiqi Wang
In the 12th International Conference on Learning Representations, 2024
Spotlight Presentation [Top %5] The out-of-distribution (OOD) problem generally arises when neural networks encounter data that significantly deviates from the training data distribution, i.e., in-distribution (InD). In this paper, we study the OOD problem from a neuron activation view. We first formulate neuron activation states by considering both the neuron output and its influence on model decisions. Then, to characterize the relationship between neurons and OOD issues, we introduce the neuron activation coverage (NAC) – a simple measure for neuron behaviors under InD data. Leveraging our NAC, we show that 1) InD and OOD inputs can be largely separated based on the neuron behavior, which significantly eases the OOD detection problem and beats the 21 previous methods over three benchmarks (CIFAR-10, CIFAR-100, and ImageNet-1K). 2) a positive correlation between NAC and model generalization ability consistently holds across architectures and datasets, which enables a NAC-based criterion for evaluating model robustness. Compared to prevalent InD validation criteria, we show that NAC not only can select more robust models, but also has a stronger correlation with OOD test performance.
-
Generalization Beyond Feature Alignment: Concept Activation-Guided Contrastive Learning
Yibing Liu, Chris Xing Tian, Haoliang Li, and Shiqi Wang
IEEE Transactions on Image Processing, 2024
Learning invariant representations via contrastive learning has seen state-of-the-art performance in domain generalization (DG). Despite such success, in this paper, we find that its core learning strategy – feature alignment – could heavily hinder model generalization. Drawing insights in neuron interpretability, we characterize this problem from a neuron activation view. Specifically, by treating feature elements as neuron activation states, we show that conventional alignment methods tend to deteriorate the diversity of learned invariant features, as they indiscriminately minimize all neuron activation differences. This instead ignores rich relations among neurons – many of them often identify the same visual concepts despite differing activation patterns. With this finding, we present a simple yet effective approach, Concept Contrast (CoCo), which relaxes element-wise feature alignments by contrasting high-level concepts encoded in neurons. Our CoCo performs in a plug-and-play fashion, thus it can be integrated into any contrastive method in DG. We evaluate CoCo over four canonical contrastive methods, showing that CoCo promotes the diversity of feature representations and consistently improves model generalization capability. By decoupling this success through neuron coverage analysis, we further find that CoCo potentially invokes more meaningful neurons during training, thereby improving model learning.
-
Rethinking Attention-Model Explainability through Faithfulness Violation Test
Yibing Liu, Haoliang Li, Yangyang Guo, Chenqi Kong, Jing Li, and Shiqi Wang
In the 39th International Conference on Machine Learning, 2022
Attention mechanisms are dominating the explainability of deep models. They produce probability distributions over the input, which are widely deemed as feature-importance indicators. However, in this paper, we find one critical limitation in attention explanations: weakness in identifying the polarity of feature impact. This would be somehow misleading – features with higher attention weights may not faithfully contribute to model predictions; instead, they can impose suppression effects. With this finding, we reflect on the explainability of current attention-based techniques, such as Attention \bigodot Gradient and LRP-based attention explanations. We first propose an actionable diagnostic methodology (henceforth faithfulness violation test) to measure the consistency between explanation weights and the impact polarity. Through the extensive experiments, we then show that most tested explanation methods are unexpectedly hindered by the faithfulness violation issue, especially the raw attention. Empirical analyses on the factors affecting violation issues further provide useful observations for adopting explanation methods in attention models.
-
Answer Questions with Right Image Regions: A Visual Attention Regularization Approach
Yibing Liu, Yangyang Guo, Jianhua Yin, Xuemeng Song, Weifeng Liu, Liqiang Nie, and Min Zhang
ACM Trans. Multimedia Comput. Commun. Appl., 2022
Visual attention in Visual Question Answering (VQA) targets at locating the right image regions regarding the answer prediction, offering a powerful technique to promote multi-modal understanding. However, recent studies have pointed out that the highlighted image regions from the visual attention are often irrelevant to the given question and answer, leading to model confusion for correct visual reasoning. To tackle this problem, existing methods mostly resort to aligning the visual attention weights with human attentions. Nevertheless, gathering such human data is laborious and expensive, making it burdensome to adapt well-developed models across datasets. To address this issue, in this article, we devise a novel visual attention regularization approach, namely, AttReg, for better visual grounding in VQA. Specifically, AttReg first identifies the image regions that are essential for question answering yet unexpectedly ignored (i.e., assigned with low attention weights) by the backbone model. And then a mask-guided learning scheme is leveraged to regularize the visual attention to focus more on these ignored key regions. The proposed method is very flexible and model-agnostic, which can be integrated into most visual attention-based VQA models and require no human attention supervision. Extensive experiments over three benchmark datasets, i.e., VQA-CP v2, VQA-CP v1, and VQA v2, have been conducted to evaluate the effectiveness of AttReg. As a by-product, when incorporating AttReg into the strong baseline LMH, our approach can achieve a new state-of-the-art accuracy of 60.00% with an absolute performance gain of 7.01% on the VQA-CP v2 benchmark dataset. In addition to the effectiveness validation, we recognize that the faithfulness of the visual attention in VQA has not been well explored in literature. In the light of this, we propose to empirically validate such property of visual attention and compare it with the prevalent gradient-based approaches.
teaching assistant
CS4187 Computer Vision for Interactivity, 2023-2024 & 2024-2025 Semester A
CS5187 Vision and Image, 2022-2023 Semester B
CS4187 Computer Vision for Interactivity, 2022-2023 Semester A
CS4296/CS5296 Cloud Computing, 2021-2022 Semester B
CS1102 Introduction to Computer Studies, 2021-2022 Semester A
professional services
Conference Reviewer: ICLR 2025, ICML 2024, ICLR 2024, NeruIPS 2023, ICML 2022
Journal Reviewer: IEEE TKDE, TCYB, TCSVT, ACM ToMM
Invited PC member for short papers track at WWW 2024
honors
Institutional Research Tuition Scholarship at CityU, 2022 & 2024
Outstanding Graduate of Shandong University, 2020
The First Prize Scholarship at Shandong University (Top 5%), 2017-2019
National Scholarship for Encouragement, 2018