๐Ÿ‘จโ€๐ŸŽ“ Education

I am a Ph.D. candidate (2022โ€“) at the College of Informatics, Huazhong Agricultural University, supervised by Prof. Hong Chen. Previously, I was pursuing the M.S. degree (2020-2022) with the assistance of Prof. Lingjuan Wu. I received the B.S. degree in Engineering from China Agricultural University in 2020.

๐Ÿ”ฌ Research Area

My research interests lie in the areas of optimization and learning theory, with emphasis on the following topics:

  • Automatic machine learning/LLM (e.g., Bilevel Optimization, Multi-Agent Communication)

  • Robust/Interpretable machine learning (e.g., Robust Metric, Sparse/Neural Additive Models)

  • Applications (e.g., Hardware Trojan Detection, Intelligent Biology Analysis)

Some suggested videos for better understanding the bilevel optimization, robust machine learning, interpretable additive models as well as the hardware Trojans.

If you are interested or have any questions about my work, please feel free to contact me: xlinml@163.com

๐Ÿ”ฅ News

  • 2026.05: ย ๐ŸŽ‰๐ŸŽ‰ New Acceptance: Three papers on Interpretable ML/LLM Evaluation/GUI Agents to appear in ICML (ccf-A).

  • 2026.04: ย ๐ŸŽ‰๐ŸŽ‰ New Acceptance: One paper on GUI Agents to appear in ACL Findings.

  • 2026.04: ย ๐ŸŽ‰๐ŸŽ‰ New Acceptance: One paper on hardware security to appear in Computer & Security (ccf-B).

  • 2026.01: ย ๐ŸŽ‰๐ŸŽ‰ New Acceptance: Two papers on interpretability to appear in Information Sciences (ccf-B) and ICASSP (ccf-B).

  • 2026.01: ย ๐ŸŽ‰๐ŸŽ‰ New Awards: Two papers are selected as โ€œBest Presentation in Sessionโ€ in ICPADS.

  • 2024.11: ย ๐ŸŽ‰๐ŸŽ‰ New Award: China Doctoral National Scholarship.

๐Ÿ“ Publications in Machine Learning

[25] ICML 2026 [ccf-A, Core A*]
sym

S2MAM: Semi-supervised Meta Additive Model for Robust Estimation and Variable Selection

Xuelin Zhang, Hong Chen*, Yingjie Wang, Tieliang Gong, Bin Gu

Forty-Third International Conference on Machine Learning 2026 [C]

  • This paper introduces the Semi-Supervised Meta Additive Model (S2MAM) to overcome the vulnerability of traditional manifold regularization to noisy data by employing a bilevel optimization scheme that automatically selects informative variables and dynamically updates the similarity matrix.
  • Supported by theoretical guarantees for convergence and generalization, extensive experiments confirm that the model achieves robust and interpretable predictions across diverse corrupted datasets.
  • Implementation: Github Link
[24] ICML 2026 [ccf-A, Core A*]
sym

UltraHorizon: Benchmarking LLM-Agent Capabilities in Ultra Long-Horizon Scenarios

Haotian Luo#, Huaisong Zhang#, Xuelin Zhang#, Haoyu Wang#, Zeyu Qin#, WenJie Lu#, Guozheng Ma, Haiying He, Yingsha Xie, Qiyang Zhou, Zixuan Hu, Hongze Mi, Yibo Wang, Naiqiang Tan, Hong Chen, Yi R. Fung, Chun Yuan, Li Shen*

Forty-Third International Conference on Machine Learning 2026 [C]

  • Existing evaluations for autonomous agents typically fail to capture the complexity of long-horizon, partially observable real-world tasks that require sustained reasoning, memory management, and tool use. To address this gap, we introduce a novel exploration-based benchmark across three environments, featuring extremely long agent trajectories with high token counts and frequent tool calls.
  • Extensive experiments reveal that current state-of-the-art agents significantly underperform compared to humans and fail to improve with simple scaling, primarily due to in-context locking and fundamental capability gaps.
  • Implementation: Github Link
[23] ICML 2026 [ccf-A, Core A*]
sym

Darwinian Memory: A Training-Free Self-Regulating Memory System for GUI Agent Evolution

Hongze Mi, Yibo Feng, WenJie Lu, Song Cao, Jinyuan Li, Yanming Li, Xuelin Zhang, Haotian Luo, Songyang Peng, He Cui, Tengfei Tian, Jun Fang, Hua Chai, Naiqiang Tan*.

Forty-Third International Conference on Machine Learning 2026 [C]

  • To overcome the limitations of current memory systems in complex GUI automation, the proposed Darwinian Memory System (DMS) uses a self-evolving, โ€œsurvival of the fittestโ€ approach to dynamically refine strategies by retaining successful actions and pruning suboptimal ones.
  • Extensive testing demonstrates that DMS significantly improves the success rate, execution stability, and speed of general-purpose Multimodal Large Language Models without requiring additional training overhead.
  • Implementation: Github Link
[22] ACL Findings 2026
sym

D-Artemis: A Deliberative Cognitive Framework for Mobile GUI Multi-Agents

Hongze Mi, Yibo Feng, WenJie Lu, Yuqi Wang, Jinyuan Li, Song Cao, He Cui, Tengfei Tian, Xuelin Zhang, Haotian Luo, Di Sun, Jun Fang, Hua Chai, Naiqiang Tan, Gang Pan.

Annual Meeting of the Association for Computational Linguistics 2026 [C]

  • Proposing the D-Artemis deliberative framework, which enhances general-purpose multimodal large language models for GUI tasks without requiring training on complex trajectory datasets, by leveraging a fine-grained, app-specific tip retrieval mechanism.
  • Designing a cognitive closed-loop mechanism of โ€œpre-execution alignment and post-execution reflection,โ€ where the collaborative operation of the TAC Check module, Action Correction Agent, and Status Reflection Agent proactively prevents execution failures and enables experiential learning.
  • Implementation: Github Link
[21] ICASSP 2026 [ccf-B, Core B]
sym

Distribution-Aware Neural Additive Models: Robust Interpretable Deep Learning with Feature Selection

Jingyi Chen, Xuelin Zhang, Peipei Yuan*, Liyuan Liu, Hong Chen.

International Conference on Acoustics, Speech, and Signal Processing 2026 [C]

  • Proposing a novel Distribution-Aware Neural Additive Model integrating kernel density estimation to flexibly capture diverse noise structures without restrictive distributional assumptions.
  • Incorporating group sparsity regularization to automatically select the most informative input features and enhance model interpretability as well as predictive performance.
  • The implementation could be found at Github Link
[20] Information Sciences 2026 [ccf-B, SCI-2]
sym

Maximum Likelihood Neural Additive Models

Jingyi Chen#, Xuelin Zhang#, Peipei Yuan, Rushi Lan, Hong Chen*.

Information Sciences 2026 [J]

  • Algorithmically, we propose the Maximum Likelihood Neural Additive Model (ML-NAM), which employs kernel density estimation for non-parametric residual modeling to achieve robust learning against non-Gaussian noise without distributional assumptions.
  • Theoretically, we establish non-asymptotic upper bounds on the excess risk, demonstrating that the model achieves minimax convergence rates with polynomial decay within Besov spaces.
  • The implementation could be found at Github Link
[19] ICPADS 2025 [ccf-C, Core B]
sym

Pairwise Generalized Importance Weighting for Metric Learning under Distribution Shift

Richeng Zhou, Xuelin Zhang*, Hong Chen, Weifu Li, Liyuan Liu.

IEEE International Conference on Parallel and Distributed Systems 2025 [C] (Best Presentation in Session)

  • Introduces an innovative solution to metric learning, effectively overcoming performance degradation under distributional shifts.
  • Identifies existing algorithmsโ€™ limitations and proves risk consistency across broad distributional changes, offering strong theoretical guarantees.
[18] ICPADS 2025 [ccf-C, Core B]
sym

Interpretable Bilevel Additive Taylor Model

Wenxing Zhou, Chao Xu, Lian Peng, Xuelin Zhang*.

IEEE International Conference on Parallel and Distributed Systems 2025 [C] (Best Presentation in Session)

  • Tucker decomposition is introduced for low-rank approximation by leveraging Taylor-series expansion to capture higher-order interactions.
  • Fusing bilevel optimization with sparse neural additive models and integrated Tucker decomposition, it robustly handles label noise and class imbalance while preserving interpretability.
[17] Journal of HZAU 2025 [Q2, T2]
sym

Citrus Pollen Viability Detection via Modified YOLOv11-FS Model

Liyuan Liu, Xuelin Zhang, Hong Chen, Weifu Li, Jianhua Liao, Kaidong Xie, Xiaomeng Wu, Yaohui Chen*.

Journal of Huazhong Agricultural University 2025 [J]

  • An improved YOLOv11-FS model is proposed that effectively overcomes detection challenges such as the small size, clumping tendency, and complex background of citrus pollen grains.
  • It provides reliable technical support for the breeding of seedless citrus varieties and can also underpin automated pollen-viability monitoring and cultivar improvement in the intelligent management of citrus orchards.
[16] CIKM 2025 [ccf-B, Core A]
sym

Interpretable Meta-weighted Sparse Neural Additive Networks

Xuelin Zhang, Hong Chen, Lingjuan Wu*.

ACM Conference on Information and Knowledge Management 2025 [C]

  • Under a bi-level optimization scheme, we jointly learn adaptive sample weights and impose sparsity on a neural additive modelโ€™s univariate functions, yielding both continual robustness and single-feature interpretability; experiments across multiple distribution shifts confirm markedly superior performance and strong resistance to forgetting.
[15] JNSPM 2025
sym

On the Convergence of Nonconcave-Nonconvex Max-Min Optimization Problem

Xuelin Zhang*

Journal of Numerical Simulations in Physics and Mathematics 2025 [J]

  • This paper introduces a novel two-sided Polyak-ลojasiewicz and Quadratic Growth framework that establishes the convergence guarantee of O(1/K) for SAGDA in nonconvex-nonconcave max-min optimization, matching convex-concave rates under milder geometry.
[14] ICML 2025 [ccf-A, Core A*]
sym

On the Generalization Ability of Next-Token-Prediction Pretraining

Zhihao Li, Xue Jiang, Liyuan Liu, Xuelin Zhang, Hong Chen and Feng Zheng.

International Conference on Machine Learning 2025 [C]

  • This study establishes a theoretical framework for Next-Token-Prediction (NTP) pre-training based on Rademacher complexity, introduces a novel decomposition method, and provides the first generalization bounds for NTP. The findings offer valuable insights into how model parameters influence generalization and have been empirically validated, advancing both the theoretical comprehension and practical application of large language models.
[13] ICDM 2024 [ccf-B, Core A*]
sym

Generalized Sparse Additive Model with Unknown Link Function

Peipei Yuan, Xinge You*, Hong Chen, Xuelin Zhang, and Qinmu Peng.

IEEE International Conference on Data Mining 2024 [C]

  • In this work, we propose a novel generalized additive model with a flexible link function automatically learned by a bilevel scheme. The proposed model is capable of nonlinear approximation, hidden interaction and feature selection, which also enjoys the theoretical guarantee of algorithmic convergence.
[12] ESWA 2024 [sci-1 Top]
sym

Error Density-dependent Empirical Risk Minimization

Hong Chen*, Xuelin Zhang, Tieliang Gong, Bin Gu, Feng Zheng.

Expert Systems With Applications 2024 [J]

  • This paper goes beyond the limitation of error value-dependent learning criterion and proposes the EDERM framework for robust regression against atypical data. Sufficient empirical evaluations validate the effectiveness of our method.
  • The implementation codes can be found at: Github Link
[11] IJCAI 2024 [ccf-A, Core A*]
sym

Fine-grained Analysis of Stability and Generalization for Stochastic Bilevel Optimization

Xuelin Zhang, Hong Chen*, Bin Gu, Tieliang Gong, Feng Zheng.

International Joint Conference on Artificial Intelligence 2024 [C] (Oral)

  • In this paper, we provide a systematic generalization analysis of the first-order gradient-based bilevel optimization methods, based on the (on-average argument) algorithmic stability technique.
  • The verification codes are provided at: Github Link
[10] IJCNN 2024 [ccf-C, Core A]
sym

Improved Concentration Bound for CVaR

Peng Sima, Hao Deng*, Xuelin Zhang, Hong Chen.

International Joint Conference on Neural Networks 2024 [C]

  • This paper introduces a novel estimator that relies on an estimator of Value at Risk (VaR) and investigates the concentration inequalities in independent scenarios where the underlying distributions are sub-Gaussian, sub-exponential, or heavy-tailed, where the inequalities we derive are bilateral, exhibit exponential decay, and are not confined to bounded scenarios.
[9] FCS 2023 [ccf-B]
sym

Neural Partially Linear Additive Model

Liangxuan Zhu, Han Li*, Xuelin Zhang, Lingjuan Wu, Hong Chen.

Frontiers of Computer Science 2023 [J]

  • This paper proposes a Neural Partially Linear Additive Model (NPLAM), which automatically distinguishes insignificant, linear, and nonlinear features in neural networks, which can realize model-level interpretability.
[8] AAAI 2023 [ccf-A, Core A*]
sym

Stepdown SLOPE for Controlled Feature Selection

Jingxuan Liang, Xuelin Zhang, Hong Chen*, Weifu Li, Xin Tang.

Association for the Advancement of Artificial Intelligence 2023 [C] (Oral)

  • This paper goes beyond the previous concern of Sorted L-One Penalized Estimation (SLOPE) limited to the false discovery rate (FDR) control by considering the stepdown-based SLOPE in order to control the probability of k or more false rejections (k-FWER) and the false discovery proportion (FDP).
  • The codes are shared at: Github Link
[7] APIN 2023 [sci-2]
sym

Robust variable structure discovery based on tilted empirical risk minimization

Xuelin Zhang, Yingjie Wang, Liangxuan Zhu, Hong Chen, Han Li, Lingjuan Wu*.

Applied Intelligence 2023 [J]

  • In this paper, we propose a new robust variable structure discovery method for group lasso based on a convergent bilevel optimization framework, where the robust tilted empirical risk minimization is adopted.
  • The implementation codes can be found at: Github Link
[6] ICCCS 2022
sym

Robustness of classifier to adversarial examples under imbalanced data

Wenqian Zhao, Han Li, Lingjuan Wu*, Liangxuan Zhu, Xuelin Zhang, Yizhi Zhao.

International Conference on Computer and Communication Systems 2022 [C]

  • In this paper, we provide a theoretical framework to analyze the robustness of a classifier to AE under an imbalanced dataset from the perspective of AUC (Area under the ROC curve), and derive an interpretable upper bound.

๐Ÿ“ Publications in Hardware Security

[5] COSE 2025 [ccf-B, SCI Q2 Top]
sym

Explainable Hardware Trojan Detection and Localization in FPGA Netlists

Lingjuan Wu, Hao Su, Siyi Tian, Xuelin Zhang, Wei Hu*.

Computers & Security 2026 [J]

  • proposing an explainable hardware Trojan detection and localization method based on graph neural networks, leveraging the initialization vector features of look-up tables (LUTs) in FPGA netlists
  • introducing Granger causality theory, the method further reveals the decision-making mechanism of the GNN model to achieve precise identification of infected IP cores and Trojan nodes in Xilinx FPGA netlists
[4] TCAD 2025 [ccf-A]
sym

Towards Precise and Explainable Hardware Trojan Localization at LUT Level

Hao Su, Wei Hu, Xuelin Zhang, Dan Zhu, Lingjuan Wu*.

IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems 2025 [J]

  • The proposed approach aims to extract the rich structural and behavioral features at the look-up-table (LUT) level to train an explainable graph neural network (GNN) model for classifying design nodes in FPGA netlists and identifying the Trojan-infected ones.
  • The implementation codes can be found at: Github Link
[3] ITC-Asia 2024 [ccf-C]
sym

Pinpointing Hardware Trojans Through Semantic Feature Extraction and Natural Language Processing

Yichen Li, Wei Hu, Hao Su, Xuelin Zhang, Yizhi Zhao, Lingjuan Wu*.

International Test Conference in Asia 2024 [C]

  • In this work, we propose a novel hardware Trojan detection method at RTL. Our approach involves the transformation of hardware design into CDFG, followed by path extraction and segmentation.
[2] ICCAD 2023 [ccf-B, Core A]
sym

Automated Hardware Trojan Detection at LUT Using Explainable Graph Neural Networks

Lingjuan Wu, Hao Su, Xuelin Zhang, Yu Tai, Han Li, Wei Hu*.

International Conference on Computer-Aided Design 2023 [C]

  • In this work, we propose a novel hardware Trojan detection method based on graph neural networks (GNNs) targeting FPGA netlists. We leverage the rich explicit structural features and behavioral characteristics at LUT, which offer an ideal abstraction level and granularity for Trojan detection. A GNN model with optimized class-balanced focal loss is trained for automated Trojan feature extraction and classification.
  • Model implementation is available at Github Link
[1] HOST 2022
sym

Hardware Trojan Detection at LUT: Where Structural Features Meet Behavioral Characteristics

Lingjuan Wu, Xuelin Zhang, Siyi Wang, Wei Hu.

International Symposium on Hardware Oriented Security and Trust 2022 [C]

  • This work proposes a novel hardware Trojan detection method that leverages static structural features and behavioral characteristics in field programmable gate array (FPGA) netlists. Mapping of hardware design sources to look-up-table (LUT) networks makes these features explicit, allowing automated feature extraction and further effective Trojan detection through machine learning.
  • The implemented codes are available at: Github Link

๐ŸŽ–๏ธ Activities and Honors

๐Ÿ› ๏ธ Authorized Patents

  • 2024.06: Hong Chen, Xuelin Zhang, Weifu Li, Feng Zheng. CN114580299A.

๐Ÿ’ฌ Academic Services

๐ŸŒ Visit Counter

Flag Counter