Publications
[Google scholar] [DBLP] [View by topic]
Summary: Nature (1), Nature portfolio journal (1), ICLR (14), NeurIPS (13), ICML (12), KDD (3), IEEE TPAMI (2), IJCAI (3), CVPR (4), IEEE TKDE (4), ACL (8), AAAI (1), ICCV (1), ECCV (1), UbiComp (3), MM (1).
Seleted Publications
- [Nature] General scales unlock ai evaluation with explanatory and predictive power.
- [NeurIPS’21] Flexmatch: Boosting semi-supervised learning with curriculum pseudo labeling. (1400+ Citations; Top 17 most cited NeurIPS papers in the past 5 years)
- [ICML’24 Oral (Top 3%)] CompeteAI: Understanding Competition Dynamics of LLM-based Agents.
- [NeurIPS’24 Spotlight (Top 5%)] Slight Corruption in Pre-training Data Makes Better Diffusion Models.
- [IEEE TPAMI’25] Impact of Noisy Supervision in Foundation Model Learning.
- [ICLR’24 Spotlight (Top 5%)] DyVal: Dynamic Evaluation of Large Language Models for Reasoning Tasks.
- [ICLR’24 Spotlight (Top 5%)] Understanding and Mitigating the Label Noise in Pre-training on Downstream Tasks.
- [ICML’24] The good, the bad, and why: Unveiling emotions in generative ai. (Reported by Forbes and many international media)
- [IEEE TPAMI’24] DIVERSIFY: A General Framework for Time Series Out-of-distribution Detection and Generalization.
- [CIKM’21] Adarnn: Adaptive learning and forecasting of time series. (Paperdigest most influencial paper)
- [ACM MM’18 Oral] Visual domain adaptation with manifold embedded distribution alignment. (800 Citations; 2nd most cited paper in MM’18)
- [ICDM’17] Balanced distribution adaptation for transfer learning. (700 Citations; most cited paper in ICDM’17)
Recent Preprints
- AgentArk: Distilling Multi-Agent Intelligence into a Single LLM Agent. Yinyi Luo, Yiqiao Jin, Weichen Yu, Mengqi Zhang, Srijan Kumar, Xiaoxiao Li, Weijie Xu, Xin Chen, Jindong Wang. [arxiv] [code]
- Thinking Makes LLM Agents Introverted: How Mandatory Thinking Can Backfire in User-Engaged Agents. Jiatong Li, Changdae Oh, Hyeong Kyu Choi, Jindong Wang, Sharon Li. [arxiv] [code]
- Evolving Collective Cognition in Human-Agent Hybrid Societies: How Agents Form Stances and Boundaries. Hanzhong Zhang, Muhua Huang, Jindong Wang. [arxiv]
- Topological Structure Learning Should Be A Research Priority for LLM-Based Multi-Agent Systems. Jiaxi Yang, Mengqi Zhang, Yiqiao Jin, Hao Chen, Qingsong Wen, Lu Lin, Yi He, Weijie Xu, James Evans, Jindong Wang. [arxiv]
- Corruption-Aware Training of Latent Video Diffusion Models for Robust Text-to-Video Generation. Chika Maduabuchi, Hao Chen, Yujin Han, Jindong Wang. [arxiv] [code]
- Understanding and Mitigating the Bias Inheritance in LLM-based Data Augmentation on Downstream Tasks. Miaomiao Li, Hao Chen, Yang Wang, Tingyuan Zhu, Weijia Zhang, Kaijie Zhu, Kam-Fai Wong, Jindong Wang. [arxiv]
- MentalArena: Self-play Training of Language Models for Diagnosis and Treatment of Mental Health Disorders. Cheng Li, May Fung, Qingyun Wang, Chi Han, Manling Li, Jindong Wang, Heng Ji. [arxiv]
- On the Diversity of Synthetic Data and its Impact on Training Large Language Models. Hao Chen, Abdul Waheed, Xiang Li, Yidong Wang, Jindong Wang, Bhiksha Raj, Marah I. Abdin. [arxiv]
Books
- CHAPTER 3: Improving Large Foundation Models in Education for Multi-cultural UnderstandingFuture of Learning with Large Language Models: Applications and Research in Education 2025 | [ HTML ]
- Activity RecognitionSpringer International Publishing 2023 | [ HTML ]
All Papers
2026
- Classroom AI: Large Language Models as Grade-Specific TeachersNPJ Artificial Intelligence (Nature Portfolio Journal) 2026 | [ arXiv ]
- Self-Corrected Image Generation with Explainable Latent RewardsIEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2026 | [ ]
- Is your (reasoning) Multimodal Language Model vulnerable toward Distractions?The Fortieth AAAI Conference on Artificial Intelligence (AAAI) 2026 | [ arXiv ]
- FedUMM: A General Framework for Federated Learning with Unified Multimodal ModelsThe Web Conference (WWW) workshop on Federated Learning 2026 | [ arXiv ]
2025
- On fairness of unified multimodal large language model for image generationThirty-Ninth Annual Conference on Neural Information Processing Systems (NeurIPS) 2025 | [ arXiv ]
- From Pretraining to Pathology: How Noise Leads to Catastrophic Inheritance in Medical ModelsThirty-Ninth Annual Conference on Neural Information Processing Systems (NeurIPS) 2025 | [ Website ]
- Exposing and Patching the Flaws of Large Language Models in Social Character SimulationConference on Language Modeling (COLM) 2025 | [ arXiv ]
- MELON: Indirect Prompt Injection Defense via Masked Re-execution and Tool ComparisonInternational Conference on Machine Learning (ICML) 2025 | [ arXiv ]
- Topology-aware Neural Flux Prediction Guided by PhysicsInternational Conference on Machine Learning (ICML) 2025 | [ arXiv ]
- SoftVQ-VAE: Efficient 1-Dimensional Continuous TokenizerProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2025 | [ arXiv ]
- CultureVLM: Characterizing and Improving Cultural Understanding of Vision-Language Models for over 100 CountriesCVPR 2025 workshop on VLMs4ALL 2025 | [ arXiv ]
2024
- RESTful-Llama: Connecting User Queries to RESTful APIsProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP): Industry Track 2024 | [ PDF ]
- On catastrophic inheritance of large foundation modelsData-centric Machine Learning Research (DMLR) 2024 | [ arXiv ]
- SpecFormer: Guarding Vision Transformer Robustness via Maximum Singular Value PenalizationECCV 2024 | [ arXiv ]
- Selective mixup helps with distribution shifts, but not (only) because of mixupInternational Conference on Machine Learning (ICML) 2024 | [ arXiv ]
- Open-Vocabulary Calibration for Vision-Language ModelsInternational Conference on Machine Learning (ICML) 2024 | [ arXiv ]
- Position Paper: What Can Large Language Models Tell Us about Time Series AnalysisInternational Conference on Machine Learning (ICML) 2024 | [ arXiv ]
- MM-Soc: Benchmarking Multimodal Large Language Models in Social Media PlatformsThe 62nd Annual Meeting of the Association for Computational Linguistics (ACL) Findings 2024 | [ arXiv ]
- LSTPrompt: Large Language Models as Zero-Shot Time Series Forecasters by Long-Short-Term PromptingThe 62nd Annual Meeting of the Association for Computational Linguistics (ACL) Findings 2024 | [ arXiv ]
- Supervised Knowledge Makes Large Language Models Better In-context LearnersInternational Conference on Learning Representation (ICLR) 2024 | [ arXiv ]
- UP-Net: An Uncertainty-Driven Prototypical Network for Few-Shot Fault DiagnosisIEEE Transactions on Neural Networks and Learning Systems (TNNLS) 2024 | [ ]
- Towards Optimization and Model Selection for Domain Generalization: A Mixup-guided SolutionSIAM Conference on Data Mining (SDM) 2024 | [ arXiv ]
- Conv-adapter: Exploring parameter efficient transfer learning for convnetsCVPR 2024 Workshop Prompting in Vision 2024 | [ arXiv ]
2023
- Out-of-Distribution Generalization in Text Classification: Past, Present, and FutureThe 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2023 | [ arXiv ]
- Boosting cross-domain speech recognition with self-supervisionIEEE Transactions on Audio, Speech and Language Processing (TASLP) 2023 | [ arXiv ]
2022
- Hierarchical knowledge amalgamation with dual discriminative feature alignmentInformation Sciences 2022 | [ Website ]
- Local and global alignments for generalizable sensor-based human activity recognitionIEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2022 | [ HTML ]
2021
- MixSpeech: Data Augmentation for Low-resource Automatic Speech RecognitionIEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2021 | [ arXiv ]
2020
2019
- DrowsyDet: A Mobile Application for Real-time Driver Drowsiness DetectionUbiquitous Intelligence Computing 2019 | [ ]
2018
2017
2016