Keynote Speakers

Meet the PAKDD 2026 keynote speakers and explore their keynote talk titles and abstracts.
Prof. Zhi-Hua ZHOU
Prof. Zhi-Hua ZHOU

Professor and Vice President, Nanjing University

Bio

Zhi-Hua Zhou is Professor of Computer Science and Artificial Intelligence, Vice President of Nanjing University. His research interests are mainly in machine learning and data mining, with significant contributions to ensemble learning, multi-label and weakly supervised learning, etc. He has authored the books "Ensemble Methods: Foundations and Algorithms", "Machine Learning", etc., and published more than 200 papers in top-tier journals or conferences, with more than 100,000 citations according to Google Scholar. Many of his inventions have been successfully deployed in industry. He is President of IJCAI Trustee, Series Editor of Springer Lecture Notes in Artificial Intelligence, Editor-in-Chief of Frontiers of Computer Science, and advisory board member of AI Magazine. ...

Learnware: Small models do big

Abstract

"Learnware = Model + Specification". Let's consider the following questions: First, do we believe that in the future (A) there will be a big model that is able to cope with all possible learning tasks, or (B) it is crucial to have many models to collaborate? Second, are these models to be developed by (A) one developer (or one company), or (B) lots of developers all over the world? Third, are training data used to train these models to be (A) freely shared, or (B) mostly not? If we choose (B) for the answers, it seems that we will encounter a mission impossible: how to identify helpful models from a growing huge pool of trained models developed by developers all over the world, and reuse or even reassemble the models to tackle new user's task, given that we could not touch developers' and users' training data? "Learnware" makes this possible. A key ingredient is the specification which enables a trained model to be adequately identified to reuse according to the requirement of new user who knows nothing about the model, while model developers' training data are preserved. Learnwares are accommodated in a learnware dock system, which enables small models do big, and enables models do things even beyond their original development purposes. This talk will briefly introduce some preliminary research advances in this direction.

Prof. Geoff Webb
Prof. Geoff Webb

Professor, Monash University

Bio

Professor Geoff Webb is an eminent and highly-cited AI researcher. He is an Australian Research Council Laureate Fellow and Professor in the Monash University Department of Data Science and Artificial Intelligence. He was editor in chief of the Data Mining and Knowledge Discovery journal, from 2005 to 2014. He has been Program Committee Chair of both ACM SIGKDD and IEEE ICDM, as well as General Chair of ICDM and member of the ACM SIGKDD Executive. He is a Technical Advisor to machine learning as a service startup BigML Inc and to recommender systems startup FROOMLE. He developed many of the key mechanisms of support-confidence association discovery in the 1980s. His OPUS search algorithm remains the state-of-the-art in rule search. ...

Large Language Models: Capabilities and Limitations

Abstract

Large Language Models (LLMs) have made extraordinary advances in recent times. This has led many people to believe that LLMs provide a pathway to attaining Artificial General Intelligence. This talk will give a high level overview of Large Language Models and present my thoughts on their capabilities and limitations. I argue that LLMs are not intelligent and the LLMs alone do not provide a likely pathway to developing truly intelligent machines.

Dr. Xin Luna Dong
Dr. Xin Luna Dong

Principal Scientist, Meta Reality Labs

Bio

Xin Luna Dong is a Principal Scientist at Meta Wearables AI, where she leads the Agentic AI efforts for building trustworthy and personalized assistants on wearable devices. Previously, she spent over a decade advancing knowledge graph technology, including the Amazon Product Graph and the Google Knowledge Graph. She is co-author of Machine Knowledge: Creation and Curation of Comprehensive Knowledge Bases and Big Data Integration. ...

From Sight to Insight: Visual Memory for Smarter Assistants

Abstract

Imagine a personal assistant that, with user permission, persistently remembers moments from daily life-answering questions like "When and where did I see this lady?" or offering personalized suggestions like "You might enjoy The Little Prince-it relates to the statue you liked in Lyon." Realizing this vision requires overcoming major challenges: capturing visual memories under hardware constraints (e.g., memory, battery, thermal limits, bandwidth), extracting meaningful personalization signals from noisy, task-agnostic visual histories, and supporting real-time question answering and recommendations under tight latency requirements. In this talk, we present our early work toward this goal. Pensieve, our memory-based QA system, improves accuracy by 11% over state-of-the-art multimodal RAG baselines. VisualLens infers user interests from casual photos, outperforming leading recommendation systems by 5-10%. We also share initial results on efficient, event-triggered memory capture and compression. Our work points to a broad landscape of research opportunities in building richer, more context-aware personal assistants capable of learning from and reasoning over users' visual experiences.

Dr. Xia "Ben" Hu
Dr. Xia "Ben" Hu

Lead Scientist, Shanghai AI Laboratory

Bio

Dr. Xia "Ben" Hu is currently the Lead Scientist at Shanghai AI Laboratory. Previously, he was Full Professor, and Director of the Data Science Institute, at Rice University, USA, and co-founder of AI POW LLC. Dedicated to machine learning and AI research, he has published over 200 papers in top conferences/journals like ICLR, NeurIPS and KDD, with over 40,000 citations. His leading works include AutoKeras (a top AutoML framework), NCF algorithm (officially recommended by TensorFlow, single paper cited over 9,000 times), and anomaly detection systems used by NVIDIA, Apple, etc. His papers have received several Best Paper (Candidate) awards from venues such as ICML, WWW, WSDM, ICDM, AMIA and INFORMS. He has won honors such as NSF CAREER Award and KDD Rising Star Award. ...

TBC

Abstract

TBC

Dr. Qingsong Wen
Dr. Qingsong Wen

Head of AI & Chief Scientist, Squirrel Ai Learning

Bio

Qingsong Wen is currently the Head of AI & Chief Scientist at Squirrel Ai Learning (a top EdTech unicorn), and PhD Supervisor at University of Oxford. Before that, he worked at Alibaba, Qualcomm, Marvell, etc., and received his M.S. and Ph.D. degrees from Georgia Institute of Technology, USA. His research interests include machine learning, data mining, and signal processing, especially AI for Time Series, AI for Education, LLM & AI Agent. He has published nearly 200 top-ranked AI conference and journal papers. Currently, he serves as Associate Vice President for Cybernetics of IEEE SMC Society, Chair of IEEE CIS Task Force on AI for Time Series and Spatio-Temporal Data, and Vice Chair of INNS AI for Education Section. ...

Agentic AI Tutors for Personalized Education at Scale

Abstract

Recent advances in large language and multimodal models have enabled agentic AI tutors to operate autonomously in real-world education systems. This keynote presents how adaptive and personalized educational AI agents are designed and deployed at scale using state-of-the-art techniques, including multi-agent LLM reasoning, multimodal error analysis, dialog-driven interaction, and longitudinal learner modeling. Drawing on large-scale applications from Squirrel Ai Learning, the talk demonstrates how these AI tutors perform fine-grained misconception diagnosis, personalized feedback generation, and adaptive learning path optimization through continuous interaction with real student data. Validated through large-scale real-world deployments, this work highlights key technical lessons on reliability, robustness, and trust in smart education ecosystems.