Skip to content

slies/WWW-26-Tutorial

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

75 Commits
 
 
 
 
 
 

Repository files navigation

  • 🌍 Conference: The Web Conference (WWW 2026)
  • 📅 Date: June 29-July 3, 2026
  • 📍 Dubai, United Arab Emirates

Generalist Model for Structured Data: Foundations, Frontiers and Applications

  • Authors: Peng Cui, Xingxuan Zhang, Hanjia Ye, Jintai Chen, Shuyang Li

  • Year: 2026

  • Abstract: Structured data constitutes one of the most ubiquitous data modalities in web-scale and enterprise applications, supporting tasks such as recommendation, forecasting, and user behavior analysis. Conventional modeling paradigms—ranging from generalized linear models and gradient boosting to deep structured networks—have provided strong baselines for predictive analytics and decision support. However, the recent emergence of foundation models and in-context learning (ICL) has sparked a new paradigm for structured modeling, moving from dataset-specific training toward universal, adaptable inference. Emerging structured foundation models illustrate how large-scale pretraining, synthetic data generation, and ICL-based inference can extend foundation-model principles to structured data. These developments open new possibilities for multi-task learning, zero-shot inference, and knowledge transfer across diverse structured settings. Yet, the space of structured foundation models remains largely unexplored, with open questions surrounding data generation, multi-task setting, pretraining objectives, and evaluation standards. This tutorial will provide a structured overview of both conventional modeling and recent ICL-based approaches. Participants will gain a comprehensive understanding of established methods, current advances in foundation models, and open research challenges. In particular, we will offer an in-depth introduction to structured ICL and review the most representative foundation models in this field. Several key topics in this field will be discussed, including pretraining data generation, multi-task learning, and other emerging directions in the modeling of structured data. This tutorial aims to bridge conventional machine learning and the emerging foundation-model paradigm, providing attendees with conceptual and practical insights into structured data modeling in the era of generalist foundation models.

🧑‍🏫 Tutors

Peng Cui

Bio: Peng Cui is an Associate Professor with tenure in the Depart- ment of Computer Science at Tsinghua University. He got his PhD degree from Tsinghua University in 2010. His research in- terests include causally-regularized machine learning, network representation learning, and social dynamics modeling. He has published more than 100 papers in prestigious conferences and journals in data mining and multimedia. His recent research won the IEEE Multimedia Best Department Paper Award, SIGKDD 2016 Best Paper Finalist, ICDM 2015 Best Student Paper Award, SIGKDD 2014 Best Paper Finalist, IEEE ICME 2014 Best Paper Award, ACM MM12 Grand Challenge Multimodal Award, and MMM13 Best Paper Award. He is PC co-chair of CIKM2019 and MMM2020, SPC or area chair of ICML, KDD, WWW, IJCAI, AAAI, etc., and Associate Editors of IEEE TKDE, IEEE TBD, ACM TIST, and ACM TOMM etc. He received ACM China Rising Star Award in 2015, and CCF-IEEE CS Young Scientist Award in 2018. He is now a Distinguished Member of ACM and CCF, and a Senior Member of IEEE.

Xingxuan Zhang

Bio: Xingxuan Zhang is an Assistant Researcher in the Department of Computer Science at Tsinghua University, where he also re- ceived his PhD. His research interests lie in Foundation Models and Trustworthy AI, with a specialized focus on developing fun- damental modeling methods for structured data. He has published more than 30 papers in top-tier conferences and journals includ- ing ICLR, Neurips, ICML, WWW, CVPR, and ICCV, where he also served as a reviewer and program committee member.

Han-Jia Ye

Bio: Han-Jia Ye is an Associate Professor in the School of Arti- ficial Intelligence at Nanjing University. His research centers on machine learning—especially representation learning, meta- learning, model reuse, and deep learning for tabular data. He received his Ph.D. in Computer Science from Nanjing University in 2019. He has served as Tutorial Co-Chair for SDM 2023 and Doctoral Forum Co-Chair for SDM 2022, and as an area chair for top-tier venues including ICML, NeurIPS, ICLR, AAAI, IJCAI, and CVPR. He leads the development of TALENT, a toolbox of representative deep tabular methods, and is organizing an AAAI 2026 tutorial on representation learning for structured tabular data.

Jintai Chen

Bio: Jintai Chen is an Assistant Professor in the AI Thrust, Informa- tion Hub at The Hong Kong University of Science and Technology (Guangzhou). He received his PhD from Zhejiang University and later conducted postdoctoral research at the University of Illinois at Urbana–Champaign. His research centers on artificial intel- ligence for tabular data and healthcare, with a particular focus on developing foundation models for heterogeneous and noisy clinical and biomedical datasets. His broader interests include medical agents, clinical trial optimisation, and drug design.

Shuyang Li

Bio: Shuyang Li is a Postdoctoral Researcher in the Department of Computer Science at Tsinghua University. He received his PhD degree in Electrical, Electronic, and Communications Engineering from Politecnico di Torino, Italy. After obtaining his PhD, he has been working as a researcher at Politecnico di Torino, collaborating on projects with Telecom Italia, and conducting research in mobile traffic forecasting and predictive maintenance. His research has been published in journals and conferences such as Computer Networks, ICC, WCNC, VTC, etc. His main research interests include time series analysis, synthetic data generation, cellular network optimization, and mobile traffic modeling.

📌 Paper List

⭐ Acknowledgements

We thank the organizers of WWW 2026, all contributors to this tutorial, and Jiawei Chen (Beihang University, https://scholar.google.com/citations?user=2803pOEAAAAJ) for assisting us in preparing the materials.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors