
Dongkuan Xu
overfitting
transformer-based models
sparse progressive distillation
knowledge-retention pruning
7
presentations
5
number of views
SHORT BIO
Dongkuan (DK) Xu is a Ph.D. student at Penn State, advised by Prof. Xiang Zhang. His research interest is resource-efficient deep learning for AI at scale, focusing on how to improve the efficiency of deep learning systems to achieve Pareto optimality between resources (e.g., parameters, data, computation) and performance (e.g., inference, training). DK has published more than 25 papers in top conferences and journals, including NeurIPS, AAAI, ACL, NAACL, and IJCAI, with more than 1400 citations. He has served as a (senior) PC member or regular reviewer for over 28 major conferences and 14 journals, and has worked as an instructor or teaching assistant for 8 courses. DK also has extensive research experience in industry. He has interned at Microsoft Research Redmond, Moffett AI, and NEC labs America, and holds 8 US patents/applications. DK's long-term research goal is to democratize AI to serve a broader range of populations and domains. More information can be found on DK’s personal website at http://www.personal.psu.edu/dux19/
Presentations

Towards Robust Pruning: An Adaptive Knowledge-Retention Pruning Strategy for Language Models | VIDEO
Jianwei Li and 3 other authors

Sparse Progressive Distillation: Resolving Overfitting under Pretrain-and-Finetune Paradigm
Shaoyi Huang and 10 other authors

Rethinking Network Pruning -- under the Pre-train and Fine-tune Paradigm
Dongkuan Xu and 3 other authors

Longitudinal Deep Kernel Gaussian Process Regression
Junjie Liang and 3 other authors

Transformer-Style Relational Reasoning with Dynamic Memory Updating for Temporal Network Modeling
Dongkuan Xu and 5 other authors

Multi-Task Recurrent Modular Networks
Dongkuan Xu and 9 other authors

How Do We Move: Modeling Human Movement with System Dynamics
Hua Wei and 3 other authors