
Sheng Shen
transformers
data augmentation
low resource
large language models
bloom
synthetic data
reservoir computing
training efficiency
finetuning
multilingual models
gpt
llm
instruction-tuning
multitask finetuning
bloomz
6
presentations
1
number of views
SHORT BIO
First-year CS Ph.D. student at UC Berkeley, focusing on efficient NLP and Vision-and-Language.
Presentations

LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement
Nicholas Lee and 8 other authors

Crosslingual Generalization through Multitask Finetuning
Niklas Muennighoff and 12 other authors

What’s Hidden in a One-layer Randomly Weighted Transformer?
Zhewei Yao and 1 other author

Reservoir Transformers
Sheng Shen and 1 other author

Noisy Self-Knowledge Distillation for Text Summarization
Yang Liu and 2 other authors

ADAHESSIAN: An Adaptive Second Order Optimizer for Machine Learning
Zhewei Yao and 5 other authors