
JingBo Zhu
machine translation
large language model
transformer
abstractive summarization
multi-task learning
generation
pre-training
ethics
knowledge distillation
curriculum learning
speech translation
multimodal machine translation
multi-view learning
translation memory
low-resource
19
presentations
8
number of views
Presentations

RoVRM: A Robust Visual Reward Model Optimized via Auxiliary Textual Preference Data
Yang Gan and 9 other authors

Forgetting Curve: A Reliable Method for Evaluating Memorization Capability for Long-Context Models
Xinyu Liu and 7 other authors

Revealing the Parallel Multilingual Learning within Large Language Models
Yongyu Mu and 10 other authors

PartialFormer: Modeling Part Instead of Whole for Machine Translation
Tong Zheng and 6 other authors

EIT: Enhanced Interactive Transformer
Tong Zheng and 4 other authors

Rethinking and Improving Multi-task Learning for End-to-end Speech Translation | VIDEO
Yuhao Zhang and 6 other authors

Augmenting Large Language Model Translators via Translation Memories
Yongyu Mu and 8 other authors

Prompting Neural Machine Translation with Translation Memories
JingBo Zhu and 5 other authors

Improving End-to-end Speech Translation by Leveraging Auxiliary Speech and Text Data
Yuhao Zhang and 5 other authors

ODE Transformer: An Ordinary Differential Equation-Inspired Model for Sequence Generation
Bei Li and 9 other authors

On Vision Features in Multimodal Machine Translation
Bei Li and 6 other authors

Bag of Tricks for Optimizing Transformer Efficiency
Ye Lin and 3 other authors

Weight Distillation: Transferring the Knowledge in Neural Network Parameters
Ye Lin and 6 other authors

Stacked Acoustic-and-Textual Encoding: Integrating the Pre-trained Models into Speech Translation Encoders
Chen Xu and 7 other authors

An Efficient Transformer Decoder with Compressed Sub-Layers
Yanyang Li and 3 other authors

Learning Light-Weight Translation Models from Deep Transformer
Bei Li and 6 other authors