profile picture

Chenglong Wang

Post-graduate student @ Northeastern University

reinforcement learning

knowledge distillation

glue

sequence generation

large language model

actor-critic

dynamic sampling

instruction tuning

evol instruct

5

presentations

1

number of views

SHORT BIO

I am currently a third-year Ph.D. student at the Natural Language Processing Laboratory, Department of Computer Science and Technology, Northeastern University, China, under the guidance of Professor Tong Xiao and Professor Jingbo Zhu.

I completed my undergraduate studies at Northeastern University and obtained my bachelor's degree in Computer Science and Technology. Subsequently, I pursued my master's degree, also at Northeastern University, graduating in 2020.

My research interests focus on modeling methods for deep reinforcement learning on sequence generation tasks, with a particular emphasis on two key areas: 1) enhancing the training efficiency of sequence generation models based on reinforcement learning, and 2) constructing and training reward models for deep reinforcement learning.

Presentations

RoVRM: A Robust Visual Reward Model Optimized via Auxiliary Textual Preference Data

Yang Gan and 9 other authors

Revealing the Parallel Multilingual Learning within Large Language Models

Yongyu Mu and 10 other authors

The NiuTrans System for the WMT 2021 Efficiency Task

Chenglong Wang

Improved Knowledge Distillation for Pre-trained Language Models via Knowledge Selection

Chenglong Wang

ESRL: Efficient Sampling-Based Reinforcement Learning for Sequence Generation

Chenglong Wang and 7 other authors

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Presentations
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2025 Underline - All rights reserved