EMNLP 2025

November 06, 2025

Suzhou, China

Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.

Recent advancements in large language models (LLMs) have shown promise in feature engineering for tabular data, but concerns about their reliability persist, especially due to variability in generated outputs. We introduce a multi-level diagnosis and evaluation framework to assess the robustness of LLMs in feature engineering, focusing on identifying key variables, relationships, and decision boundaries for predicting target classes across diverse domains. We demonstrate that the robustness of LLMs varies significantly over different datasets, and that high-quality LLM-generated features can improve prediction performance by up to 10.52%. This work opens a new direction for assessing and enhancing the reliability of LLM-driven feature engineering.

Downloads

SlidesPaperTranscript English (automatic)

Next from EMNLP 2025

Dream to Chat: Model-based Reinforcement Learning on Dialogues with User Belief Modeling
poster

Dream to Chat: Model-based Reinforcement Learning on Dialogues with User Belief Modeling

EMNLP 2025

+7
Teng Chen and 9 other authors

06 November 2025

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Presentations
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2026 Underline - All rights reserved