Content not yet available
This lecture has no active video or poster.
Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.
The automation of diagram generation has gained significant attention in recent years.In previous work on automated diagram generation, most studies focused on generating diagrams from natural language descriptions. Many of these methods produced diagrams that were not user-friendly for editing, lacking drag-and-drop functionality. This paper proposes a novel task of generating editable, high-fidelity diagrams from either text descriptions or raster images (non-editable images), and it may be the first to introduce the tasks of diagram restoration and style transfer in this context. To address these tasks, we created the Diagram-mxGraph dataset, covering three main tasks: diagram restoration, text-to-diagram generation, and diagram style transfer. We introduced two key innovations: Fine-grained Adaptive Background Suppression (FABS) and Component-Aware Adaptive Loss (CAAL). By leveraging the capabilities of pre-trained Vision Transformers (ViTs) and the Diagram Adapter module, we effectively align the features of diagrams with a Large Language Model (LLM) to generate mxGraph format textual descriptions, ultimately producing editable diagrams.
