Content not yet available
This lecture has no active video or poster.
Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.
Autonomous driving must cope with motion blur, low light, and dynamic agents, where RGB frames and event cameras offer complementary strengths. This thesis investigates how to fuse them across the perception–reasoning–planning pipeline. It introduces FlexEvent, a frequency-robust detector with adaptive fusion and label-efficient training; Talk2Event, the first benchmark for event–language grounding with attribute-aware modeling; and the ongoing EventChat, an event–frame VLM for perception, spatial relations, and ego reasoning. Future work will extend this framework with iterative perception and reinforcement learning for long-horizon decision making. Together, these efforts aim to deliver robust perception, interpretable reasoning, and planning support through event–frame fusion.
