My research interests lie at the intersection of cognitive neuroscience, learning, and robotics. I am particularly focused on how robots can acquire, adapt, and retain motor skills similarly to how humans do. Inspired by the Language of Thought theory, my work is trying to explore the primitives of motor skills, both at the perception level (such as dominant features of object affordance) and the movement level (such as muscle synergies). By leveraging VLMs, I aim to build generalized robotic skills that can be adapted and transferred across a variety of tasks based on these primitives. Additionally, I'm working on the RLHF process of CogVLM.
A proactive robot-human interaction framework that uses Vision Language Models and Affective Theory of Mind to autonomously detect and fulfill unspoken human needs, achieving high human satisfaction in need detection and task execution.
This study reveals a functional gradient along the hippocampus that encodes detailed experiences into abstract knowledge for decision-making, showing that the posterior hippocampus handles memory formation while the anterior hippocampus utilizes abstract knowledge for future planning.
This study uses computational modeling to show how selective attention and contextual inference help construct efficient relational memory, enabling humans to navigate complex information while overcoming memory capacity limitations.
This work introduces ES-ImageNet, a large-scale event-stream dataset generated from ILSVRC2012 using the Omnidirectional Discrete Gradient (ODG) algorithm, providing a low-cost, high-speed solution for advancing neuromorphic vision processing with spiking neural networks.
This study systematically compares spiking neural networks (SNNs) and recurrent neural networks (RNNs) on neuromorphic vision data, highlighting their similarities and differences in spatiotemporal processing through unified benchmarking experiments.