Director: Hanseok Ko, Ph.D.
The Multimodal Artificial Intelligence Laboratory conducts advanced research at the intersection of perception, reasoning, and autonomous action across multiple sensory modalities. We develop intelligent systems that integrate and interpret data from diverse sources to operate effectively in complex, dynamic environments by exploring advanced machine learning algorithms and human-computer interactions.
Our work includes:
- Real-time multimodal sensor fusion and computer vision for robust perception.
- Advanced acoustic and audio signal processing for accurate modeling, detection, and tracking.
- Development and fine-tuning of private large language models (LLMs) and agentic AI systems for autonomous planning, adaptive decision-making, and dialogue management and Implementation of retrieval-augmented generation (RAG) and long-term memory architectures to enhance AI capabilities.
- Robotics and autonomous vehicles (ground, aerial, and underwater) for navigation in unstructured terrain and adverse environment.
We build AI that doesn’t just analyze the world but engages with it intelligently.