Explainable AI
Explainable AI aims to improve the transparency of AI decisions through interpretable methods. We explore vulnerabilities in current explanation methods using adversarial attacks and develop intuitive visualizations to enhance user trust.
Responsible AI
Safe Image Generation
Safe Image Generation focuses on developing algorithms to create unbiased and ethical images. We work on automatically identifying and eliminating harmful or biased outputs to ensure responsible AI usage.
Machine Unlearning
Machine Unlearning deals with the secure removal of specific data from AI models. Our research includes developing methods to comply with privacy regulations and provide transparent control of data.
Efficient AI
Learned Image Compression
Learned Image Compression utilizes AI techniques to significantly reduce the size of image data while maintaining visual quality. We focus on compression methods suitable for real-time applications and resource-limited environments.
Knowledge Distillation
Knowledge Distillation transfers knowledge from large AI models to smaller, efficient versions. Our goal is to create lightweight models ideal for mobile and edge computing, ensuring high accuracy and robustness.
Brain-Computer Interfaces
BCI is a powerful way to understand users' perception directly by monitoring brain activities. We have been working on investigating effective BCI techniques using EEG and fNIRS for implicit monitoring of users' perception of given multimedia content. Here, perceptual dimensions include satisfaction, quality, immersiveness, emotion, sense of reality, etc. In particular, we focus on understanding perception for novel types of media such as 3D, high dynamic range (HDR), 4K UHD, etc. Furthermore, peripheral physiological signals such as skin temperature, skin conductance, and heart rate are also considered, which have potential to complement brain activities.
Dynamic Vision Sensors
Dynamic Vision Sensors detect rapid visual changes through event-based imaging. We research algorithms for efficient interpretation of DVS data, applicable to robotics, autonomous driving, and real-time monitoring due to their speed and efficiency.
Multimodal Deep Learning
Often, heterogeneous sensors are simultaneously involved, which produce multimodal data. For instance, human actions can be recorded by cameras, depth cameras, accelerometers, and gyroscopes at the same time. They are mutually complementary, thus learning from them is beneficial to maximize performance. We have recently developed a deep learning architecture that can integrate multimodal data. It is not only effective in terms of performance but also robust against partial absence of data and modalities.