Dr. Hansung Kim is an Associate Professor in the School of Electronics & Computer Science at the University of Southampton. He has published over 100 peer-reviewed papers (Over 60 papers as the first author). These publications include top journals and conferences in computer vision such as IJCV, IEEE Trans. Image Processing, CVPR, ICCV and ECCV. His research over the last 20 years spanned a wide range of 3D computer vision and neighbouring areas including camera systems, spherical imaging, 3D reconstruction, Augmented/Virtual Reality (AR/VR), multi-modal processing, audio-visual data processing, artificial Intelligence (AI) and media production.
- 3D Computer Vision
- Artificial intelligence (AI) for scene understanding
- Audio-visual data processing
- Virtual Reality (VR)
- Multi-modal data analysis
Dr. Hansung Kim’s research aims to unlock the creative potential of Audio-Visual Machine Perception (AI) to deliver a step change in immersive VR experiences for entertainment and training. This requires highly intelligent technologies, including machine learning (deep audio-visual data learning), computer vision (object and action recognition), and audio signal processing (audio/speech analysis).
His current research addresses the open-challenge of machine understanding of complex dynamic real-world scenes combining the complementary information available from audio and visual sensor to achieve robust interpretation. These research advances are of central interest to both the audio and vision research communities and bring together advances in machine perception. Joint audio-visual processing is essential to overcome the inherent ambiguities in either sensing modality.
COMP6223 Computer Vision
COMP3200: Part III Individual Project
COMP6200 MSc Project
ELEC6200: Group Design Project
I am an Associate Professor in the School of Electronics & Computer Science at the University of Southampton. I received the MS and Ph.D degrees in electronic and electrical engineering from Yonsei University, Seoul, Korea, in 2001 and 2005, respectively. I was employed as a researcher of Knowledge Science Lab (KSL) at Advanced Telecommunications Research Institute International (ATR), Japan, from 2005 to 2008, and as a senior research fellow (RA-2) at the Centre for Vision, Speech, and Signal Processing (CVSSP) at the University of Surrey, UK, from 2008 to 2020.
My researches for last 20 years include 3D computer vision, Stereo image processing, 3D reconstruction, Augmented/Mixed Reality, Multi-modal data processing, Audio-visual data processing and Media production with over 100 published articles in 16 projects. Recently, I was involved in EU FP7 IMPART project for multi-modal big data processing as a work package leader in 2012-2015. I lead the computer vision team in the UK EPSRC project - S3A: Future Spatial Audio (2013-2019). I also served as a Principal Investigator for BEIS Global Partnership programme - Intelligent Virtual Reality: Deep Audio-Visual Representation Learning for Multimedia Perception and Reproduction (2017-2019). Currently I'm serving as the Principal Investigator for EPSRC NIA project for Immersive Audio-Visual 3D Scene Reproduction (2021-2024) and Real-time 3D scene understanding and human pose estimation (2022-2023).
- 3rd Prize (2020)
- Best Paper Award (2013)
- Best paper nomination (2019)
- Best Paper Award from IEEE Signal Processing Society (2013)
- Best Paper nomination at IEEE VR conference (2019)
- - Third Prize at the AHSC Competition for Innovative Ideas to Solve Real World Problems Related to COVID-19 (2020)