{"id":199589,"date":"2025-07-11T16:49:04","date_gmt":"2025-07-11T07:49:04","guid":{"rendered":"http:\/\/ee.presscat.kr\/?post_type=research-achieve&#038;p=199589"},"modified":"2026-04-13T03:48:14","modified_gmt":"2026-04-12T18:48:14","slug":"ee-professor-jung-woo-chois-research-team-wins-the-ieee-dcase-2025-challenge-the-worlds-leading-acoustic-ai-competition","status":"publish","type":"research-achieve","link":"http:\/\/ee.presscat.kr\/en\/research-achieve\/ee-professor-jung-woo-chois-research-team-wins-the-ieee-dcase-2025-challenge-the-worlds-leading-acoustic-ai-competition\/","title":{"rendered":"EE Professor Jung-Woo Choi\u2019s Research Team Wins the IEEE DCASE 2025 Challenge, the World\u2019s Leading Acoustic AI Competition"},"content":{"rendered":"<figure id=\"attachment_199578\" aria-describedby=\"caption-attachment-199578\" style=\"width: 750px\" class=\"wp-caption aligncenter\"><img fetchpriority=\"high\" decoding=\"async\" class=\"size-full wp-image-199578\" src=\"http:\/\/ee.presscat.kr\/wp-content\/uploads\/2025\/07\/\ucd5c\uc815\uc6b0-\uad50\uc218\ub2d8\ud300_750.jpg\" alt=\"\" width=\"750\" height=\"587\" title=\"\"><figcaption id=\"caption-attachment-199578\" class=\"wp-caption-text\"><span style=\"color: #000000;font-size: 12pt\">&lt;(Left to right) Younghoo Kwon (Integrated Master&#8217;s and Ph.D. program), Dohwan Kim (Master&#8217;s program), Professor Jung-Woo Choi, Dongheon Lee (Ph.D.)&gt;<\/span><\/figcaption><\/figure>\n<p><span style=\"color: #000000;font-size: 14pt\">Acoustic source separation and classification is a key next-generation AI technology for early detection of anomalies in drone operations piping faults or border surveillance and for enabling spatial audio editing in AR VR content production.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;font-size: 14pt\">Professor Jung-Woo Choi\u2019s research team from the School of Electrical Engineering won first place in the \u201cSpatial Semantic Segmentation of Sound Scenes\u201d task of the \u201cIEEE DCASE Challenge 2025.\u201d<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;font-size: 14pt\">This year\u2019s challenge featured 86 teams competing across six tasks. In their first-ever participation, KAIST\u2019s team ranked first in Task 4: Spatial Semantic Segmentation of Sound Scenes\u2014a highly demanding task requiring the analysis of spatial information in multi-channel audio signals with overlapping sound sources. The goal was to separate individual sounds and classify them into 18 predefined categories. The team, composed of Dr. Dongheon Lee, integrated MS-PhD student Younghoo Kwon, and MS student Dohwan Kim, will present their results at the DCASE Workshop in Barcelona this October.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;font-size: 14pt\">Earlier this year, Dr. Dongheon Lee developed a state-of-the-art sound source separation AI combining Transformer and Mamba architectures. Furthermore, at the challenge, led by Younghoo Kwon, the team established the chain-of-inference architecture that first separates waveforms and source types and then refines the estimation by utilizing the estimated waveforms and classes as clues for target signal extraction in the next stage.<\/span><\/p>\n<p>&nbsp;<\/p>\n<figure id=\"attachment_199580\" aria-describedby=\"caption-attachment-199580\" style=\"width: 1033px\" class=\"wp-caption aligncenter\"><img decoding=\"async\" class=\"size-full wp-image-199580\" src=\"http:\/\/ee.presscat.kr\/wp-content\/uploads\/2025\/07\/1.-\uc5ec\ub7ec-\uc18c\ub9ac\uac00-\ud63c\ud569\ub41c-\uc74c\ud5a5-\uc7a5\uba74\uc758-\uc608.jpg\" alt=\"\" width=\"1033\" height=\"544\" title=\"\"><figcaption id=\"caption-attachment-199580\" class=\"wp-caption-text\"><span style=\"color: #000000;font-size: 12pt\">&lt; Figure 1. Example of an acoustic scene with multiple mixed sounds &gt;<\/span><\/figcaption><\/figure>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;font-size: 14pt\">This chain-of-inference approach is inspired by human\u2019s auditory scene analysis mechanism that isolates individual sounds by focusing on incomplete clues such as sound type, rhythm, or direction.\u00a0<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;font-size: 14pt\">In the evaluation metric CA-SDRi (Class-aware Signal-to-distortion Ratio improvement)*, the team was the only participant to achieve a double-digit improvement of 11 dB, demonstrating their technical excellence.\u00a0*CA-SDRi (Class-aware Signal-to-distortion Ratio improvement) measures how much clearer and less distorted the target sound is compared with the original mix.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;font-size: 14pt\">Professor Choi remarked, \u201cI am proud that our team\u2019s world leading acoustic separation AI models over the past three years have now received formal recognition. Despite the greatly increased difficulty and the limited development window due to other conference schedules and final exams, each member demonstrated focused research that led to first place.\u201d<\/span><\/p>\n<p>&nbsp;<\/p>\n<figure id=\"attachment_199582\" aria-describedby=\"caption-attachment-199582\" style=\"width: 1326px\" class=\"wp-caption aligncenter\"><img decoding=\"async\" class=\"size-full wp-image-199582\" src=\"http:\/\/ee.presscat.kr\/wp-content\/uploads\/2025\/07\/2.-\ud63c\ud569-\uc74c\uc6d0\uc73c\ub85c\ubd80\ud130-\ubd84\ub9ac\ub41c-\uc74c\uc6d0\ub4e4\uc758-\uc2dc\uac04-\uc8fc\ud30c\uc218-\ud328\ud134.jpg\" alt=\"\" width=\"1326\" height=\"687\" title=\"\"><figcaption id=\"caption-attachment-199582\" class=\"wp-caption-text\"><span style=\"color: #000000;font-size: 12pt\">&lt; Figure 2. Time frequency patterns of separated sound sources &gt;<\/span><\/figcaption><\/figure>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;font-size: 14pt\">The \u201cIEEE DCASE Challenge 2025\u201d was held online from April 1<sup>st<\/sup>\u00a0to June 15<sup>th<\/sup>\u00a0for submissions, with results announced on June 30<sup>th<\/sup>. Since its inception in 2013 under the IEEE Signal Processing Society, the challenge has served as a global stage for AI models in the acoustic field.<\/span><\/p>\n<p><span style=\"color: #000000;font-size: 14pt\"><a style=\"color: #000000\" href=\"https:\/\/dcase.community\/challenge2025\/task-spatial-semantic-segmentation-of-sound-scenes\" target=\"_blank\" rel=\"noopener noreferrer\">Go to the IEEE DCASE Challenge 2025 website (Click)<\/a><\/span><\/p>\n<p><span style=\"color: #000000;font-size: 14pt\">This research was supported by the National Research Foundation of Korea\u2019s Mid-Career Researcher Program and STEAM Research Project, funded by the Ministry of Education, and the Future Defense Research Center, funded by the Defense Acquisition Program Administration and the Agency for Defense Development.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><figure id=\"attachment_199584\" aria-describedby=\"caption-attachment-199584\" style=\"width: 1046px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"size-full wp-image-199584\" src=\"http:\/\/ee.presscat.kr\/wp-content\/uploads\/2025\/07\/\uadf8\ub9bc-3.-\uc5f0\uad6c\uc9c4\uc774-\uac1c\ubc1c\ud55c-\uc74c\ud5a5\uc758-\ubd84\ub9ac-\ubc0f-\ubd84\ub958-AI-\uad6c\uc870.png\" alt=\"\" width=\"1046\" height=\"604\" title=\"\"><figcaption id=\"caption-attachment-199584\" class=\"wp-caption-text\"><span style=\"color: #000000;font-size: 12pt\">&lt; Figure 3. AI architecture for sound separation and classification &gt;<\/span><\/figcaption><\/figure> <figure id=\"attachment_199586\" aria-describedby=\"caption-attachment-199586\" style=\"width: 1044px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"size-full wp-image-199586\" src=\"http:\/\/ee.presscat.kr\/wp-content\/uploads\/2025\/07\/images_000102_image333.png.png\" alt=\"\" width=\"1044\" height=\"520\" title=\"\"><figcaption id=\"caption-attachment-199586\" class=\"wp-caption-text\"><span style=\"color: #000000;font-size: 12pt\">&lt; Competition Results Rankings. Higher CA-SDRi indicates a better score (Unit: decibels dB) &gt;<\/span><\/figcaption><\/figure><\/p>\n","protected":false},"excerpt":{"rendered":"<p>341<\/p>\n","protected":false},"featured_media":199590,"template":"","research_category":[364],"class_list":["post-199589","research-achieve","type-research-achieve","status-publish","has-post-thumbnail","hentry","research_category-multimedia-signal-processing-en"],"acf":[],"_links":{"self":[{"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/research-achieve\/199589","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/research-achieve"}],"about":[{"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/types\/research-achieve"}],"wp:featuredmedia":[{"embeddable":true,"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/media\/199590"}],"wp:attachment":[{"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/media?parent=199589"}],"wp:term":[{"taxonomy":"research_category","embeddable":true,"href":"http:\/\/ee.presscat.kr\/en\/wp-json\/wp\/v2\/research_category?post=199589"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}