Full metadata record

DC Field Value Language
dc.contributor.authorKim, Jisoo-
dc.contributor.authorYoo, Byounghyun-
dc.date.accessioned2025-09-04T02:30:25Z-
dc.date.available2025-09-04T02:30:25Z-
dc.date.created2025-09-04-
dc.date.issued2025-08-
dc.identifier.issn2288-4300-
dc.identifier.urihttps://pubs.kist.re.kr/handle/201004/153118-
dc.description.abstractHuman Activity Recognition (HAR) plays a crucial role in identifying and digitizing human behaviors. Among various approaches, sound-based HAR offers distinct advantages, such as overcoming visual limitations and enabling recognition in diverse environments. This study introduces an innovative application of sound segmentation with SegNet, originally designed for image segmentation, to sound-based HAR. Traditionally, labeling sound data has been challenging due to its limited scope, often restricted to specific events or time frames. To address this issue, a novel labeling approach was developed, allowing detailed annotations across the entire temporal and frequency domains. This method facilitates the use of SegNet, which requires pixel-level labeling for accurate segmentation, leading to more granular and explainable activity recognition. A dataset comprising six distinct human activities-speech, groaning, screaming, coughing, toilet and snoring-was constructed to enable comprehensive evaluation. The trained neural network, utilizing this annotated dataset, achieved F1 scores ranging from 0.68 to 0.95. The model's practical applicability was further validated through recognition tests conducted in a professional office environment. This study presents a novel framework for quantifying daily human activities through sound segmentation, contributing to advancements in intelligent system technology.-
dc.languageEnglish-
dc.publisher한국CDE학회-
dc.titleGranular and explainable human activity recognition through sound segmentation and deep learning-
dc.typeArticle-
dc.identifier.doi10.1093/jcde/qwaf075-
dc.description.journalClass1-
dc.identifier.bibliographicCitationJournal of Computational Design and Engineering, v.12, no.8, pp.252 - 269-
dc.citation.titleJournal of Computational Design and Engineering-
dc.citation.volume12-
dc.citation.number8-
dc.citation.startPage252-
dc.citation.endPage269-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.identifier.wosid001555098700001-
dc.identifier.scopusid2-s2.0-105014025877-
dc.relation.journalWebOfScienceCategoryComputer Science, Interdisciplinary Applications-
dc.relation.journalWebOfScienceCategoryEngineering, Multidisciplinary-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalResearchAreaEngineering-
dc.type.docTypeArticle-
dc.subject.keywordPlusEVENT DETECTION-
dc.subject.keywordPlusFALL DETECTION-
dc.subject.keywordPlusCLASSIFICATION-
dc.subject.keywordPlusLOCALIZATION-
dc.subject.keywordPlusAUDIO-
dc.subject.keywordPlusCNN-
dc.subject.keywordAuthorSound Segmentation-
dc.subject.keywordAuthorHuman Activity Recognition-
dc.subject.keywordAuthorSound Data Labeling-
dc.subject.keywordAuthorExplainable Artificial Intelligence-
Appears in Collections:
KIST Article > Others
Files in This Item:
There are no files associated with this item.
Export
RIS (EndNote)
XLS (Excel)
XML

qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE