<?xml version="1.0" encoding="utf-8" standalone="no"?>
<dublin_core schema="dc">
<dcvalue element="contributor" qualifier="author">Kwon,&#x20;Jangho</dcvalue>
<dcvalue element="contributor" qualifier="author">Ha,&#x20;Jihyeon</dcvalue>
<dcvalue element="contributor" qualifier="author">Kim,&#x20;Da-Hye</dcvalue>
<dcvalue element="contributor" qualifier="author">Choi,&#x20;Jun&#x20;Won</dcvalue>
<dcvalue element="contributor" qualifier="author">Kim,&#x20;Laehyun</dcvalue>
<dcvalue element="date" qualifier="accessioned">2024-01-19T13:34:02Z</dcvalue>
<dcvalue element="date" qualifier="available">2024-01-19T13:34:02Z</dcvalue>
<dcvalue element="date" qualifier="created">2022-01-10</dcvalue>
<dcvalue element="date" qualifier="issued">2021-10</dcvalue>
<dcvalue element="identifier" qualifier="issn">2169-3536</dcvalue>
<dcvalue element="identifier" qualifier="uri">https:&#x2F;&#x2F;pubs.kist.re.kr&#x2F;handle&#x2F;201004&#x2F;116340</dcvalue>
<dcvalue element="description" qualifier="abstract">We&#x20;present&#x20;a&#x20;glasses-type&#x20;wearable&#x20;device&#x20;to&#x20;detect&#x20;emotions&#x20;from&#x20;a&#x20;human&#x20;face&#x20;in&#x20;an&#x20;unobtrusive&#x20;manner.&#x20;The&#x20;device&#x20;is&#x20;designed&#x20;to&#x20;gather&#x20;multi-channel&#x20;responses&#x20;from&#x20;the&#x20;user&amp;apos;s&#x20;face&#x20;naturally&#x20;and&#x20;continuously&#x20;while&#x20;he&#x2F;she&#x20;is&#x20;wearing&#x20;it.&#x20;The&#x20;multi-channel&#x20;facial&#x20;responses&#x20;consist&#x20;of&#x20;local&#x20;facial&#x20;images&#x20;and&#x20;biosignals&#x20;including&#x20;electrodermal&#x20;activity&#x20;(EDA)&#x20;and&#x20;photoplethysmogram&#x20;(PPG).&#x20;We&#x20;had&#x20;conducted&#x20;experiments&#x20;to&#x20;determine&#x20;the&#x20;optimal&#x20;positions&#x20;of&#x20;EDA&#x20;sensors&#x20;on&#x20;the&#x20;wearable&#x20;device&#x20;because&#x20;EDA&#x20;signal&#x20;quality&#x20;is&#x20;very&#x20;sensitive&#x20;to&#x20;the&#x20;sensing&#x20;position.&#x20;In&#x20;addition&#x20;to&#x20;the&#x20;physiological&#x20;data,&#x20;the&#x20;device&#x20;can&#x20;capture&#x20;the&#x20;image&#x20;region&#x20;representing&#x20;local&#x20;facial&#x20;expressions&#x20;around&#x20;the&#x20;left&#x20;eye&#x20;via&#x20;a&#x20;built-in&#x20;camera.&#x20;In&#x20;this&#x20;study,&#x20;we&#x20;developed&#x20;and&#x20;validated&#x20;an&#x20;algorithm&#x20;to&#x20;recognize&#x20;emotions&#x20;using&#x20;multi-channel&#x20;responses&#x20;obtained&#x20;from&#x20;the&#x20;device.&#x20;The&#x20;results&#x20;show&#x20;that&#x20;the&#x20;emotion&#x20;recognition&#x20;algorithm&#x20;using&#x20;only&#x20;local&#x20;facial&#x20;images&#x20;has&#x20;an&#x20;accuracy&#x20;of&#x20;76.09%&#x20;at&#x20;classifying&#x20;emotions.&#x20;Using&#x20;multi-channel&#x20;data&#x20;including&#x20;EDA&#x20;and&#x20;PPG,&#x20;this&#x20;accuracy&#x20;was&#x20;increased&#x20;by&#x20;8.46%&#x20;compared&#x20;to&#x20;using&#x20;the&#x20;local&#x20;facial&#x20;expression&#x20;alone.&#x20;This&#x20;glasses-type&#x20;wearable&#x20;system&#x20;measuring&#x20;multi-channel&#x20;facial&#x20;responses&#x20;in&#x20;a&#x20;natural&#x20;manner&#x20;is&#x20;very&#x20;useful&#x20;for&#x20;monitoring&#x20;a&#x20;user&amp;apos;s&#x20;emotions&#x20;in&#x20;daily&#x20;life,&#x20;which&#x20;has&#x20;a&#x20;huge&#x20;potential&#x20;for&#x20;use&#x20;in&#x20;the&#x20;healthcare&#x20;industry.</dcvalue>
<dcvalue element="language" qualifier="none">English</dcvalue>
<dcvalue element="publisher" qualifier="none">IEEE-INST&#x20;ELECTRICAL&#x20;ELECTRONICS&#x20;ENGINEERS&#x20;INC</dcvalue>
<dcvalue element="subject" qualifier="none">HEART-RATE-VARIABILITY</dcvalue>
<dcvalue element="subject" qualifier="none">SENSOR</dcvalue>
<dcvalue element="title" qualifier="none">Emotion&#x20;Recognition&#x20;Using&#x20;a&#x20;Glasses-Type&#x20;Wearable&#x20;Device&#x20;via&#x20;Multi-Channel&#x20;Facial&#x20;Responses</dcvalue>
<dcvalue element="type" qualifier="none">Article</dcvalue>
<dcvalue element="identifier" qualifier="doi">10.1109&#x2F;ACCESS.2021.3121543</dcvalue>
<dcvalue element="description" qualifier="journalClass">1</dcvalue>
<dcvalue element="identifier" qualifier="bibliographicCitation">IEEE&#x20;ACCESS,&#x20;v.9,&#x20;pp.146392&#x20;-&#x20;146403</dcvalue>
<dcvalue element="citation" qualifier="title">IEEE&#x20;ACCESS</dcvalue>
<dcvalue element="citation" qualifier="volume">9</dcvalue>
<dcvalue element="citation" qualifier="startPage">146392</dcvalue>
<dcvalue element="citation" qualifier="endPage">146403</dcvalue>
<dcvalue element="description" qualifier="journalRegisteredClass">scie</dcvalue>
<dcvalue element="description" qualifier="journalRegisteredClass">scopus</dcvalue>
<dcvalue element="identifier" qualifier="wosid">000714706800001</dcvalue>
<dcvalue element="identifier" qualifier="scopusid">2-s2.0-85118246850</dcvalue>
<dcvalue element="relation" qualifier="journalWebOfScienceCategory">Computer&#x20;Science,&#x20;Information&#x20;Systems</dcvalue>
<dcvalue element="relation" qualifier="journalWebOfScienceCategory">Engineering,&#x20;Electrical&#x20;&amp;&#x20;Electronic</dcvalue>
<dcvalue element="relation" qualifier="journalWebOfScienceCategory">Telecommunications</dcvalue>
<dcvalue element="relation" qualifier="journalResearchArea">Computer&#x20;Science</dcvalue>
<dcvalue element="relation" qualifier="journalResearchArea">Engineering</dcvalue>
<dcvalue element="relation" qualifier="journalResearchArea">Telecommunications</dcvalue>
<dcvalue element="type" qualifier="docType">Article</dcvalue>
<dcvalue element="subject" qualifier="keywordPlus">HEART-RATE-VARIABILITY</dcvalue>
<dcvalue element="subject" qualifier="keywordPlus">SENSOR</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">Wearable&#x20;computers</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">Emotion&#x20;recognition</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">Sensors</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">Cameras</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">Biomedical&#x20;monitoring</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">Glass</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">Motion&#x20;pictures</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">Wearable&#x20;device</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">emotion&#x20;recognition</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">affective&#x20;computing</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">facial&#x20;expression</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">biosignal</dcvalue>
<dcvalue element="subject" qualifier="keywordAuthor">physiological&#x20;responses</dcvalue>
</dublin_core>
