Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kang, Minsoo | - |
dc.contributor.author | Kang, Minkoo | - |
dc.contributor.author | Lee, Seong-Whan | - |
dc.contributor.author | Kim, Suhyun | - |
dc.date.accessioned | 2024-06-07T02:30:06Z | - |
dc.date.available | 2024-06-07T02:30:06Z | - |
dc.date.created | 2024-06-07 | - |
dc.date.issued | 2024-06 | - |
dc.identifier.issn | 0262-8856 | - |
dc.identifier.uri | https://pubs.kist.re.kr/handle/201004/150019 | - |
dc.description.abstract | The inherent complexity and extensive architecture of deep neural networks often lead to overfitting, compromising their ability to generalize to new, unseen data. One of the regularization techniques, data augmentation, is now considered vital to alleviate this, and mixup, which blends pairs of images and labels, has proven effective in enhancing model generalization. Recently, incorporating saliency in mixups has shown performance gains by retaining salient regions in mixed results. While these methods have become mainstream at the input level, their applications at the feature level remain under-explored. Our observations indicate that outcomes from naive applications of input saliency-based methods did not consistently lead to enhancements in performance. In this paper, we attribute these observations primarily to two challenges: 'Hard Boundary Issue' and 'Saliency Mismatch.' The Hard Boundary Issue describes a situation where masks with distinct, sharp edges work well at the input level, but lead to unintended distortions in the deeper layers. The Saliency Mismatch points to the disparity between saliency masks generated from input images and the saliency of feature maps. To tackle these challenges, we present a novel method called 'attention-based mixup mask adaptation' (MMA). This approach employs an attention mechanism to effectively adapt mixup masks, which are designed to maximize saliency at the input level, for feature augmentation purposes. We reduce the Saliency Mismatch problem by incorporating the spatial significance of the feature map into the mixup mask. Additionally, we address the Hard Boundary Issue by applying softmax to smoothen the adjusted mixup mask. Through comprehensive experiments, we validate our observations and confirm the effectiveness of applying MMA to saliency-aware mixup approaches at the feature level, as evidenced by the performance improvements on multiple benchmarks and the robustness improvements against corruption and deformation. | - |
dc.language | English | - |
dc.publisher | Elsevier BV | - |
dc.title | Mixup Mask Adaptation: Bridging the gap between input saliency and representations via attention mechanism in feature mixup | - |
dc.type | Article | - |
dc.identifier.doi | 10.1016/j.imavis.2024.105013 | - |
dc.description.journalClass | 1 | - |
dc.identifier.bibliographicCitation | Image and Vision Computing, v.146 | - |
dc.citation.title | Image and Vision Computing | - |
dc.citation.volume | 146 | - |
dc.description.isOpenAccess | Y | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.identifier.wosid | 001231786400001 | - |
dc.identifier.scopusid | 2-s2.0-85190334440 | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Artificial Intelligence | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Software Engineering | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Theory & Methods | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Optics | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Optics | - |
dc.type.docType | Article | - |
dc.subject.keywordPlus | NETWORKS | - |
dc.subject.keywordAuthor | Regularization | - |
dc.subject.keywordAuthor | Data augmentation | - |
dc.subject.keywordAuthor | Mixup | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.