Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Ha, Junhyoung | - |
dc.contributor.author | An, Byungchul | - |
dc.contributor.author | Kim, Soon kyum | - |
dc.date.accessioned | 2024-01-12T02:31:59Z | - |
dc.date.available | 2024-01-12T02:31:59Z | - |
dc.date.created | 2022-09-25 | - |
dc.date.issued | 2023-03 | - |
dc.identifier.issn | 1551-3203 | - |
dc.identifier.uri | https://pubs.kist.re.kr/handle/201004/75792 | - |
dc.description.abstract | In a graph search algorithm, a given environment is represented as a graph comprising a set of feasible system configurations and their neighboring connections. A path is generated by connecting the initial and goal configurations through graph exploration, whereby the path is often desired to be optimal or suboptimal. The computational performance of the optimal path generation depends on the avoidance of unnecessary explorations. Accordingly, heuristic functions have been widely adopted to guide the exploration efficiently by providing estimated costs to the goal configurations. The exploration is efficient when the heuristic functions estimate the optimal cost closely which remains challenging because it requires a comprehensive understanding of the environment. However, this challenge presents the scope to improve the computational efficiency over the existing methods. Herein, we propose Reinforcement Learning Heuristic A* (RLHA*), which adopts an artificial neural network as a learning heuristic function to closely estimate the optimal cost, while achieving a bounded suboptimal path. Instead of being trained by pre-computed paths, the learning heuristic function keeps improving by using self-generated paths. Numerous simulations were performed to demonstrate the consistent and robust performance of RLHA* by comparing it with existing methods. IEEE | - |
dc.language | English | - |
dc.publisher | Institute of Electrical and Electronics Engineers | - |
dc.title | Reinforcement Learning Heuristic A | - |
dc.type | Article | - |
dc.identifier.doi | 10.1109/TII.2022.3188359 | - |
dc.description.journalClass | 1 | - |
dc.identifier.bibliographicCitation | IEEE Transactions on Industrial Informatics, v.19, no.3, pp.2307 - 2316 | - |
dc.citation.title | IEEE Transactions on Industrial Informatics | - |
dc.citation.volume | 19 | - |
dc.citation.number | 3 | - |
dc.citation.startPage | 2307 | - |
dc.citation.endPage | 2316 | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.identifier.wosid | 000967277300001 | - |
dc.identifier.scopusid | 2-s2.0-85134225033 | - |
dc.relation.journalWebOfScienceCategory | Automation & Control Systems | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Interdisciplinary Applications | - |
dc.relation.journalWebOfScienceCategory | Engineering, Industrial | - |
dc.relation.journalResearchArea | Automation & Control Systems | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.type.docType | Article | - |
dc.subject.keywordAuthor | Costs | - |
dc.subject.keywordAuthor | Graph Search | - |
dc.subject.keywordAuthor | Heuristic algorithms | - |
dc.subject.keywordAuthor | Path planning | - |
dc.subject.keywordAuthor | Path Planning | - |
dc.subject.keywordAuthor | Planning | - |
dc.subject.keywordAuthor | Reinforcement learning | - |
dc.subject.keywordAuthor | Reinforcement Learning | - |
dc.subject.keywordAuthor | Robots | - |
dc.subject.keywordAuthor | Signal processing algorithms | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.