以視覺為基礎之示範學習與協作機器人系統

dc.contributor許陳鑑zh_TW
dc.contributorHsu, Chen-Chienen_US
dc.contributor.author黃品叡zh_TW
dc.contributor.authorHwang, Pin-Juien_US
dc.date.accessioned2023-12-08T07:47:23Z
dc.date.available2026-07-01
dc.date.available2023-12-08T07:47:23Z
dc.date.issued2023
dc.description.abstractnonezh_TW
dc.description.abstractRobot arms have been widely used in many automated factories over the past decades. However, most conventional robots operate on the basis of pre-defined programs, limiting their responsiveness and adaptability to changes in the environment. When new tasks are deployed, weeks of reprogramming by robotic engineers/operators would be inevitable, with the detriment of downtime, high cost, and time consumption. To address this problem, this dissertation proposes a more intuitive way for robots to perform tasks through learning from human demonstration (LfD), based on two major components: understanding human behavior and reproducing the task by a robot. For the understanding of human behavior/intent, two approaches are presented. The first method uses a multi-action recognition carried out by an inflated 3D network (I3D) followed by a proposed statistically fragmented approach to enhance the action recognition results. The second method is a vision-based spatial-temporal action detection method to detect human actions focusing on meticulous hand movement in real time to establish an action base. For robot reproduction according to the descriptions in the action base, we integrate the sequence of actions in the action base and the key path derived by an object trajectory inductive method for motion planning to reproduce the task demonstrated by the human user. In addition to static industrial robot arms, collaborative robots (cobots) intended for human-robot interaction are playing an increasingly important role in intelligent manufacturing. Though promising for many industrial and home service applications, there are still issues to be solved for collaborative robots, including understanding human intention in a natural way, adaptability to execute tasks when the environment changes, and robot mobility to navigate around a working environment. Thus, this dissertation proposes a modularized solution for mobile collaborative robot systems, where the cobot equipped with a multi-camera localization scheme for self-localization can understand the human intention in a natural way via voice commands to execute the tasks instructed by the human operator in an unseen scenario when the environment changes. To validate the proposed approaches, comprehensive experiments are conducted and presented in this dissertation.en_US
dc.description.sponsorship電機工程學系zh_TW
dc.identifier80775002H-43035
dc.identifier.urihttps://etds.lib.ntnu.edu.tw/thesis/detail/232505a30dc4f720e7da87a16d0d9aa1/
dc.identifier.urihttp://rportal.lib.ntnu.edu.tw/handle/20.500.12235/120354
dc.language英文
dc.subjectnonezh_TW
dc.subjectrobotic systemsen_US
dc.subjectlearning from demonstrationen_US
dc.subjectaction recognitionen_US
dc.subjectobject detectionen_US
dc.subjecttrajectory planningen_US
dc.subjectcollaborative robotsen_US
dc.subjecthuman-robot interactionen_US
dc.subjectrobot localizationen_US
dc.title以視覺為基礎之示範學習與協作機器人系統zh_TW
dc.titleVision-Based Learning from Demonstration and Collaborative Robotic Systemsen_US
dc.typeetd

Files

Collections