Show simple item record

dc.contributor.authorImtiaz, Muhammad Babar
dc.contributor.authorQiao, Yuansong
dc.contributor.authorLee, Brian
dc.date.accessioned2023-04-25T10:58:36Z
dc.date.available2023-04-25T10:58:36Z
dc.date.copyright2023
dc.date.issued2023-01-29
dc.identifier.citationImtiaz, M.B., Qiao, Y., Lee, B, (2023). Prehensile and non-prehensile robotic pick-and-place of objects in clutter using deep reinforcement learning. Sensors, 23, 1513. https://doi.org/10.3390/s23031513en_US
dc.identifier.urihttps://research.thea.ie/handle/20.500.12065/4490
dc.description.abstractIn this study, we develop a framework for an intelligent and self-supervised industrial pick-and-place operation for cluttered environments. Our target is to have the agent learn to perform prehensile and non-prehensile robotic manipulations to improve the efficiency and throughput of the pick-and-place task. To achieve this target, we specify the problem as a Markov decision process (MDP) and deploy a deep reinforcement learning (RL) temporal difference model-free algorithm known as the deep Q-network (DQN). We consider three actions in our MDP; one is ‘grasping’ from the prehensile manipulation category and the other two are ‘left-slide’ and ‘right-slide’ from the nonprehensile manipulation category. Our DQN is composed of three fully convolutional networks (FCN) based on the memory-efficient architecture of DenseNet-121 which are trained together without causing any bottleneck situations. Each FCN corresponds to each discrete action and outputs a pixel-wise map of affordances for the relevant action. Rewards are allocated after every forward pass and backpropagation is carried out for weight tuning in the corresponding FCN. In this manner, non-prehensile manipulations are learnt which can, in turn, lead to possible successful prehensile manipulations in the near future and vice versa, thus increasing the efficiency and throughput of the pick-and-place task. The Results section shows performance comparisons of our approach to a baseline deep learning approach and a ResNet architecture-based approach, along with very promising test results at varying clutter densities across a range of complex scenario test cases.en_US
dc.formatPDFen_US
dc.language.isoengen_US
dc.publisherMDPIen_US
dc.relation.ispartofSensorsen_US
dc.rightsAttribution- 3.0 United States*
dc.rights.urihttp://creativecommons.org/licenses/by/3.0/us/*
dc.subjectPrehensileen_US
dc.subjectNon-prehensileen_US
dc.subjectRobotic manipulationen_US
dc.subjectMarkov decision processen_US
dc.subjectDeep reinforcement learningen_US
dc.subjectDeep Q-networken_US
dc.subjectFully convolutional networken_US
dc.subjectDenseNet-121en_US
dc.titlePrehensile and non-prehensile robotic pick-and-place of objects in clutter using deep reinforcement learningen_US
dc.typeinfo:eu-repo/semantics/articleen_US
dc.contributor.affiliationTechnological University of the Shannon: Midlands Midwesten_US
dc.contributor.sponsorScience Foundation Ireland (SFI)en_US
dc.description.peerreviewyesen_US
dc.identifier.doi10.3390/s23031513en_US
dc.identifier.eissn1424-8220
dc.identifier.orcidhttps://orcid.org/0000-0003-4775-9033en_US
dc.identifier.orcidhttps://orcid.org/0000-0002-1543-1589en_US
dc.identifier.orcidhttps://orcid.org/0000-0002-8475-4074en_US
dc.rights.accessrightsinfo:eu-repo/semantics/openAccessen_US
dc.subject.departmentSoftware Research Institute: TUS MIdlandsen_US
dc.type.versioninfo:eu-repo/semantics/publishedVersionen_US
dc.relation.projectidSFI 16/RC/3918en_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record

Attribution- 3.0 United States
Except where otherwise noted, this item's license is described as Attribution- 3.0 United States