Publication:
Comparing Fusion Methods for 3D Object Detection

dc.contributor.authorArıcan, Erkut
dc.contributor.authorAydin, Tarkan
dc.contributor.editorKahraman, C.
dc.contributor.editorCebi, S.
dc.contributor.editorCevik Onar, S.
dc.contributor.editorOztaysi, B.
dc.contributor.editorTolga, A.C.
dc.contributor.editorSari, I.U.
dc.contributor.institutionArıcan, Erkut, Faculty of Engineering and Natural Sciences, Bahçeşehir Üniversitesi, Istanbul, Turkey
dc.contributor.institutionAydin, Tarkan, Faculty of Engineering and Natural Sciences, Bahçeşehir Üniversitesi, Istanbul, Turkey
dc.date.accessioned2025-10-05T15:26:13Z
dc.date.issued2022
dc.description.abstractOne of the main problems in computer vision that leads digital technologies to transform our business and social life is object detection. The solutions to the problem have various application areas such as security systems, surveillance, shopping applications, and much more. Significant performance gain have been achieved by using the popular deep learning methods that are applied on 2D RGB images. With the availability of low-cost 3D sensors, methods that incorporates 3D data with 2D RGB data using existing deep network architectures became more popular to increase performance further. In this work, different data level and feature level fusion strategies have been analyzed to incorporate 3D depth and 2D RGB data with existing architectures to assess their effects on the performance. These methods were tested on the real RGB-D benchmark datasets available in the literature, and the accuracy results were compared to each other in their object types of group. © 2021 Elsevier B.V., All rights reserved.
dc.identifier.conferenceNameInternational Conference on Intelligent and Fuzzy Systems, INFUS 2021
dc.identifier.conferencePlaceIstanbul
dc.identifier.doi10.1007/978-3-030-85577-2_16
dc.identifier.endpage146
dc.identifier.issn23673389
dc.identifier.issn23673370
dc.identifier.scopus2-s2.0-85115261209
dc.identifier.startpage138
dc.identifier.urihttps://doi.org/10.1007/978-3-030-85577-2_16
dc.identifier.urihttps://hdl.handle.net/20.500.14719/9264
dc.identifier.volume308
dc.language.isoen
dc.publisherSpringer Science and Business Media Deutschland GmbH
dc.relation.sourceLecture Notes in Networks and Systems
dc.subject.authorkeywords2d
dc.subject.authorkeywords3d
dc.subject.authorkeywordsComputer Vision
dc.subject.authorkeywordsDeep Learning
dc.subject.authorkeywordsDepth
dc.subject.authorkeywordsFusion
dc.subject.authorkeywordsObject Detection
dc.subject.authorkeywordsRgb
dc.subject.authorkeywordsRgb-d
dc.titleComparing Fusion Methods for 3D Object Detection
dc.typeConference Paper
dcterms.referencesProceedings of the Dagm Symposium, (1995), Bay, Herbert, SURF: Speeded up robust features, Lecture Notes in Computer Science, 3951 LNCS, pp. 404-417, (2006), Drost, Bertram H., Model globally, match locally: Efficient and robust 3D object recognition, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 998-1005, (2010), undefined, Intel Intel Realsense, (2020), Janoch, Allison, A category-level 3-D object dataset: Putting the Kinect to work, Proceedings of the IEEE International Conference on Computer Vision, pp. 1168-1174, (2011), Johnson, Andrew Edie, Using spin images for efficient object recognition in cluttered 3D scenes, IEEE Transactions on Pattern Analysis and Machine Intelligence, 21, 5, pp. 433-449, (1999), Advances in Neural Information Processing Systems, (2012), Lai, Kevin, A large-scale hierarchical multi-view RGB-D object dataset, Proceedings - IEEE International Conference on Robotics and Automation, pp. 1817-1824, (2011), Leutenegger, Stefan, BRISK: Binary Robust invariant scalable keypoints, Proceedings of the IEEE International Conference on Computer Vision, pp. 2548-2555, (2011)
dspace.entity.typePublication
local.indexed.atScopus
person.identifier.scopus-author-id55807668300
person.identifier.scopus-author-id35106687700

Files