Отрывок: The matching points between the reference Kinect and the rest (obtained in the point cloud matching step) are the 3D points in the world reference and will be denoted by PWi for i ∈ {2, 3, 4}, the 2D projections of these points on the image plane are denoted by ui = (u, v) and are known from the acquisition step. In homogeneous coordinates, the mapping between points PW = (x, y, z) and their 2D projections u = (u, v) in the image plane is given by ( ) 0 0 1 W R t...
Полная запись метаданных
Поле DC Значение Язык
dc.contributor.authorRuchay, A.N.-
dc.contributor.authorDorofeev, K.A.-
dc.contributor.authorKolpakov, V.I.-
dc.date.accessioned2018-11-13 15:56:18-
dc.date.available2018-11-13 15:56:18-
dc.date.issued2018-
dc.identifierDspace\SGAU\20181112\72391ru
dc.identifier.citationRuchay AN, Dorofeev KA, Kolpakov VI. Fusion of information from multiple Kinect sensors for 3D object reconstruction. Computer Optics 2018; 42(5): 898-903. DOI: 10.18287/2412-6179-2018-42-5-898-903.ru
dc.identifier.urihttps://dx.doi.org/10.18287/2412-6179-2018-42-5-898-903-
dc.identifier.urihttp://repo.ssau.ru/handle/Zhurnal-Komputernaya-optika/Fusion-of-information-from-multiple-Kinect-sensors-for-3D-object-reconstruction-72391-
dc.description.abstractIn this paper, we estimate the accuracy of 3D object reconstruction using multiple Kinect sensors. First, we discuss the calibration of multiple Kinect sensors, and provide an analysis of the accuracy and resolution of the depth data. Next, the precision of coordinate mapping between sensors data for registration of depth and color images is evaluated. We test a proposed system for 3D object reconstruction with four Kinect V2 sensors and present reconstruction accuracy results. Experiments and computer simulation are carried out using Matlab and Kinect V2.ru
dc.description.sponsorshipThe Russian Science Foundation (project #17-76-20045) financially supported the work.ru
dc.language.isoen_USru
dc.publisherНовая техникаru
dc.relation.ispartofseries42;5-
dc.subjectmultiple sensorsru
dc.subjectKinectru
dc.subject3D object reconstructionru
dc.subjectfusionru
dc.titleFusion of information from multiple Kinect sensors for 3D object reconstructionru
dc.typeArticleru
dc.textpartThe matching points between the reference Kinect and the rest (obtained in the point cloud matching step) are the 3D points in the world reference and will be denoted by PWi for i ∈ {2, 3, 4}, the 2D projections of these points on the image plane are denoted by ui = (u, v) and are known from the acquisition step. In homogeneous coordinates, the mapping between points PW = (x, y, z) and their 2D projections u = (u, v) in the image plane is given by ( ) 0 0 1 W R t...-
dc.classindex.scsti28.17.19-
Располагается в коллекциях: Журнал "Компьютерная оптика"

Файлы этого ресурса:
Файл Описание Размер Формат  
420521.pdf153.29 kBAdobe PDFПросмотреть/Открыть



Все ресурсы в архиве электронных ресурсов защищены авторским правом, все права сохранены.