Отрывок: The matching points between the reference Kinect and the rest (obtained in the point cloud matching step) are the 3D points in the world reference and will be denoted by PWi for i ∈ {2, 3, 4}, the 2D projections of these points on the image plane are denoted by ui = (u, v) and are known from the acquisition step. In homogeneous coordinates, the mapping between points PW = (x, y, z) and their 2D projections u = (u, v) in the image plane is given by ( ) 0 0 1 W R t...
Полная запись метаданных
Поле DC | Значение | Язык |
---|---|---|
dc.contributor.author | Ruchay, A.N. | - |
dc.contributor.author | Dorofeev, K.A. | - |
dc.contributor.author | Kolpakov, V.I. | - |
dc.date.accessioned | 2018-11-13 15:56:18 | - |
dc.date.available | 2018-11-13 15:56:18 | - |
dc.date.issued | 2018 | - |
dc.identifier | Dspace\SGAU\20181112\72391 | ru |
dc.identifier.citation | Ruchay AN, Dorofeev KA, Kolpakov VI. Fusion of information from multiple Kinect sensors for 3D object reconstruction. Computer Optics 2018; 42(5): 898-903. DOI: 10.18287/2412-6179-2018-42-5-898-903. | ru |
dc.identifier.uri | https://dx.doi.org/10.18287/2412-6179-2018-42-5-898-903 | - |
dc.identifier.uri | http://repo.ssau.ru/handle/Zhurnal-Komputernaya-optika/Fusion-of-information-from-multiple-Kinect-sensors-for-3D-object-reconstruction-72391 | - |
dc.description.abstract | In this paper, we estimate the accuracy of 3D object reconstruction using multiple Kinect sensors. First, we discuss the calibration of multiple Kinect sensors, and provide an analysis of the accuracy and resolution of the depth data. Next, the precision of coordinate mapping between sensors data for registration of depth and color images is evaluated. We test a proposed system for 3D object reconstruction with four Kinect V2 sensors and present reconstruction accuracy results. Experiments and computer simulation are carried out using Matlab and Kinect V2. | ru |
dc.description.sponsorship | The Russian Science Foundation (project #17-76-20045) financially supported the work. | ru |
dc.language.iso | en_US | ru |
dc.publisher | Новая техника | ru |
dc.relation.ispartofseries | 42;5 | - |
dc.subject | multiple sensors | ru |
dc.subject | Kinect | ru |
dc.subject | 3D object reconstruction | ru |
dc.subject | fusion | ru |
dc.title | Fusion of information from multiple Kinect sensors for 3D object reconstruction | ru |
dc.type | Article | ru |
dc.textpart | The matching points between the reference Kinect and the rest (obtained in the point cloud matching step) are the 3D points in the world reference and will be denoted by PWi for i ∈ {2, 3, 4}, the 2D projections of these points on the image plane are denoted by ui = (u, v) and are known from the acquisition step. In homogeneous coordinates, the mapping between points PW = (x, y, z) and their 2D projections u = (u, v) in the image plane is given by ( ) 0 0 1 W R t... | - |
dc.classindex.scsti | 28.17.19 | - |
Располагается в коллекциях: | Журнал "Компьютерная оптика" |
Файлы этого ресурса:
Файл | Описание | Размер | Формат | |
---|---|---|---|---|
420521.pdf | 153.29 kB | Adobe PDF | Просмотреть/Открыть |
Показать базовое описание ресурса
Просмотр статистики
Поделиться:
Все ресурсы в архиве электронных ресурсов защищены авторским правом, все права сохранены.