• Aucun résultat trouvé

Online Sparse Scene Coordinates Learning for Real-Time Camera Relocalization

N/A
N/A
Protected

Academic year: 2021

Partager "Online Sparse Scene Coordinates Learning for Real-Time Camera Relocalization"

Copied!
3
0
0

Texte intégral

(1)

HAL Id: hal-02048750

https://hal.archives-ouvertes.fr/hal-02048750

Submitted on 25 Feb 2019

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of

sci-entific research documents, whether they are

pub-lished or not. The documents may come from

teaching and research institutions in France or

abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est

destinée au dépôt et à la diffusion de documents

scientifiques de niveau recherche, publiés ou non,

émanant des établissements d’enseignement et de

recherche français ou étrangers, des laboratoires

publics ou privés.

Online Sparse Scene Coordinates Learning for

Real-Time Camera Relocalization

Nam-Duong Duong, Amine Kacete, Catherine Soladie, Pierre-Yves Richard,

Jérôme Royan

To cite this version:

Nam-Duong Duong, Amine Kacete, Catherine Soladie, Pierre-Yves Richard, Jérôme Royan. Online

Sparse Scene Coordinates Learning for Real-Time Camera Relocalization. 6th International

Confer-ence on 3D Vision (3DV), Sep 2018, Verona, Italy. �hal-02048750�

(2)

Online Sparse Scene Coordinates Learning for Real-Time Camera Relocalization

Nam-Duong Duong, Amine Kacete, Catherine Soladie, Pierre-Yves Richard, J´erˆome Royan

Institute of Research and Technology b-com

Nam-Duong.Duong@b-com.com

Abstract

Camera relocalization refers to the problematics of defining the camera pose in known scenes. It is needed in several applications such as augmented reality or robot navigation. However, current camera relocalization sys-tems require an off-line step, that uses the labeled images to construct a scene model. This step takes time, so that is dif-ficult to apply in the augmented reality applications. Thus, we develop an online learning for real-time camera relo-calization system. We present a hybrid method combining machine learning approach and geometric approach for ac-curate and real-time camera relocalization from each single RGB image independently. For machine learning part, we propose a sparse scene coordinates regression forest, which is able to be fast learned during images capturing. Our sys-tem presents how easily user can experience on scenes in desktop-scale.

1. Introduction

In computer vision, the main solution of camera pose es-timation for augmented reality systems is known as Simul-taneously Localization And Mapping (SLAM). SLAM is a method of camera localization. It faces tracking failures in the case of fast camera motion or sudden change of view-point, such as in a hand-held camera. Camera relocalization is then essential.

That is why nowadays, some research works focus on camera relocalization solution from RGB images. We can split the solutions into 3 groups: geometric approaches, machine learning approaches and hybrid approaches. Ge-ometric approaches first construct a scene model of a set of key-points attached to feature vectors by performing Struc-ture from Motion (SfM) on a set of images of the scene. Then camera relocalization is performed from each frame independently by matching feature vector of each frame to key-frames (frame-to-frame approach) or the scene model (frame-to-model approach). To accelerate feature matching, [3] uses a visual vocabulary for efficient 2D-to-3D match-ing. Even so, the matching time depends on the size of

the model. Machine learning approaches for camera re-localization have appeared to challenge these constraints. Deep learning based methods [2] can predict camera pose from each whole RGB image in real-time. However, lim-itation of these methods lies in their accuracy. Last year, [1] presented hybrid method combining deep learning ap-proach and geometric apap-proach for camera relocalization using only RGB images with high accuracy. However, they took more time to optimize camera pose from thousands of correspondences and consequently cannot address aug-mented reality applications.

In general, current works have two main limits. Firstly, learning phase requires more time to create a scene model by SfM or machine learning. Secondly, testing phase is still challenging to have a both real-time and accurate method. Our demonstration overcomes these two limits. We propose an online learning solution, that uses scene coordinates re-gression forest for accurate and real-time camera relocaliza-tion in augmented reality system.

2. Methodology

We present a hybrid method combining machine learn-ing approach and geometric approach. For the former, we propose a sparse feature learning using regression forest to predict 3D world coordinates corresponding to 2D key-point detection. That aims to define 2D-3D key-point corre-spondences. The geometric part then estimates camera pose from these correspondences.

In the training phase, our sparse feature regression forest can be learned immediately during images capturing. Our method belongs to supervised learning approaches. So that, we use a RGB-D camera and a marker (be attached in the scene) to fast estimate camera pose and to label 3D scene coordinates for each pixel. From each captured RGB-D frame with estimated camera pose from the marker, we ex-tract a set of SURF (Speed Up Robust Features) features. Each feature is labeled by projecting 2D image coordinates in the world coordinates system. Once our system reaches sufficient amount of feature, we create a new thread to train a tree of our regression forest in the same time of captur-ing image. Our traincaptur-ing step finishes when all trees are

(3)

Frame1 Frame2 ... ... ... ... FrameK

Pose1 Pose2 ... ... ... ... PoseN

f1 f2 ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... fM

...

Tree1

Thread N

TreeN

Read Frame

Camera Pose

Estimation

Feature Extraction

and Labeling

Training

Capturing Time

Thread 0

Thread 1

Figure 1. Our on-line training system. Our sparse feature regression forest is learned by multi-threads in the same time of capturing data.

learned. Capturing and training phase takes about 2 min-utes for a desktop-scale scene with the configuration of our sparse feature regression forest including: extraction up to 500 SURF features for each image; 4 trees in the forest; 16 depth of each tree. Figure 1shows the online learning of our system.

Sparse Feature Regression Forest SURF Feature Extraction

Camera Pose 3D World Coordinates

World Coordinates System

PnP RANSAC

Outliers Inliers

Figure 2. Our real-time camera relocalization in the augmented reality.

In the testing phase, we use only RGB camera. All fea-tures extracted from RGB image pass through our regres-sion forest to obtain a set of 2D-3D correspondences. We then use Perspective-n-Point (PnP) and RANdom SAmple Consensus (RANSAC) algorithms to calculate camera pose. Our method can perform in real-time at 50ms per frame. The testing time of our demonstration is shown in Figure2.

3. Conclusion

In this demonstration, we presented our hybrid method combining machine learning and geometric approach, where machine learning part is online learned during cap-turing data. Our method shows both accurate and real-time result for camera relocalization from only RGB images.

4. Videos

We made a demonstrative video1 to show the training and testing performance of our method for camera relocal-ization. The video 2 presents the ability to help user

ex-perience easily thanks to our solution. Finally, the video3

shows the robustness of our method to challenges of gener-alization, occlusion and illumination changes.

References

[1] E. Brachmann, A. Krull, S. Nowozin, J. Shotton, F. Michel, S. Gumhold, and C. Rother. Dsac - differentiable ransac for camera localization. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.1 [2] A. Kendall and R. Cipolla. Geometric loss functions for

cam-era pose regression with deep learning. Proceedings of the IEEE Conference on Computer Vision and Pattern Recogni-tion, 2017.1

[3] T. Sattler, B. Leibe, and L. Kobbelt. Efficient & effective prioritized matching for large-scale image-based localization. IEEE transactions on pattern analysis and machine intelli-gence, 39(9):1744–1756, 2017.1

1https://youtu.be/mBcpzHEkQLc 2https://youtu.be/fsyGV1Ca5Uo 3https://youtu.be/2oOL-BSemmQ

Figure

Figure 2. Our real-time camera relocalization in the augmented reality.

Références

Documents relatifs

Een geïntegreerde dienst voor thuiszorg is een instelling voor geneeskundige verzorging die bijdraagt tot het totale zorgaanbod voor patiënten, onder meer door de

The algorithm has been tested on various images sequences and for various applications (visual servoing, augmented reality,...) which demonstrates a real usability of this

Given those pros and cons, this thesis aims at improving the performance of camera relocalization in terms of both runtime and accuracy as well as handling challenges of

In this paper, a new distance feature ‘s’ is considered as a set of distances between local point features obtained from a fast image processing step and the contours of a more

The principle of our algorithm is to iteratively modify using a visual servoing control law the position of a virtual camera in order to register the desired features extracted from

La compatibilité du protocole 802.1x sur les systèmes d’exploitation des clients est assurée sur Windows depuis le service pack 3 de XP. Sur Mac, le protocole est

Initially, planes poses and camera motion are computed from a vertical and a horizontal blob by filtering their intersec- tion line and performing the SVD + 8-parameters

As a second stage of experiments, the principle point ( u 0 ; v 0 ) was xed to the centre of the image for the images provided by GETRIS, that is, at (360,288) and the