Meet Inspiring Speakers and Experts at our 3000+ Global Conference Series Events with over 1000+ Conferences, 1000+ Symposiums
and 1000+ Workshops on Medical, Pharma, Engineering, Science, Technology and Business.

Explore and learn more about Conference Series : World's leading Event Organizer

Back

Rafael Diniz

University of Brasília, Brazil

Title: Depth sensors and volumetric video meet deep learning

Biography

Biography: Rafael Diniz

Abstract

The new visual immersive media formats provide a 3D visual representation of real objects and scenes. In this new visual format, objects can be captured, compressed, transmitted and visualized in real-time not anymore as a flat 2D image, but as 3D content. This new dimension of image capture is easily possible given the introduction of affordable yet powerful depth (or range) sensors. One of the most popular formats for immersive media is Point Cloud (PC), which is composed by points with 3 geometry coordinates plus color information, and sometimes, other information like reflectance and transparency. The proposed work considers the introduction of novel PC feature extractors [Figure 1] which are obtained by using both color and geometry information of a given scene or selected volumetric element of a larger object or scene. Initially, these feature descriptors were design for the purpose of objective quality assessment of volumetric images, but they can also be used to provide valuable information of the volumetric content. The use of PCs as input for machine learning algorithms is showing potential in providing greater accuracy for many types of machine learning algorithms for computer vision. The proposed descriptors are based on local-neighborhood luminance and surface geometry distances among the volumetric elements (voxels). The joint use of features based on color texture and geometry information is also proposed, and presents a better correlation to the Human Vision System (HVS) when such features are used as quality estimation.