Depth map refinement using multiple patch-based depth image completion via local stereo warping

2010 ◽  
Vol 49 (7) ◽  
pp. 077003
Author(s):  
Ilkwon Park
Entropy ◽  
2021 ◽  
Vol 23 (5) ◽  
pp. 546
Author(s):  
Zhenni Li ◽  
Haoyi Sun ◽  
Yuliang Gao ◽  
Jiao Wang

Depth maps obtained through sensors are often unsatisfactory because of their low-resolution and noise interference. In this paper, we propose a real-time depth map enhancement system based on a residual network which uses dual channels to process depth maps and intensity maps respectively and cancels the preprocessing process, and the algorithm proposed can achieve real-time processing speed at more than 30 fps. Furthermore, the FPGA design and implementation for depth sensing is also introduced. In this FPGA design, intensity image and depth image are captured by the dual-camera synchronous acquisition system as the input of neural network. Experiments on various depth map restoration shows our algorithms has better performance than existing LRMC, DE-CNN and DDTF algorithms on standard datasets and has a better depth map super-resolution, and our FPGA completed the test of the system to ensure that the data throughput of the USB 3.0 interface of the acquisition system is stable at 226 Mbps, and support dual-camera to work at full speed, that is, 54 fps@ (1280 × 960 + 328 × 248 × 3).


2019 ◽  
Vol 11 (10) ◽  
pp. 204 ◽  
Author(s):  
Dogan ◽  
Haddad ◽  
Ekmekcioglu ◽  
Kondoz

When it comes to evaluating perceptual quality of digital media for overall quality of experience assessment in immersive video applications, typically two main approaches stand out: Subjective and objective quality evaluation. On one hand, subjective quality evaluation offers the best representation of perceived video quality assessed by the real viewers. On the other hand, it consumes a significant amount of time and effort, due to the involvement of real users with lengthy and laborious assessment procedures. Thus, it is essential that an objective quality evaluation model is developed. The speed-up advantage offered by an objective quality evaluation model, which can predict the quality of rendered virtual views based on the depth maps used in the rendering process, allows for faster quality assessments for immersive video applications. This is particularly important given the lack of a suitable reference or ground truth for comparing the available depth maps, especially when live content services are offered in those applications. This paper presents a no-reference depth map quality evaluation model based on a proposed depth map edge confidence measurement technique to assist with accurately estimating the quality of rendered (virtual) views in immersive multi-view video content. The model is applied for depth image-based rendering in multi-view video format, providing comparable evaluation results to those existing in the literature, and often exceeding their performance.


2001 ◽  
Vol 41 (1) ◽  
pp. 429
Author(s):  
R.J.W. Bunt ◽  
W.D. Powell ◽  
T. Scholefield

Difficulties in defining the structural character of the reservoir horizons at the Tubridgi Gas Field arise from gas charging of thin, often laterally discontinuous, silts and sands within the overburden. The gas charging of these shallow, low permeability units results in a seismic representation of the field as a time low. Historically, conversion from time to a reliable depth image has been problematic due to the variable nature of the gas charging, the relatively sparse, multi-vintage 2D seismic coverage and the corresponding difficulties in defining an accurate velocity field.After the unsuccessful drilling program in 1997 when three out of the five wells were plugged and abandoned, a revised interpretation methodology was developed, incorporating all available geophysical data, but placing a much greater emphasis on geological information from each of the wells in the area.The new depth map and geological model were tested by the drilling of Tubridgi–16 to –18 in August 1999. These three wells intersected the Birdrong Sandstone within one metre of prognosis, with two wells located structurally up-dip of the previous 17 wells drilled on the field. This accuracy resulted in a 97% increase in remaining reserves and a much higher level of confidence in the structural configuration of the Tubridgi field.A core of the Lower Gearle Sandstone in the Tubridgi 18 well highlighted the potential of this zone which has subsequently been evaluated in greater detail and potentially represents an additional productive horizon for the field.


2019 ◽  
Vol 5 (9) ◽  
pp. 73 ◽  
Author(s):  
Wen-Nung Lie ◽  
Chia-Che Ho

In this paper, a multi-focus image stack captured by varying positions of the imaging plane is processed to synthesize an all-in-focus (AIF) image and estimate its corresponding depth map. Compared with traditional methods (e.g., pixel- and block-based techniques), our focus-based measures are calculated based on irregularly shaped regions that have been refined or split in an iterative manner, to adapt to different image contents. An initial all-focus image is first computed, which is then segmented to get a region map. Spatial-focal property for each region is then analyzed to determine whether a region should be iteratively split into sub-regions. After iterative splitting, the final region map is used to perform regionally best focusing, based on the Winner-take-all (WTA) strategy, i.e., choosing the best focused pixels from image stack. The depth image can be easily converted from the resulting label image, where the label for each pixel represents the image index from which the pixel with the best focus is chosen. Regions whose focus profiles are not confident in getting a winner of the best focus will resort to spatial propagation from neighboring confident regions. Our experiments show that the adaptive region-splitting algorithm outperforms other state-of-the-art methods or commercial software in synthesis quality (in terms of a well-known Q metric), depth maps (in terms of subjective quality), and processing speed (with a gain of 17.81~40.43%).


2020 ◽  
Vol 43 (1) ◽  
pp. 59-78 ◽  
Author(s):  
David Johnson ◽  
Daniela Damian ◽  
George Tzanetakis

We present research for automatic assessment of pianist hand posture that is intended to help beginning piano students improve their piano-playing technique during practice sessions. To automatically assess a student's hand posture, we propose a system that is able to recognize three categories of postures from a single depth map containing a pianist's hands during performance. This is achieved through a computer vision pipeline that uses machine learning on the depth maps for both hand segmentation and detection of hand posture. First, we segment the left and right hands from the scene captured in the depth map using per-pixel classification. To train the hand-segmentation models, we experiment with two feature descriptors, depth image features and depth context features, that describe the context of individual pixels' neighborhoods. After the hands have been segmented from the depth map, a posture-detection model classifies each hand as one of three possible posture categories: correct posture, low wrists, or flat hands. Two methods are tested for extracting descriptors from the segmented hands, histograms of oriented gradients and histograms of normal vectors. To account for variation in hand size and practice space, detection models are individually built for each student using support vector machines with the extracted descriptors. We validate this approach using a data set that was collected by recording four beginning piano students while performing standard practice exercises. The results presented in this article show the effectiveness of this approach, with depth context features and histograms of normal vectors performing the best.


2014 ◽  
Vol 1006-1007 ◽  
pp. 797-801
Author(s):  
Ying Sun ◽  
Guang Lin Gao

The depth map is a basic diagram of the intrinsic; each pixel value represents the scene graph the elevation position of the object point. In this paper, the analysis methods for target classification elevation map. Figure elevation are visible depth image, the depth of the image is the distance from each point in the scene to the image capture device values ​​of the image as an image pixel value.


Author(s):  
Takuya Matsuo ◽  
Norishige Fukushima ◽  
Yutaka Ishibashi

Sign in / Sign up

Export Citation Format

Share Document