CN102447917A - Three-dimensional image matching method and equipment thereof - Google Patents
Three-dimensional image matching method and equipment thereof Download PDFInfo
- Publication number
- CN102447917A CN102447917A CN2010105025844A CN201010502584A CN102447917A CN 102447917 A CN102447917 A CN 102447917A CN 2010105025844 A CN2010105025844 A CN 2010105025844A CN 201010502584 A CN201010502584 A CN 201010502584A CN 102447917 A CN102447917 A CN 102447917A
- Authority
- CN
- China
- Prior art keywords
- pixel
- cost
- unit
- coupling
- target image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Landscapes
- Image Processing (AREA)
Abstract
提供一种立体图像匹配方法及其设备,其中,所述设备包括:代价计算单元,用于计算参考图像的像素与目标图像的像素之间的匹配代价,其中,代价计算单元分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;代价聚集单元,用于将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及视差确定单元,用于基于代价聚集单元对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。
A stereoscopic image matching method and device thereof are provided, wherein the device includes: a cost calculation unit configured to calculate a matching cost between pixels of a reference image and pixels of a target image, wherein the cost calculation unit calculates the matching costs of the reference image respectively A non-parametric statistical cost function and a color mean absolute difference function between pixels and pixels of the target image, and linearly combining the non-parametric statistical cost function and the color mean absolute difference function to use the combined result as a reference image The matching cost between the pixel of the pixel and the pixel of the target image; the cost aggregation unit is used to gather the matching cost between the pixel of the reference image calculated by the cost calculation unit and the pixel of the target image; and the parallax determination unit is used for The disparity between the pixels of the reference image and the target image is determined based on the result of the aggregation of the matching cost by the cost aggregation unit, so as to realize stereoscopic image matching.
Description
技术领域 technical field
本发明涉及用于图像处理的方法和设备,尤其涉及一种在参考图像与目标图像之间进行立体图像匹配的方法和设备。The present invention relates to a method and device for image processing, in particular to a method and device for stereoscopic image matching between a reference image and a target image.
背景技术 Background technique
立体图像匹配是计算机视觉领域的重要技术课题,在参考图像(例如,左图像)与目标图像(例如,右图像)之间实施立体图像匹配的过程中获取视差图,其中,视差(即,双目视差)是指在双目视觉中,同一空间点在双目分别看到的左图像和右图像中不同投影之间的位置矢量。实际中,由于人眼对于深度的感觉取决于水平方向的视差,所以视差也可仅仅指位置矢量的水平分量。Stereo image matching is an important technical topic in the field of computer vision. The disparity map is obtained during the process of implementing stereo image matching between a reference image (for example, left image) and a target image (for example, right image). Visual disparity) refers to the position vector between different projections of the same spatial point in the left image and right image seen by the binocular vision in binocular vision. In practice, since the perception of depth by the human eye depends on the parallax in the horizontal direction, the parallax may only refer to the horizontal component of the position vector.
如上所述,视差为产生深度感觉的依据,因此,在3D图像的技术发展中,通过视差估计实现立体图像匹配是计算场景深度、场景三维重建以及三维显示中的核心技术。在现有的立体图像匹配实施方案中,主要关注匹配精度和处理效率这两方面的需求。然而,尽管目前存在很多用于实现立体图像匹配的系统,但是这些系统均难以同时在匹配精度和处理效率这两方面实现较大的改进,往往是某一方面的改进导致牺牲了另一方面的性能。例如,许多精度较好的立体图像匹配方案具有以下共同点:1)采用较大的支持窗口来进行强健的代价聚集运算;2)在视差图的计算中,利用复杂的全局优化器来进行能量最优运算;3)主要依赖于匹配图像的分割结果来实现匹配精度的提高。上述特点虽然能够成功地抑制匹配中的模糊现象,但是却均需要复杂的计算方法为基础,使得计算处理需要耗费大量的时间,以致难以应用于对实时性有要求的情况。而另外一些立体图像匹配方案为了达到实时匹配的性能,采取了简化的措施,例如,图像金字塔、渐次匹配、强健测量和基于简单动态编程的优化器等,其结果是这些系统中的匹配精度不得不降低。As mentioned above, parallax is the basis for generating depth perception. Therefore, in the development of 3D image technology, stereoscopic image matching through parallax estimation is the core technology in calculating scene depth, 3D scene reconstruction and 3D display. In existing implementations of stereo image matching, the main focus is on the requirements of matching accuracy and processing efficiency. However, although there are many systems for stereoscopic image matching, it is difficult for these systems to achieve greater improvements in matching accuracy and processing efficiency at the same time. Often, the improvement in one aspect leads to the sacrifice of the other. performance. For example, many stereo image matching schemes with better accuracy have the following things in common: 1) use a large support window for robust cost aggregation operations; 2) use complex global optimizers to perform energy Optimal operation; 3) Mainly rely on the segmentation results of matching images to achieve the improvement of matching accuracy. Although the above features can successfully suppress the fuzzy phenomenon in matching, they all require complex calculation methods as the basis, which makes calculation and processing take a lot of time, making it difficult to apply to situations that require real-time performance. In order to achieve real-time matching performance, some other stereo image matching schemes adopt simplified measures, such as image pyramid, gradual matching, robust measurement and optimizer based on simple dynamic programming, etc., as a result, the matching accuracy in these systems cannot be achieved Not lowered.
因此,需要提供一种能够在匹配精度和处理速度两方面均有显著改进的立体图像匹配方案,从而能够在实现高精度匹配的同时达到实时处理的运算要求。Therefore, it is necessary to provide a stereoscopic image matching scheme that can significantly improve both matching accuracy and processing speed, so as to achieve high-precision matching and meet the computing requirements for real-time processing.
发明内容 Contents of the invention
本发明的目的在于提供一种能够实现高精度匹配的快速立体图像匹配方法和系统,此外,可通过对生成的视差图进行细化处理来进一步提高立体图像匹配的性能。The purpose of the present invention is to provide a fast stereoscopic image matching method and system capable of achieving high-precision matching. In addition, the performance of stereoscopic image matching can be further improved by performing thinning processing on the generated disparity map.
根据本发明的一方面,提供一种立体图像匹配设备,所述设备包括:代价计算单元,用于计算参考图像的像素与目标图像的像素之间的匹配代价,其中,代价计算单元分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;代价聚集单元,用于将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及视差确定单元,用于基于代价聚集单元对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。According to an aspect of the present invention, there is provided a stereoscopic image matching device, the device comprising: a cost calculation unit for calculating matching costs between pixels of a reference image and pixels of a target image, wherein the cost calculation unit calculates the reference The non-parametric statistical cost function and the color mean absolute difference function between the pixels of the image and the pixels of the target image, and the non-parametric statistical cost function and the color mean absolute difference function are linearly combined to use the result of the combination as a matching cost between the pixels of the reference image and the pixels of the target image; a cost aggregation unit for gathering the matching costs between the pixels of the reference image calculated by the cost calculation unit and the pixels of the target image; and a parallax determination unit, It is used to determine the disparity between the pixels of the reference image and the target image based on the result of the aggregation of the matching cost by the cost aggregation unit, so as to realize stereoscopic image matching.
所述代价聚集单元针可对参考图像中的每个像素确定该像素的动态支持域,在确定的每个动态支持域中,将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集。For each pixel in the reference image, the cost aggregation unit can determine the dynamic support region of the pixel, and in each determined dynamic support region, the difference between the pixel of the reference image calculated by the cost calculation unit and the pixel of the target image The matching costs among them are aggregated.
所述视差确定单元可构建附带深度平滑性约束的匹配代价的全局能量,并将使全局能量取得最小值的视差级别集合确定为最终的立体图像匹配视差图。The disparity determination unit may construct the global energy of the matching cost with a depth smoothness constraint, and determine the set of disparity levels that minimize the global energy as the final stereoscopic image matching disparity map.
所述代价聚集单元可包括:像素动态支持域构建单元,用于针对参考图像中的每个像素构建动态支持域;动态支持域代价聚集单元,用于针对构建的每个动态支持域,对其中的所有像素进行代价聚集。The cost aggregation unit may include: a pixel dynamic support domain construction unit for constructing a dynamic support domain for each pixel in the reference image; a dynamic support domain cost aggregation unit for constructing each dynamic support domain for which All pixels of are cost aggregated.
所述立体图像匹配设备,还可包括:视差图细化单元,用于对由视差确定单元输出的视差图进行细化处理。The stereo image matching device may further include: a disparity map refinement unit, configured to refine the disparity map output by the disparity determination unit.
所述视差图细化单元可包括以下项中的至少一个:遮挡点处理单元,用于检测视差图中的遮挡点,利用遮挡点周围的图像来估计遮挡点的视差;非连续边界修正单元,用于在视差图中检测视差值出现跳变的非连续边界,对于非连续边界上的像素,利用在所述非连续边界两侧与所述像素间隔一定距离的像素来重新确定非连续边界上的像素的视差;子像素级差值单元,用于对视差图进行子像素级别的差值运算。The disparity map refinement unit may include at least one of the following items: an occlusion point processing unit, configured to detect an occlusion point in the disparity map, and use images around the occlusion point to estimate the disparity of the occlusion point; a discontinuous boundary correction unit, It is used to detect a discontinuous boundary in which the disparity value jumps in the disparity map, and for pixels on the discontinuous boundary, use pixels on both sides of the discontinuous boundary that are at a certain distance from the pixel to re-determine the discontinuous boundary The disparity of the pixels above; the sub-pixel level difference unit is used to perform sub-pixel level difference operation on the disparity map.
在所述线性组合中,所述非参数统计代价函数与彩色平均绝对差值函数之间的比例可基于每个像素的局部纹理、局部梯度和/或匹配置信信息来自适应性地设置。In the linear combination, the ratio between the non-parametric statistical cost function and the color mean absolute difference function may be adaptively set based on each pixel's local texture, local gradient and/or matching confidence information.
根据本发明的另一方面,提供一种立体图像匹配方法,所述方法包括:计算参考图像的像素与目标图像的像素之间的匹配代价,其中,分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;将计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及基于对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。According to another aspect of the present invention, a stereoscopic image matching method is provided, the method comprising: calculating the matching cost between the pixels of the reference image and the pixels of the target image, wherein the pixels of the reference image and the pixels of the target image are respectively calculated Between the non-parametric statistical cost function and the color mean absolute difference function, and the non-parametric statistical cost function and the color mean absolute difference function are linearly combined to use the result of the combination as the pixel of the reference image and the pixel of the target image The matching cost between the pixels; the calculated matching costs between the pixels of the reference image and the pixels of the target image are aggregated; and the disparity of the pixels of the reference image to the target image is determined based on the result of the aggregation of the matching costs, to Realize stereo image matching.
进行聚集的步骤可包括:针对参考图像中的每个像素确定该像素的动态支持域,在确定的每个动态支持域中,将计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集。The step of aggregating may include: for each pixel in the reference image, determine the dynamic support region of the pixel, and in each determined dynamic support region, the calculated matching between the pixel of the reference image and the pixel of the target image The cost is aggregated.
确定视差的步骤可包括:构建附带深度平滑性约束的匹配代价的全局能量,并将使全局能量取得最小值的视差级别集合确定为最终的立体图像匹配视差图。The step of determining the disparity may include: constructing a global energy of a matching cost with a depth smoothness constraint, and determining a set of disparity levels that minimize the global energy as a final stereoscopic image matching disparity map.
所述立体图像匹配方法还可包括:对最终的立体图像匹配视差图进行细化处理。The stereo image matching method may further include: performing thinning processing on the final stereo image matching disparity map.
对最终的立体图像匹配视差图进行细化处理的步骤可包括以下步骤中的至少一个:检测视差图中的遮挡点,利用遮挡点周围的图像来估计遮挡点的视差;在视差图中检测视差值出现跳变的非连续边界,对于非连续边界上的像素,利用在所述非连续边界两侧与所述像素间隔一定距离的像素来重新确定非连续边界上的像素的视差;对视差图进行子像素级别的差值运算。The step of refining the final stereo image matching disparity map may include at least one of the following steps: detecting an occlusion point in the disparity map, using images around the occlusion point to estimate the disparity of the occlusion point; detecting a disparity in the disparity map The discontinuous boundary where the difference value jumps, for the pixels on the discontinuous boundary, use the pixels at a certain distance from the pixel on both sides of the discontinuous boundary to re-determine the disparity of the pixels on the discontinuous boundary; for the disparity The image performs the difference operation at the sub-pixel level.
根据本发明的另一方面,提供一种用于3D内容的生成装置,包括:3D数据输入单元、3D数据分析单元、3D内容产生单元,所述生成装置还包括:根据本发明的立体图像匹配设备。According to another aspect of the present invention, there is provided a generating device for 3D content, including: a 3D data input unit, a 3D data analysis unit, and a 3D content generating unit, and the generating device also includes: the stereoscopic image matching according to the present invention equipment.
根据本发明的另一方面,提供一种用于3D内容的显示装置,包括:3D数据输入单元、3D数据分析单元、3D内容产生单元和3D内容显示单元,所述显示装置还包括:根据本发明的立体图像匹配设备。According to another aspect of the present invention, there is provided a display device for 3D content, including: a 3D data input unit, a 3D data analysis unit, a 3D content generation unit, and a 3D content display unit, and the display device further includes: according to the present invention Invented stereo image matching device.
根据本发明的另一方面,提供一种立体图像匹配设备,所述设备包括:代价计算单元,用于计算参考图像的像素与目标图像的像素之间的匹配代价;代价聚集单元,用于针对参考图像中的每个像素确定该像素的动态支持域,在确定的每个动态支持域中,将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及视差确定单元,用于基于代价聚集单元对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。According to another aspect of the present invention, there is provided a stereoscopic image matching device, which includes: a cost calculation unit for calculating the matching cost between pixels of a reference image and pixels of a target image; a cost aggregation unit for Each pixel in the reference image determines the dynamic support region of the pixel, and in each determined dynamic support region, the matching costs between the pixels of the reference image calculated by the cost calculation unit and the pixels of the target image are gathered; and The disparity determination unit is configured to determine the disparity between the pixels of the reference image and the target image based on the result of the aggregation of the matching cost by the cost aggregation unit, so as to achieve stereoscopic image matching.
根据本发明的另一方面,提供一种立体图像匹配方法,所述方法包括:计算参考图像的像素与目标图像的像素之间的匹配代价;针对参考图像中的每个像素确定该像素的动态支持域,在确定的每个动态支持域中,将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及基于对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。According to another aspect of the present invention, there is provided a stereoscopic image matching method, the method comprising: calculating the matching cost between a pixel of a reference image and a pixel of a target image; a support domain, in each determined dynamic support domain, the matching costs between the pixels of the reference image calculated by the cost calculation unit and the pixels of the target image are aggregated; and the reference image is determined based on the result of the aggregation of the matching costs The pixel-to-target image disparity for stereo image matching.
根据本发明,能够基于特定的规则计算匹配代价函数,并进一步通过针对每个像素的动态支持域进行代价聚集,在代价聚集的基础上可考虑全局的深度平滑性而产生参考图像的视差图,从而在实现立体图像匹配的过程中同时提高了匹配精度和处理速度。此外,通过对产生的视差图进行细化修正处理,能够进一步提高图像匹配的性能。According to the present invention, the matching cost function can be calculated based on specific rules, and the cost aggregation is further carried out for the dynamic support domain of each pixel. On the basis of the cost aggregation, the global depth smoothness can be considered to generate the disparity map of the reference image. Therefore, the matching accuracy and the processing speed are simultaneously improved in the process of realizing stereoscopic image matching. In addition, the performance of image matching can be further improved by performing thinning and correction processing on the generated disparity map.
附图说明 Description of drawings
通过下面结合附图进行的对实施例的描述,本发明的上述和/或其它目的和优点将会变得更加清楚,其中:The above and/or other objects and advantages of the present invention will become more clear through the following description of the embodiments in conjunction with the accompanying drawings, wherein:
图1是示出根据本发明示例性实施例的立体图像匹配设备的框图;FIG. 1 is a block diagram illustrating a stereoscopic image matching device according to an exemplary embodiment of the present invention;
图2是示出根据本发明示例性实施例的立体图像匹配方法的流程图;FIG. 2 is a flowchart illustrating a stereo image matching method according to an exemplary embodiment of the present invention;
图3是示出根据本发明示例性实施例的代价聚集单元的详细结构的框图;3 is a block diagram showing a detailed structure of a cost aggregation unit according to an exemplary embodiment of the present invention;
图4示出根据本发明示例性实施例来构建像素动态支持域的示例;FIG. 4 shows an example of constructing a pixel dynamic support domain according to an exemplary embodiment of the present invention;
图5示出根据本发明示例性实施例的基于动态支持域进行代价聚集的示例;FIG. 5 shows an example of cost aggregation based on dynamic support domains according to an exemplary embodiment of the present invention;
图6是示出根据本发明示例性实施例的视差图细化单元的详细结构的框图;6 is a block diagram illustrating a detailed structure of a disparity map refinement unit according to an exemplary embodiment of the present invention;
图7示出根据本发明示例性实施例的遮挡点处理的示例;Fig. 7 shows an example of occlusion point processing according to an exemplary embodiment of the present invention;
图8示出根据本发明示例性实施例的非连续边界修正的示例;以及FIG. 8 shows an example of discontinuous boundary correction according to an exemplary embodiment of the present invention; and
图9示出根据本发明示例性实施例的立体图像匹配系统与现有技术相比在性能方面的改进。FIG. 9 shows the improvement in performance of the stereo image matching system according to the exemplary embodiment of the present invention compared with the prior art.
具体实施方式 Detailed ways
现将详细描述本发明的实施例,所述实施例的示例在附图中示出,其中,相同的标号始终指的是相同的部件。以下将通过参照附图来说明所述实施例,以便解释本发明。Reference will now be made in detail to embodiments of the invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like parts throughout. The embodiments are described below in order to explain the present invention by referring to the figures.
图1是示出根据本发明示例性实施例的立体图像匹配设备的框图。如图1所示,根据本发明示例性实施例的立体图像匹配设备包括:代价计算单元10,用于计算参考图像的像素与目标图像的像素之间的匹配代价,其中,代价计算单元10分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;代价聚集单元20,用于将由代价计算单元10计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及视差确定单元30,用于基于代价聚集单元20对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。作为优选但非必要的实施方式,还可额外包括视差图细化单元40,用于对由视差确定单元30输出的视差构成的视差图进行细化处理,从而进一步提高立体图像匹配的性能。FIG. 1 is a block diagram illustrating a stereoscopic image matching device according to an exemplary embodiment of the present invention. As shown in FIG. 1 , the stereoscopic image matching device according to an exemplary embodiment of the present invention includes: a
以下将参照图2来描述利用图1所示的立体图像匹配设备来实现根据本发明的立体图像匹配方法的示例。An example of implementing the stereo image matching method according to the present invention by using the stereo image matching device shown in FIG. 1 will be described below with reference to FIG. 2 .
图2是示出根据本发明示例性实施例的立体图像匹配方法的流程图。参照图2,在步骤S100,由代价计算单元10计算参考图像的像素与目标图像的像素之间的匹配代价,其中,代价计算单元10分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;在步骤S200,由代价聚集单元20将由代价计算单元10计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;在步骤S300,由视差确定单元30基于代价聚集单元20对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。作为优选但非必要的实施方式,还可额外包括步骤S400,其中,由视差图细化单元40对由视差确定单元30输出的视差构成的视差图进行细化处理,从而进一步提高立体图像匹配的性能。FIG. 2 is a flowchart illustrating a stereoscopic image matching method according to an exemplary embodiment of the present invention. Referring to Fig. 2, in step S100, the matching cost between the pixels of the reference image and the pixels of the target image is calculated by the
以下,首先描述在步骤S100,由代价计算单元10计算参考图像的像素与目标图像的像素之间的匹配代价的处理。In the following, first, the process of calculating the matching cost between the pixels of the reference image and the pixels of the target image by the
代价计算单元10接收参考图像和目标图像,其中,作为示例,参考图像为左图像,而目标图像为右图像。然后,对于左图像中的每个像素p(x,y),计算其与右图像中的像素pd(x-d,y)的匹配代价,其中,x、y分别指示像素p的横坐标和纵坐标,d表示给定的视差级别,d的取值范围可根据实际需要以整像素为单位而设置。The
在根据本发明的当前示例性实施例中,代价计算单元10分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价。In the current exemplary embodiment according to the present invention, the
具体说来,代价计算单元10首先计算像素p(x,y)与像素pd(x-d,y)之间的非参数统计代价函数,在这种情况下,选择像素p(x,y)周围的方形窗口Np以及对应的像素pd(x-d,y)周围的方形窗口Npd,并分别将所选择的方形窗口Np和Npd中的像素按照一定的顺序表示为比特串的形式,在所述比特串中,通过比特Bitp(p1)表示像素p(x,y)的强度与其方形窗口Np中的任一像素p1的强度之间的相对关系(例如,大于、小于或近似),举例说来,当像素p(x,y)的强度大于像素p1的强度时,将Bitp(p1)值设置为1,否则将其设置为0,其中,可根据实际需要来设定各种具体的比较关系;类似地,通过比特Bitpd(pd1)表示像素pd(x-d,y)的强度与其方形窗口Npd中的任一像素pd1的强度之间的相对关系(例如,大于、小于或近似),举例说来,当像素pd(x-d,y)的强度大于像素pd1的强度时,将Bitpd(pd1)值设置为1,否则将其设置为0,其中,可根据实际需要来设定各种具体的比较关系。Specifically, the cost calculation unit 10 first calculates the non-parametric statistical cost function between the pixel p(x, y) and the pixel pd(xd, y), in this case, selects the pixels around the pixel p(x, y) square window N p and the square window N pd around the corresponding pixel pd(xd, y), and respectively represent the pixels in the selected square window N p and N pd as a bit string in a certain order, in the In the above bit string, bit Bit p (p 1 ) represents the relative relationship between the intensity of pixel p(x, y) and the intensity of any pixel p 1 in its square window N p (for example, greater than, less than or approximately ), for example, when the intensity of the pixel p(x, y) is greater than the intensity of the pixel p 1 , the value of Bit p (p 1 ) is set to 1, otherwise it is set to 0, which can be determined according to actual needs Set various specific comparison relationships; similarly, the relative relationship between the intensity of pixel pd (xd, y) and the intensity of any pixel pd1 in its square window N pd is represented by bit Bit pd (pd 1 ) (for example , greater than, less than or approximately), for example, when the intensity of pixel pd(xd, y) is greater than the intensity of pixel pd 1 , the value of Bit pd (pd 1 ) is set to 1, otherwise it is set to 0, where , various specific comparison relationships can be set according to actual needs.
基于上述设定,代价计算单元10按照下面的等式1来计算像素p(x,y)与像素pd(x-d,y)之间的非参数统计代价函数Ccensus(p,d):Based on the above settings, the
其中,W(p1,p)和W(pd1,pd)分别为方形窗口Np与Npd之间的汉明距离的变化权重,W(p1,p)值可根据像素p1与像素p之间在空间和色彩上的关系来计算得出,具体说来,可根据以下规则来得出W(p1,p)值:当像素p1与像素p之间的距离越近,W(p1,p)值越大,同时,当像素p1与像素p的颜色越相近,W(p1,p)值也越大;类似地,W(pd1,pd)值可根据像素pd1与像素pd之间在空间和色彩上的关系来计算得出,具体说来,可根据以下规则来得出W(pd1,pd)值:当像素pd1与像素pd之间的距离越近,W(pd1,pd)值越大,同时,当像素pd1与像素pd的颜色越相近,W(pd1,pd)值也越大。Among them, W(p 1 , p) and W(pd 1 , pd) are the change weights of the Hamming distance between the square window N p and N pd respectively, and the value of W(p 1 , p) can be determined according to the pixel p 1 and The relationship between pixels p in space and color can be calculated. Specifically, the value of W(p 1 , p) can be obtained according to the following rules: when the distance between pixel p 1 and pixel p is closer, W The larger the value of (p 1 , p) is, at the same time, when the color of pixel p 1 is closer to that of pixel p, the value of W(p 1 , p) is also larger; similarly, the value of W(pd 1 , pd) can be calculated according to pixel The relationship between pd 1 and pixel pd in terms of space and color can be calculated. Specifically, the value of W(pd 1 , pd) can be obtained according to the following rules: when the distance between pixel pd 1 and pixel pd is greater Recently, the larger the value of W(pd 1 , pd), at the same time, the closer the color of the pixel pd 1 to the pixel pd, the larger the value of W(pd 1 , pd).
在代价计算单元10如上计算出像素p(x,y)与像素pd(x-d,y)之间的非参数统计代价函数Ccensus(p,d)之后,所述代价计算单元10来计算参考图像的像素与目标图像的像素之间的彩色平均绝对差值函数。作为示例,可通过等式2来计算像素p(x,y)与像素pd(x-d,y)在R(红)、G(绿)、B(蓝)三色通道上的彩色平均绝对差值函数CAD(p,d):After the
其中,I(p)和I(pd)分别表示像素p(x,y)与像素pd(x-d,y)在对应的R、G、B通道下的强度值,λAD为截断阈值。Among them, I(p) and I(pd) represent the intensity values of pixel p(x, y) and pixel pd(xd, y) under the corresponding R, G, B channels respectively, and λ AD is the truncation threshold.
然后,代价计算单元10将所述非参数统计代价函数Ccensus(p,d)与彩色平均绝对差值函数CAD(p,d)进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价C(p,d)。Then, the
作为示例,代价计算单元10按照等式3来计算参考图像的像素与目标图像的像素之间的匹配代价C(p,d):As an example, the
C(p,d)=z(p)·Ccensus(p,d)+(1-z(p))·CAD(p,d)等式3C(p,d)=z(p)·C census (p,d)+(1-z(p))·C AD (p,d)
其中,z(p)为用于控制非参数统计代价函数Ccensus(p,d)与彩色平均绝对差值函数CAD(p,d)的比例关系的权重,其值在0与1之间。例如,可考虑每个像素p(x,y)的局部纹理、局部梯度和/或匹配置信(matching confidence)信息来自适应性地设置z(p)。Among them, z(p) is the weight used to control the proportional relationship between the non-parametric statistical cost function C census (p, d) and the color average absolute difference function C AD (p, d), and its value is between 0 and 1 . For example, z(p) may be adaptively set in consideration of local texture, local gradient and/or matching confidence information of each pixel p(x,y).
在如上所述得到参考图像中的像素匹配代价函数之后,可在步骤S200,由代价聚集单元20将由代价计算单元10计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集,从而产生代价聚集结果Caggr(p,d)。After obtaining the pixel matching cost function in the reference image as described above, in step S200, the
这里,作为一种可选方式,可采用根据本发明示例性实施例的优选聚集方案,即,基于每个像素的动态支持域来进行代价聚集。以下,将参照图3来详细描述根据本发明示例性实施例的代价聚集单元的详细结构。Here, as an optional manner, the preferred aggregation scheme according to the exemplary embodiment of the present invention may be adopted, that is, the cost aggregation is performed based on the dynamic support region of each pixel. Hereinafter, the detailed structure of the cost aggregation unit according to the exemplary embodiment of the present invention will be described in detail with reference to FIG. 3 .
图3是示出根据本发明示例性实施例的代价聚集单元的详细结构的框图。参照图3,作为一种示例性结构,代价聚集单元20可包括:像素动态支持域构建单元201,用于针对参考图像中的每个像素构建动态支持域;动态支持域代价聚集单元202,用于针对构建的每个动态支持域,对其中的所有像素进行代价聚集。FIG. 3 is a block diagram illustrating a detailed structure of a cost aggregation unit according to an exemplary embodiment of the present invention. Referring to FIG. 3 , as an exemplary structure, the
如本领域技术人员所知,构建的动态支持域作为每个像素进行代价聚合的基础,其形状和大小对匹配结果的精确性有很大的影响。例如,对于纹理较少的区域需要构建较大的支持域,但是如果所述支持域跨过图像中某对象的边界,则在对象边界处的匹配性能较差。另一方面,较小的支持域对于图像噪声和拐点也比较敏感。因此,综合上述因素,像素动态支持域构建单元201可遵循以下两条准则来构建每个像素的动态支持域:(1)动态支持域中的像素具有与该支持域的中心像素相类似的彩色信息;(2)动态支持域的边界应遵循参考图像中的已有边缘,其中,可利用像素局部的图像梯度信息来检测参考图像中的已有边缘。图4示出根据本发明示例性实施例来构建像素动态支持域的示例。如图4所示,针对像素P和像素Q的动态支持域分别基于上述两条准则被构建。As known to those skilled in the art, the constructed dynamic support domain serves as the basis for cost aggregation for each pixel, and its shape and size have a great influence on the accuracy of the matching result. For example, a larger support domain needs to be constructed for an area with less texture, but if the support domain crosses the boundary of an object in the image, the matching performance at the object boundary will be poor. On the other hand, smaller support domains are also more sensitive to image noise and inflection points. Therefore, considering the above factors, the pixel dynamic support
在像素动态支持域构建单元201构建了各个像素的动态支持域之后,动态支持域代价聚集单元202针对构建的每个动态支持域,对其中的所有像素进行代价聚集。After the pixel dynamic support
作为示例,动态支持域代价聚集单元202可按照图5所示的方式对像素P的动态支持域中的所有像素进行代价聚集。如图5所示,动态支持域代价聚集单元202可按照两个通路对像素P的动态支持域中的所有像素进行代价聚集:首先,沿水平方向分别对每一行的所有像素进行代价聚集,得到并存储水平方向的各个聚集结果,然后,将存储的水平方向的聚焦结果沿垂直方向进行聚合。As an example, the dynamic support region
本领域技术人员应理解,上述通过图3所示的代价聚集单元20基于像素的动态支持域进行代价匹配处理的示例并不依赖于特定的像素匹配代价算法,即,进行聚集的像素匹配代价函数并不受限于由图1的代价计算单元10产生的具体代价函数,而是可以应用于各种像素的匹配代价函数。虽然上述动态域为二维空间域,还可将空间和视差域的三维支持域应用于本发明中。此外,本领域技术人员还应理解,图3所示的像素动态支持域构建单元201和动态支持域代价聚集单元202仅仅是一种示例,这种示例并不构成对代价聚集单元20的结构限制。也就是说,代价聚集单元20完全可以作为单个的实体单元来实现像素动态支持域构建单元201和动态支持域代价聚集单元202整体执行的操作,或者,代价聚集单元20还可被进一步细分为更多的单元来执行所述操作。Those skilled in the art should understand that the above-mentioned example of performing cost matching processing based on the dynamic support domain of pixels by the
参照回图2,在如上所述得到代价函数的聚集结果之后,可在步骤S300,由视差确定单元30基于代价聚集单元20对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差。具体说来,视差确定单元30可利用能量优化器来计算视差图,即,基于特定的能量运算来确定使得匹配代价最小的视差级别集合其中,p表示参考图像中的任一像素,dp为该像素p的视差值。如果直接对聚集的代价结果应用局部WTA(Winner Take All,胜者通吃)算法来选择分别对每个像素p而言具有最小匹配代价的视差,则产生的视差图中会存在很多噪声深度值。因此,作为一种优选的方式,可采用包括对深度平滑性进行附加约束的全局算法,将确定视差的处理构建为确定使全局能量得到最小值的视差级别集合作为确定的视差图。Referring back to FIG. 2, after the aggregation result of the cost function is obtained as described above, in step S300, the parallax of the pixels of the reference image to the target image can be determined by the
作为示例,视差确定单元30可按照等式4来构建附带深度平滑性约束的匹配代价的全局能量,而使得该全局能量取得最小值的视差级别集合则被确定为最终的立体图像匹配视差图:As an example, the
其中,Caggr(p,dp)为匹配代价的聚集结果,N为像素p的邻域,可根据实际需要来确定,q为邻域N中的像素,dq为像素q的视差值,T(A)为真值函数,即,当A等式成立时,T(A)值为1,否则,T(A)值为0,P1和P2为惩罚参数,可根据实际需要来设置。此外,作为优选方式,可采用4向扫描线优化方式来有效地执行上述确定的处理流程。Among them, C aggr (p, d p ) is the aggregation result of the matching cost, N is the neighborhood of pixel p, which can be determined according to actual needs, q is the pixel in the neighborhood N, and d q is the disparity value of pixel q , T(A) is a truth function, that is, when the A equation is established, the value of T(A) is 1, otherwise, the value of T(A) is 0, and P 1 and P 2 are penalty parameters, which can be adjusted according to actual needs to set. In addition, as a preferred method, the 4-direction scan line optimization method can be used to effectively perform the above determination processing flow.
参照回图2,在由视差确定单元30确定了参考图像的像素到目标图像的视差图之后,作为优选但非必要的实施方式,还可额外包括步骤S400,其中,由视差图细化单元40对由视差确定单元30输出的视差图进行细化处理,从而进一步提高立体图像匹配的性能。Referring back to FIG. 2 , after the disparity map from the pixels of the reference image to the target image is determined by the
以下,将参照图6来根据本发明示例性实施例的视差图细化单元40的详细结构。Hereinafter, the detailed structure of the disparity
图6是示出根据本发明示例性实施例的视差图细化单元的详细结果的框图。参照图6,作为一种示例性结构,视差图细化单元40可包括:遮挡点处理单元401,用于检测视差图中的遮挡点,利用遮挡点周围的图像来估计遮挡点的视差;非连续边界修正单元402,用于在由遮挡点处理单元401处理过的视差图中检测视差值出现跳变的非连续边界,对于非连续边界上的像素,利用在所述非连续边界两侧与所述像素间隔一定距离的像素来重新确定非连续边界上的像素的视差;子像素级差值单元403,用于对由非连续边界修正单元402修正过的视差图进行子像素级别的差值运算。FIG. 6 is a block diagram illustrating detailed results of a disparity map refinement unit according to an exemplary embodiment of the present invention. Referring to FIG. 6, as an exemplary structure, the disparity
具体说来,遮挡点处理单元401通过检查参考图像与目标图像的左右一致性来检测视差图中的遮挡点(即,无法通过目标图像中的相应像素确定其视差值的点)。然后,如图7中的(a)所示,首先利用该遮挡点的动态支持域中的其它像素来估计遮挡点的视差值,例如,可将所述动态支持域的所有像素的视差值中,最为普遍的视差值作为遮挡点的视差值。通过这种方式,可确定某些或全部遮挡点的视差值,对于仍旧无法确定视差值的遮挡点,还可如图7中的(b)所示,从遮挡点出发沿多个方向(例如,16个方向)进行扫描,将沿这些方向扫描到的像素中,色彩与遮挡点最为接近的像素的视差值确定为所述遮挡点的视差值。Specifically, the occlusion
然后,非连续边界修正单元402在由遮挡点处理单元401处理过的视差图中检测视差值出现跳变(例如,视差值的差>2)的非连续边界,然后,对于非连续边界上的像素,非连续边界修正单元402利用在所述非连续边界两侧与所述像素间隔一定距离(例如,4个像素)的两个像素来重新确定非连续边界上的像素的视差。例如,可将所述两个像素的彩色值分别与非连续边界上的所述像素进行比较,将与非连续边界上的所述像素的颜色较为相近的像素的视差值选为所述边界上的像素的视差值。作为优选方式,可利用像素的动态支持域的彩色信息来代表像素的彩色值,从而减少相应的运算消耗。上述非连续边界修正处理的结果如图8所示,在图8中,左图示出修正前检测出的非连续边界,右图示出修正后的结果。Then, the discontinuous
然后,子像素级差值单元403可对由非连续边界修正单元402修正过的视差图进行子像素级别的差值运算。由于之前的处理均基于整数像素级别,为了进一步对产生的视差图进行细化,例如,可按照等式5进行子像素级别的差值运算:Then, the sub-pixel
其中,d*为子像素级别的视差值,E表示通过等式4计算的能量。where d * is the disparity value at the sub-pixel level, and E represents the energy calculated by
本领域技术人员应理解,上述通过图6所示的遮挡点处理单元401、非连续边界修正单元402和子像素级差值单元403仅仅是一种示例,这种示例并不构成对视差图细化单元40的结构限制。也就是说,视差图细化单元40完全可以作为单个的实体单元来实现遮挡点处理单元401、非连续边界修正单元402和子像素级差值单元403整体执行的操作,或者,视差图细化单元40还可被进一步细分为更多的单元来执行所述操作。另外,上述遮挡点处理单元401、非连续边界修正单元402和子像素级差值单元403分别执行了三种独立的视差图细化处理,因此,视差图细化单元40可仅包括上述三个单元中的任意一个或两个,而且三个单元之间并无固定的设置顺序。Those skilled in the art should understand that the above-mentioned occlusion
以上示出了根据本发明示例性实施例进行立体图像匹配的设备和方法。图9示出根据本发明示例性实施例的立体图像匹配系统与现有技术相比在性能方面的改进,其中,图9中的(a)和(b)分别示出在Middlebury网站针对误差阈值=1以及误差阈值=0.5的测试结果,如图所示,根据本发明示例性实施例的立体图像匹配方案(在测试结果界面中显示为“YOUR METHOD”)在匹配精度和处理效率方面都明显领先于现有技术中的其它方案。The above shows the device and method for stereo image matching according to the exemplary embodiments of the present invention. Fig. 9 shows the improvement in performance of the stereo image matching system according to an exemplary embodiment of the present invention compared with the prior art, wherein (a) and (b) in Fig. 9 respectively show the error threshold =1 and the test result of error threshold=0.5, as shown in the figure, the stereoscopic image matching scheme (displayed as "YOUR METHOD" in the test result interface) according to the exemplary embodiment of the present invention is all obvious in terms of matching accuracy and processing efficiency ahead of other solutions in the prior art.
根据本发明,能够基于特定的规则计算匹配代价函数,并进一步通过针对每个像素的动态支持域进行代价聚集,在代价聚集的基础上可考虑全局的深度平滑性而产生参考图像的视差图,从而在实现立体图像匹配的过程中同时提高了匹配精度和处理速度。此外,通过对产生的视差图进行细化修正处理,能够进一步提高图像匹配的性能。According to the present invention, the matching cost function can be calculated based on specific rules, and the cost aggregation is further carried out for the dynamic support domain of each pixel. On the basis of the cost aggregation, the global depth smoothness can be considered to generate the disparity map of the reference image. Therefore, the matching accuracy and the processing speed are simultaneously improved in the process of realizing stereoscopic image matching. In addition, the performance of image matching can be further improved by performing thinning and correction processing on the generated disparity map.
应注意,根据本发明示例性实施例的立体图像匹配方法和设备可被包括在用于3D内容的生成装置中,也可被包括在用于3D内容的显示装置中。在上述装置中,除了根据本发明示例性实施例的立体图像匹配设备之外,还包括3D数据输入单元、3D数据分析单元、3D内容产生单元或3D内容显示单元,由于这些单元均属于本发明以外的现有技术,因此,为了避免对本发明的主题造成混淆,在此不进行详细说明。It should be noted that the stereoscopic image matching method and apparatus according to the exemplary embodiments of the present invention may be included in a generating device for 3D content, and may also be included in a display device for 3D content. In the above device, in addition to the stereoscopic image matching device according to the exemplary embodiment of the present invention, it also includes a 3D data input unit, a 3D data analysis unit, a 3D content generation unit or a 3D content display unit, since these units all belong to the present invention Therefore, in order to avoid confusing the subject matter of the present invention, no detailed description is given here.
本发明的以上各个实施例仅仅是示例性的,而本发明并不受限于此。本领域技术人员应该理解:任何分别涉及上述匹配代价处理、像素动态支持域构建、视差图生成和细化来进行立体图像匹配的方式均落入本发明的范围之中。在不脱离本发明的原理和精神的情况下,可对这些实施例进行改变,其中,本发明的范围在权利要求及其等同物中限定。The above respective embodiments of the present invention are merely exemplary, and the present invention is not limited thereto. Those skilled in the art should understand that any method for performing stereo image matching involving the above-mentioned matching cost processing, pixel dynamic support domain construction, disparity map generation and thinning respectively falls within the scope of the present invention. Changes may be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the claims and their equivalents.
Claims (15)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN2010105025844A CN102447917A (en) | 2010-10-08 | 2010-10-08 | Three-dimensional image matching method and equipment thereof |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN2010105025844A CN102447917A (en) | 2010-10-08 | 2010-10-08 | Three-dimensional image matching method and equipment thereof |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN102447917A true CN102447917A (en) | 2012-05-09 |
Family
ID=46009941
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN2010105025844A Pending CN102447917A (en) | 2010-10-08 | 2010-10-08 | Three-dimensional image matching method and equipment thereof |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN102447917A (en) |
Cited By (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102831601A (en) * | 2012-07-26 | 2012-12-19 | 中北大学 | Three-dimensional matching method based on union similarity measure and self-adaptive support weighting |
| CN102881014A (en) * | 2012-09-07 | 2013-01-16 | 航天恒星科技有限公司 | Quick stereo matching method based on graph cut |
| CN103167306A (en) * | 2013-03-22 | 2013-06-19 | 上海大学 | A device and method for real-time extraction of high-resolution depth maps based on image matching |
| CN104380338A (en) * | 2012-05-22 | 2015-02-25 | 索尼电脑娱乐公司 | Information processor and information processing method |
| WO2017067390A1 (en) * | 2015-10-20 | 2017-04-27 | 努比亚技术有限公司 | Method and terminal for obtaining depth information of low-texture regions in image |
| CN107588721A (en) * | 2017-08-28 | 2018-01-16 | 武汉科技大学 | The measuring method and system of a kind of more sizes of part based on binocular vision |
| CN107770512A (en) * | 2016-08-22 | 2018-03-06 | 现代自动车株式会社 | Systems and methods for generating disparity maps by matching stereoscopic images |
| CN108230273A (en) * | 2018-01-05 | 2018-06-29 | 西南交通大学 | A kind of artificial compound eye camera three dimensional image processing method based on geological information |
| CN109544611A (en) * | 2018-11-06 | 2019-03-29 | 深圳市爱培科技术股份有限公司 | A kind of binocular vision solid matching method and system based on bit feature |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20100142828A1 (en) * | 2008-12-10 | 2010-06-10 | Electronics And Telecommunications Research Institute | Image matching apparatus and method |
| CN101841730A (en) * | 2010-05-28 | 2010-09-22 | 浙江大学 | Real-time stereoscopic vision implementation method based on FPGA |
-
2010
- 2010-10-08 CN CN2010105025844A patent/CN102447917A/en active Pending
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20100142828A1 (en) * | 2008-12-10 | 2010-06-10 | Electronics And Telecommunications Research Institute | Image matching apparatus and method |
| CN101841730A (en) * | 2010-05-28 | 2010-09-22 | 浙江大学 | Real-time stereoscopic vision implementation method based on FPGA |
Non-Patent Citations (1)
| Title |
|---|
| KRISTIAN AMBROSCH等: "Parameter Optimization of the SAD-IGMCT forStereo Vision in RGB and HSV Color Spaces", 《ELMAR, 2010 PROCEEDINGS》, 17 September 2010 (2010-09-17) * |
Cited By (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10469829B2 (en) | 2012-05-22 | 2019-11-05 | Sony Interactive Entertainment Inc. | Information processor and information processing method |
| CN104380338A (en) * | 2012-05-22 | 2015-02-25 | 索尼电脑娱乐公司 | Information processor and information processing method |
| CN102831601A (en) * | 2012-07-26 | 2012-12-19 | 中北大学 | Three-dimensional matching method based on union similarity measure and self-adaptive support weighting |
| CN102881014A (en) * | 2012-09-07 | 2013-01-16 | 航天恒星科技有限公司 | Quick stereo matching method based on graph cut |
| CN102881014B (en) * | 2012-09-07 | 2015-02-11 | 航天恒星科技有限公司 | Quick stereo matching method based on graph cut |
| CN103167306A (en) * | 2013-03-22 | 2013-06-19 | 上海大学 | A device and method for real-time extraction of high-resolution depth maps based on image matching |
| WO2017067390A1 (en) * | 2015-10-20 | 2017-04-27 | 努比亚技术有限公司 | Method and terminal for obtaining depth information of low-texture regions in image |
| CN107770512A (en) * | 2016-08-22 | 2018-03-06 | 现代自动车株式会社 | Systems and methods for generating disparity maps by matching stereoscopic images |
| CN107770512B (en) * | 2016-08-22 | 2020-11-06 | 现代自动车株式会社 | System and method for generating disparity map by matching stereo images |
| CN107588721A (en) * | 2017-08-28 | 2018-01-16 | 武汉科技大学 | The measuring method and system of a kind of more sizes of part based on binocular vision |
| CN108230273B (en) * | 2018-01-05 | 2020-04-07 | 西南交通大学 | Three-dimensional image processing method of artificial compound eye camera based on geometric information |
| CN108230273A (en) * | 2018-01-05 | 2018-06-29 | 西南交通大学 | A kind of artificial compound eye camera three dimensional image processing method based on geological information |
| CN109544611A (en) * | 2018-11-06 | 2019-03-29 | 深圳市爱培科技术股份有限公司 | A kind of binocular vision solid matching method and system based on bit feature |
| CN109544611B (en) * | 2018-11-06 | 2021-05-14 | 深圳市爱培科技术股份有限公司 | Binocular vision stereo matching method and system based on bit characteristics |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN102447917A (en) | Three-dimensional image matching method and equipment thereof | |
| CN108596965B (en) | Light field image depth estimation method | |
| CN102999913B (en) | A kind of sectional perspective matching process based on credible propagation | |
| CN103177451B (en) | Based on the self-adapting window of image border and the Stereo Matching Algorithm of weight | |
| US9350969B2 (en) | Target region filling involving source regions, depth information, or occlusions | |
| WO2018127007A1 (en) | Depth image acquisition method and system | |
| CN106408513B (en) | Depth Map Super-Resolution Reconstruction Method | |
| CN109146946B (en) | Image non-local stereo matching method | |
| US9380286B2 (en) | Stereoscopic target region filling | |
| CN105144234A (en) | Generate a depth map for an input image using an example approximate depth map associated with an example similar image | |
| CN106408596B (en) | Edge-Based Local Stereo Matching Method | |
| JP2014150521A (en) | Method and system for increasing resolution of depth image | |
| CN107133914B (en) | Apparatus for generating a three-dimensional color image and method for generating a three-dimensional color image | |
| CN102831601A (en) | Three-dimensional matching method based on union similarity measure and self-adaptive support weighting | |
| Shivakumar et al. | Real time dense depth estimation by fusing stereo with sparse depth measurements | |
| CN102436671A (en) | Virtual viewpoint drawing method based on depth value nonlinear transformation | |
| CN109887008B (en) | Method, device and equipment for parallax stereo matching based on forward and backward smoothing and O (1) complexity | |
| CN112991421B (en) | A Stereo Matching Method for Robot Vision | |
| CN102129567A (en) | Fast stereo matching method based on color partitioning and self-adaptive window | |
| CN104200453A (en) | Parallax image correcting method based on image segmentation and credibility | |
| US8884951B2 (en) | Depth estimation data generating apparatus, depth estimation data generating method, and depth estimation data generating program, and pseudo three-dimensional image generating apparatus, pseudo three-dimensional image generating method, and pseudo three-dimensional image generating program | |
| CN102857772B (en) | Image treatment method and image processor | |
| CN107610070B (en) | Free stereo matching method based on three-camera collection | |
| CN118015059A (en) | A noise-resistant stereo matching method and system based on improved Census transformation | |
| WO2013100789A1 (en) | Method of and apparatus for local optimization texture synthesis 3-d inpainting |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| C06 | Publication | ||
| PB01 | Publication | ||
| C10 | Entry into substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| AD01 | Patent right deemed abandoned |
Effective date of abandoning: 20160302 |
|
| C20 | Patent right or utility model deemed to be abandoned or is abandoned |


