CN102447917A - Three-dimensional image matching method and equipment thereof - Google Patents

Three-dimensional image matching method and equipment thereof Download PDF

Info

Publication number
CN102447917A
CN102447917A CN2010105025844A CN201010502584A CN102447917A CN 102447917 A CN102447917 A CN 102447917A CN 2010105025844 A CN2010105025844 A CN 2010105025844A CN 201010502584 A CN201010502584 A CN 201010502584A CN 102447917 A CN102447917 A CN 102447917A
Authority
CN
China
Prior art keywords
pixel
cost
unit
coupling
target image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2010105025844A
Other languages
Chinese (zh)
Inventor
梅星
崔超
王海涛
孙迅
马赓宇
周明才
金智渊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Samsung Telecommunications Technology Research Co Ltd
Samsung Electronics Co Ltd
Original Assignee
Beijing Samsung Telecommunications Technology Research Co Ltd
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Samsung Telecommunications Technology Research Co Ltd, Samsung Electronics Co Ltd filed Critical Beijing Samsung Telecommunications Technology Research Co Ltd
Priority to CN2010105025844A priority Critical patent/CN102447917A/en
Publication of CN102447917A publication Critical patent/CN102447917A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Processing (AREA)

Abstract

提供一种立体图像匹配方法及其设备,其中,所述设备包括:代价计算单元,用于计算参考图像的像素与目标图像的像素之间的匹配代价,其中,代价计算单元分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;代价聚集单元,用于将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及视差确定单元,用于基于代价聚集单元对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。

Figure 201010502584

A stereoscopic image matching method and device thereof are provided, wherein the device includes: a cost calculation unit configured to calculate a matching cost between pixels of a reference image and pixels of a target image, wherein the cost calculation unit calculates the matching costs of the reference image respectively A non-parametric statistical cost function and a color mean absolute difference function between pixels and pixels of the target image, and linearly combining the non-parametric statistical cost function and the color mean absolute difference function to use the combined result as a reference image The matching cost between the pixel of the pixel and the pixel of the target image; the cost aggregation unit is used to gather the matching cost between the pixel of the reference image calculated by the cost calculation unit and the pixel of the target image; and the parallax determination unit is used for The disparity between the pixels of the reference image and the target image is determined based on the result of the aggregation of the matching cost by the cost aggregation unit, so as to realize stereoscopic image matching.

Figure 201010502584

Description

立体图像匹配方法及其设备Stereo image matching method and device thereof

技术领域 technical field

本发明涉及用于图像处理的方法和设备,尤其涉及一种在参考图像与目标图像之间进行立体图像匹配的方法和设备。The present invention relates to a method and device for image processing, in particular to a method and device for stereoscopic image matching between a reference image and a target image.

背景技术 Background technique

立体图像匹配是计算机视觉领域的重要技术课题,在参考图像(例如,左图像)与目标图像(例如,右图像)之间实施立体图像匹配的过程中获取视差图,其中,视差(即,双目视差)是指在双目视觉中,同一空间点在双目分别看到的左图像和右图像中不同投影之间的位置矢量。实际中,由于人眼对于深度的感觉取决于水平方向的视差,所以视差也可仅仅指位置矢量的水平分量。Stereo image matching is an important technical topic in the field of computer vision. The disparity map is obtained during the process of implementing stereo image matching between a reference image (for example, left image) and a target image (for example, right image). Visual disparity) refers to the position vector between different projections of the same spatial point in the left image and right image seen by the binocular vision in binocular vision. In practice, since the perception of depth by the human eye depends on the parallax in the horizontal direction, the parallax may only refer to the horizontal component of the position vector.

如上所述,视差为产生深度感觉的依据,因此,在3D图像的技术发展中,通过视差估计实现立体图像匹配是计算场景深度、场景三维重建以及三维显示中的核心技术。在现有的立体图像匹配实施方案中,主要关注匹配精度和处理效率这两方面的需求。然而,尽管目前存在很多用于实现立体图像匹配的系统,但是这些系统均难以同时在匹配精度和处理效率这两方面实现较大的改进,往往是某一方面的改进导致牺牲了另一方面的性能。例如,许多精度较好的立体图像匹配方案具有以下共同点:1)采用较大的支持窗口来进行强健的代价聚集运算;2)在视差图的计算中,利用复杂的全局优化器来进行能量最优运算;3)主要依赖于匹配图像的分割结果来实现匹配精度的提高。上述特点虽然能够成功地抑制匹配中的模糊现象,但是却均需要复杂的计算方法为基础,使得计算处理需要耗费大量的时间,以致难以应用于对实时性有要求的情况。而另外一些立体图像匹配方案为了达到实时匹配的性能,采取了简化的措施,例如,图像金字塔、渐次匹配、强健测量和基于简单动态编程的优化器等,其结果是这些系统中的匹配精度不得不降低。As mentioned above, parallax is the basis for generating depth perception. Therefore, in the development of 3D image technology, stereoscopic image matching through parallax estimation is the core technology in calculating scene depth, 3D scene reconstruction and 3D display. In existing implementations of stereo image matching, the main focus is on the requirements of matching accuracy and processing efficiency. However, although there are many systems for stereoscopic image matching, it is difficult for these systems to achieve greater improvements in matching accuracy and processing efficiency at the same time. Often, the improvement in one aspect leads to the sacrifice of the other. performance. For example, many stereo image matching schemes with better accuracy have the following things in common: 1) use a large support window for robust cost aggregation operations; 2) use complex global optimizers to perform energy Optimal operation; 3) Mainly rely on the segmentation results of matching images to achieve the improvement of matching accuracy. Although the above features can successfully suppress the fuzzy phenomenon in matching, they all require complex calculation methods as the basis, which makes calculation and processing take a lot of time, making it difficult to apply to situations that require real-time performance. In order to achieve real-time matching performance, some other stereo image matching schemes adopt simplified measures, such as image pyramid, gradual matching, robust measurement and optimizer based on simple dynamic programming, etc., as a result, the matching accuracy in these systems cannot be achieved Not lowered.

因此,需要提供一种能够在匹配精度和处理速度两方面均有显著改进的立体图像匹配方案,从而能够在实现高精度匹配的同时达到实时处理的运算要求。Therefore, it is necessary to provide a stereoscopic image matching scheme that can significantly improve both matching accuracy and processing speed, so as to achieve high-precision matching and meet the computing requirements for real-time processing.

发明内容 Contents of the invention

本发明的目的在于提供一种能够实现高精度匹配的快速立体图像匹配方法和系统,此外,可通过对生成的视差图进行细化处理来进一步提高立体图像匹配的性能。The purpose of the present invention is to provide a fast stereoscopic image matching method and system capable of achieving high-precision matching. In addition, the performance of stereoscopic image matching can be further improved by performing thinning processing on the generated disparity map.

根据本发明的一方面,提供一种立体图像匹配设备,所述设备包括:代价计算单元,用于计算参考图像的像素与目标图像的像素之间的匹配代价,其中,代价计算单元分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;代价聚集单元,用于将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及视差确定单元,用于基于代价聚集单元对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。According to an aspect of the present invention, there is provided a stereoscopic image matching device, the device comprising: a cost calculation unit for calculating matching costs between pixels of a reference image and pixels of a target image, wherein the cost calculation unit calculates the reference The non-parametric statistical cost function and the color mean absolute difference function between the pixels of the image and the pixels of the target image, and the non-parametric statistical cost function and the color mean absolute difference function are linearly combined to use the result of the combination as a matching cost between the pixels of the reference image and the pixels of the target image; a cost aggregation unit for gathering the matching costs between the pixels of the reference image calculated by the cost calculation unit and the pixels of the target image; and a parallax determination unit, It is used to determine the disparity between the pixels of the reference image and the target image based on the result of the aggregation of the matching cost by the cost aggregation unit, so as to realize stereoscopic image matching.

所述代价聚集单元针可对参考图像中的每个像素确定该像素的动态支持域,在确定的每个动态支持域中,将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集。For each pixel in the reference image, the cost aggregation unit can determine the dynamic support region of the pixel, and in each determined dynamic support region, the difference between the pixel of the reference image calculated by the cost calculation unit and the pixel of the target image The matching costs among them are aggregated.

所述视差确定单元可构建附带深度平滑性约束的匹配代价的全局能量,并将使全局能量取得最小值的视差级别集合确定为最终的立体图像匹配视差图。The disparity determination unit may construct the global energy of the matching cost with a depth smoothness constraint, and determine the set of disparity levels that minimize the global energy as the final stereoscopic image matching disparity map.

所述代价聚集单元可包括:像素动态支持域构建单元,用于针对参考图像中的每个像素构建动态支持域;动态支持域代价聚集单元,用于针对构建的每个动态支持域,对其中的所有像素进行代价聚集。The cost aggregation unit may include: a pixel dynamic support domain construction unit for constructing a dynamic support domain for each pixel in the reference image; a dynamic support domain cost aggregation unit for constructing each dynamic support domain for which All pixels of are cost aggregated.

所述立体图像匹配设备,还可包括:视差图细化单元,用于对由视差确定单元输出的视差图进行细化处理。The stereo image matching device may further include: a disparity map refinement unit, configured to refine the disparity map output by the disparity determination unit.

所述视差图细化单元可包括以下项中的至少一个:遮挡点处理单元,用于检测视差图中的遮挡点,利用遮挡点周围的图像来估计遮挡点的视差;非连续边界修正单元,用于在视差图中检测视差值出现跳变的非连续边界,对于非连续边界上的像素,利用在所述非连续边界两侧与所述像素间隔一定距离的像素来重新确定非连续边界上的像素的视差;子像素级差值单元,用于对视差图进行子像素级别的差值运算。The disparity map refinement unit may include at least one of the following items: an occlusion point processing unit, configured to detect an occlusion point in the disparity map, and use images around the occlusion point to estimate the disparity of the occlusion point; a discontinuous boundary correction unit, It is used to detect a discontinuous boundary in which the disparity value jumps in the disparity map, and for pixels on the discontinuous boundary, use pixels on both sides of the discontinuous boundary that are at a certain distance from the pixel to re-determine the discontinuous boundary The disparity of the pixels above; the sub-pixel level difference unit is used to perform sub-pixel level difference operation on the disparity map.

在所述线性组合中,所述非参数统计代价函数与彩色平均绝对差值函数之间的比例可基于每个像素的局部纹理、局部梯度和/或匹配置信信息来自适应性地设置。In the linear combination, the ratio between the non-parametric statistical cost function and the color mean absolute difference function may be adaptively set based on each pixel's local texture, local gradient and/or matching confidence information.

根据本发明的另一方面,提供一种立体图像匹配方法,所述方法包括:计算参考图像的像素与目标图像的像素之间的匹配代价,其中,分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;将计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及基于对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。According to another aspect of the present invention, a stereoscopic image matching method is provided, the method comprising: calculating the matching cost between the pixels of the reference image and the pixels of the target image, wherein the pixels of the reference image and the pixels of the target image are respectively calculated Between the non-parametric statistical cost function and the color mean absolute difference function, and the non-parametric statistical cost function and the color mean absolute difference function are linearly combined to use the result of the combination as the pixel of the reference image and the pixel of the target image The matching cost between the pixels; the calculated matching costs between the pixels of the reference image and the pixels of the target image are aggregated; and the disparity of the pixels of the reference image to the target image is determined based on the result of the aggregation of the matching costs, to Realize stereo image matching.

进行聚集的步骤可包括:针对参考图像中的每个像素确定该像素的动态支持域,在确定的每个动态支持域中,将计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集。The step of aggregating may include: for each pixel in the reference image, determine the dynamic support region of the pixel, and in each determined dynamic support region, the calculated matching between the pixel of the reference image and the pixel of the target image The cost is aggregated.

确定视差的步骤可包括:构建附带深度平滑性约束的匹配代价的全局能量,并将使全局能量取得最小值的视差级别集合确定为最终的立体图像匹配视差图。The step of determining the disparity may include: constructing a global energy of a matching cost with a depth smoothness constraint, and determining a set of disparity levels that minimize the global energy as a final stereoscopic image matching disparity map.

所述立体图像匹配方法还可包括:对最终的立体图像匹配视差图进行细化处理。The stereo image matching method may further include: performing thinning processing on the final stereo image matching disparity map.

对最终的立体图像匹配视差图进行细化处理的步骤可包括以下步骤中的至少一个:检测视差图中的遮挡点,利用遮挡点周围的图像来估计遮挡点的视差;在视差图中检测视差值出现跳变的非连续边界,对于非连续边界上的像素,利用在所述非连续边界两侧与所述像素间隔一定距离的像素来重新确定非连续边界上的像素的视差;对视差图进行子像素级别的差值运算。The step of refining the final stereo image matching disparity map may include at least one of the following steps: detecting an occlusion point in the disparity map, using images around the occlusion point to estimate the disparity of the occlusion point; detecting a disparity in the disparity map The discontinuous boundary where the difference value jumps, for the pixels on the discontinuous boundary, use the pixels at a certain distance from the pixel on both sides of the discontinuous boundary to re-determine the disparity of the pixels on the discontinuous boundary; for the disparity The image performs the difference operation at the sub-pixel level.

根据本发明的另一方面,提供一种用于3D内容的生成装置,包括:3D数据输入单元、3D数据分析单元、3D内容产生单元,所述生成装置还包括:根据本发明的立体图像匹配设备。According to another aspect of the present invention, there is provided a generating device for 3D content, including: a 3D data input unit, a 3D data analysis unit, and a 3D content generating unit, and the generating device also includes: the stereoscopic image matching according to the present invention equipment.

根据本发明的另一方面,提供一种用于3D内容的显示装置,包括:3D数据输入单元、3D数据分析单元、3D内容产生单元和3D内容显示单元,所述显示装置还包括:根据本发明的立体图像匹配设备。According to another aspect of the present invention, there is provided a display device for 3D content, including: a 3D data input unit, a 3D data analysis unit, a 3D content generation unit, and a 3D content display unit, and the display device further includes: according to the present invention Invented stereo image matching device.

根据本发明的另一方面,提供一种立体图像匹配设备,所述设备包括:代价计算单元,用于计算参考图像的像素与目标图像的像素之间的匹配代价;代价聚集单元,用于针对参考图像中的每个像素确定该像素的动态支持域,在确定的每个动态支持域中,将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及视差确定单元,用于基于代价聚集单元对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。According to another aspect of the present invention, there is provided a stereoscopic image matching device, which includes: a cost calculation unit for calculating the matching cost between pixels of a reference image and pixels of a target image; a cost aggregation unit for Each pixel in the reference image determines the dynamic support region of the pixel, and in each determined dynamic support region, the matching costs between the pixels of the reference image calculated by the cost calculation unit and the pixels of the target image are gathered; and The disparity determination unit is configured to determine the disparity between the pixels of the reference image and the target image based on the result of the aggregation of the matching cost by the cost aggregation unit, so as to achieve stereoscopic image matching.

根据本发明的另一方面,提供一种立体图像匹配方法,所述方法包括:计算参考图像的像素与目标图像的像素之间的匹配代价;针对参考图像中的每个像素确定该像素的动态支持域,在确定的每个动态支持域中,将由代价计算单元计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及基于对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。According to another aspect of the present invention, there is provided a stereoscopic image matching method, the method comprising: calculating the matching cost between a pixel of a reference image and a pixel of a target image; a support domain, in each determined dynamic support domain, the matching costs between the pixels of the reference image calculated by the cost calculation unit and the pixels of the target image are aggregated; and the reference image is determined based on the result of the aggregation of the matching costs The pixel-to-target image disparity for stereo image matching.

根据本发明,能够基于特定的规则计算匹配代价函数,并进一步通过针对每个像素的动态支持域进行代价聚集,在代价聚集的基础上可考虑全局的深度平滑性而产生参考图像的视差图,从而在实现立体图像匹配的过程中同时提高了匹配精度和处理速度。此外,通过对产生的视差图进行细化修正处理,能够进一步提高图像匹配的性能。According to the present invention, the matching cost function can be calculated based on specific rules, and the cost aggregation is further carried out for the dynamic support domain of each pixel. On the basis of the cost aggregation, the global depth smoothness can be considered to generate the disparity map of the reference image. Therefore, the matching accuracy and the processing speed are simultaneously improved in the process of realizing stereoscopic image matching. In addition, the performance of image matching can be further improved by performing thinning and correction processing on the generated disparity map.

附图说明 Description of drawings

通过下面结合附图进行的对实施例的描述,本发明的上述和/或其它目的和优点将会变得更加清楚,其中:The above and/or other objects and advantages of the present invention will become more clear through the following description of the embodiments in conjunction with the accompanying drawings, wherein:

图1是示出根据本发明示例性实施例的立体图像匹配设备的框图;FIG. 1 is a block diagram illustrating a stereoscopic image matching device according to an exemplary embodiment of the present invention;

图2是示出根据本发明示例性实施例的立体图像匹配方法的流程图;FIG. 2 is a flowchart illustrating a stereo image matching method according to an exemplary embodiment of the present invention;

图3是示出根据本发明示例性实施例的代价聚集单元的详细结构的框图;3 is a block diagram showing a detailed structure of a cost aggregation unit according to an exemplary embodiment of the present invention;

图4示出根据本发明示例性实施例来构建像素动态支持域的示例;FIG. 4 shows an example of constructing a pixel dynamic support domain according to an exemplary embodiment of the present invention;

图5示出根据本发明示例性实施例的基于动态支持域进行代价聚集的示例;FIG. 5 shows an example of cost aggregation based on dynamic support domains according to an exemplary embodiment of the present invention;

图6是示出根据本发明示例性实施例的视差图细化单元的详细结构的框图;6 is a block diagram illustrating a detailed structure of a disparity map refinement unit according to an exemplary embodiment of the present invention;

图7示出根据本发明示例性实施例的遮挡点处理的示例;Fig. 7 shows an example of occlusion point processing according to an exemplary embodiment of the present invention;

图8示出根据本发明示例性实施例的非连续边界修正的示例;以及FIG. 8 shows an example of discontinuous boundary correction according to an exemplary embodiment of the present invention; and

图9示出根据本发明示例性实施例的立体图像匹配系统与现有技术相比在性能方面的改进。FIG. 9 shows the improvement in performance of the stereo image matching system according to the exemplary embodiment of the present invention compared with the prior art.

具体实施方式 Detailed ways

现将详细描述本发明的实施例,所述实施例的示例在附图中示出,其中,相同的标号始终指的是相同的部件。以下将通过参照附图来说明所述实施例,以便解释本发明。Reference will now be made in detail to embodiments of the invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like parts throughout. The embodiments are described below in order to explain the present invention by referring to the figures.

图1是示出根据本发明示例性实施例的立体图像匹配设备的框图。如图1所示,根据本发明示例性实施例的立体图像匹配设备包括:代价计算单元10,用于计算参考图像的像素与目标图像的像素之间的匹配代价,其中,代价计算单元10分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;代价聚集单元20,用于将由代价计算单元10计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;以及视差确定单元30,用于基于代价聚集单元20对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。作为优选但非必要的实施方式,还可额外包括视差图细化单元40,用于对由视差确定单元30输出的视差构成的视差图进行细化处理,从而进一步提高立体图像匹配的性能。FIG. 1 is a block diagram illustrating a stereoscopic image matching device according to an exemplary embodiment of the present invention. As shown in FIG. 1 , the stereoscopic image matching device according to an exemplary embodiment of the present invention includes: a cost calculation unit 10 configured to calculate a matching cost between pixels of a reference image and pixels of a target image, wherein the cost calculation unit 10 is respectively Calculate the non-parametric statistical cost function and the color mean absolute difference function between the pixels of the reference image and the pixels of the target image, and linearly combine the non-parametric statistical cost function and the color mean absolute difference function to combine the combined The result is used as the matching cost between the pixels of the reference image and the pixels of the target image; the cost aggregation unit 20 is used to gather the matching costs between the pixels of the reference image calculated by the cost calculation unit 10 and the pixels of the target image; and The disparity determination unit 30 is configured to determine the disparity between the pixels of the reference image and the target image based on the result of the aggregation of the matching costs by the cost aggregation unit 20, so as to achieve stereoscopic image matching. As a preferred but non-essential implementation, a disparity map refinement unit 40 may be additionally included for performing refinement processing on the disparity map formed by the disparity output by the disparity determination unit 30, so as to further improve the performance of stereoscopic image matching.

以下将参照图2来描述利用图1所示的立体图像匹配设备来实现根据本发明的立体图像匹配方法的示例。An example of implementing the stereo image matching method according to the present invention by using the stereo image matching device shown in FIG. 1 will be described below with reference to FIG. 2 .

图2是示出根据本发明示例性实施例的立体图像匹配方法的流程图。参照图2,在步骤S100,由代价计算单元10计算参考图像的像素与目标图像的像素之间的匹配代价,其中,代价计算单元10分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价;在步骤S200,由代价聚集单元20将由代价计算单元10计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集;在步骤S300,由视差确定单元30基于代价聚集单元20对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差,以实现立体图像匹配。作为优选但非必要的实施方式,还可额外包括步骤S400,其中,由视差图细化单元40对由视差确定单元30输出的视差构成的视差图进行细化处理,从而进一步提高立体图像匹配的性能。FIG. 2 is a flowchart illustrating a stereoscopic image matching method according to an exemplary embodiment of the present invention. Referring to Fig. 2, in step S100, the matching cost between the pixels of the reference image and the pixels of the target image is calculated by the cost calculation unit 10, wherein, the cost calculation unit 10 calculates the difference between the pixels of the reference image and the pixels of the target image respectively A parametric statistical cost function and a color mean absolute difference function, and the non-parametric statistical cost function and the color mean absolute difference function are linearly combined, so that the result of the combination is used as the difference between the pixels of the reference image and the pixels of the target image Matching cost; in step S200, the matching cost between the pixels of the reference image calculated by the cost calculation unit 10 and the pixels of the target image is gathered by the cost aggregation unit 20; in step S300, the parallax determination unit 30 is based on the cost aggregation unit 20 Aggregate the matching cost to determine the disparity between the pixels of the reference image and the target image, so as to achieve stereo image matching. As a preferred but non-essential implementation, step S400 may also be additionally included, wherein the disparity map refinement unit 40 performs refinement processing on the disparity map formed by the disparity output by the disparity determination unit 30, thereby further improving the accuracy of stereo image matching. performance.

以下,首先描述在步骤S100,由代价计算单元10计算参考图像的像素与目标图像的像素之间的匹配代价的处理。In the following, first, the process of calculating the matching cost between the pixels of the reference image and the pixels of the target image by the cost calculation unit 10 at step S100 will be described.

代价计算单元10接收参考图像和目标图像,其中,作为示例,参考图像为左图像,而目标图像为右图像。然后,对于左图像中的每个像素p(x,y),计算其与右图像中的像素pd(x-d,y)的匹配代价,其中,x、y分别指示像素p的横坐标和纵坐标,d表示给定的视差级别,d的取值范围可根据实际需要以整像素为单位而设置。The cost calculation unit 10 receives a reference image and a target image, wherein, as an example, the reference image is a left image and the target image is a right image. Then, for each pixel p(x, y) in the left image, calculate its matching cost with the pixel pd(x-d, y) in the right image, where x, y indicate the abscissa and ordinate of pixel p respectively , d represents a given disparity level, and the value range of d can be set in units of integer pixels according to actual needs.

在根据本发明的当前示例性实施例中,代价计算单元10分别计算参考图像的像素与目标图像的像素之间的非参数统计代价函数和彩色平均绝对差值函数,并将所述非参数统计代价函数与彩色平均绝对差值函数进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价。In the current exemplary embodiment according to the present invention, the cost calculation unit 10 respectively calculates the non-parametric statistical cost function and the color mean absolute difference function between the pixels of the reference image and the pixels of the target image, and calculates the non-parametric statistical cost function The cost function is linearly combined with the color mean absolute difference function, so that the result of the combination is used as the matching cost between the pixels of the reference image and the pixels of the target image.

具体说来,代价计算单元10首先计算像素p(x,y)与像素pd(x-d,y)之间的非参数统计代价函数,在这种情况下,选择像素p(x,y)周围的方形窗口Np以及对应的像素pd(x-d,y)周围的方形窗口Npd,并分别将所选择的方形窗口Np和Npd中的像素按照一定的顺序表示为比特串的形式,在所述比特串中,通过比特Bitp(p1)表示像素p(x,y)的强度与其方形窗口Np中的任一像素p1的强度之间的相对关系(例如,大于、小于或近似),举例说来,当像素p(x,y)的强度大于像素p1的强度时,将Bitp(p1)值设置为1,否则将其设置为0,其中,可根据实际需要来设定各种具体的比较关系;类似地,通过比特Bitpd(pd1)表示像素pd(x-d,y)的强度与其方形窗口Npd中的任一像素pd1的强度之间的相对关系(例如,大于、小于或近似),举例说来,当像素pd(x-d,y)的强度大于像素pd1的强度时,将Bitpd(pd1)值设置为1,否则将其设置为0,其中,可根据实际需要来设定各种具体的比较关系。Specifically, the cost calculation unit 10 first calculates the non-parametric statistical cost function between the pixel p(x, y) and the pixel pd(xd, y), in this case, selects the pixels around the pixel p(x, y) square window N p and the square window N pd around the corresponding pixel pd(xd, y), and respectively represent the pixels in the selected square window N p and N pd as a bit string in a certain order, in the In the above bit string, bit Bit p (p 1 ) represents the relative relationship between the intensity of pixel p(x, y) and the intensity of any pixel p 1 in its square window N p (for example, greater than, less than or approximately ), for example, when the intensity of the pixel p(x, y) is greater than the intensity of the pixel p 1 , the value of Bit p (p 1 ) is set to 1, otherwise it is set to 0, which can be determined according to actual needs Set various specific comparison relationships; similarly, the relative relationship between the intensity of pixel pd (xd, y) and the intensity of any pixel pd1 in its square window N pd is represented by bit Bit pd (pd 1 ) (for example , greater than, less than or approximately), for example, when the intensity of pixel pd(xd, y) is greater than the intensity of pixel pd 1 , the value of Bit pd (pd 1 ) is set to 1, otherwise it is set to 0, where , various specific comparison relationships can be set according to actual needs.

基于上述设定,代价计算单元10按照下面的等式1来计算像素p(x,y)与像素pd(x-d,y)之间的非参数统计代价函数Ccensus(p,d):Based on the above settings, the cost calculation unit 10 calculates the non-parametric statistical cost function C census (p, d) between the pixel p (x, y) and the pixel pd (xd, y) according to the following equation 1:

C census ( p , d ) = Σ p 1 ∈ N , pd 1 ∈ N pd W ( p 1 , p ) · W ( pd 1 , pd ) · ( Bit p ( p 1 ) ⊗ Bit pd ( pd 1 ) ) 等式1 C census ( p , d ) = Σ p 1 ∈ N , pd 1 ∈ N pd W ( p 1 , p ) · W ( pd 1 , pd ) &Center Dot; ( bit p ( p 1 ) ⊗ bit pd ( pd 1 ) ) Equation 1

其中,W(p1,p)和W(pd1,pd)分别为方形窗口Np与Npd之间的汉明距离的变化权重,W(p1,p)值可根据像素p1与像素p之间在空间和色彩上的关系来计算得出,具体说来,可根据以下规则来得出W(p1,p)值:当像素p1与像素p之间的距离越近,W(p1,p)值越大,同时,当像素p1与像素p的颜色越相近,W(p1,p)值也越大;类似地,W(pd1,pd)值可根据像素pd1与像素pd之间在空间和色彩上的关系来计算得出,具体说来,可根据以下规则来得出W(pd1,pd)值:当像素pd1与像素pd之间的距离越近,W(pd1,pd)值越大,同时,当像素pd1与像素pd的颜色越相近,W(pd1,pd)值也越大。Among them, W(p 1 , p) and W(pd 1 , pd) are the change weights of the Hamming distance between the square window N p and N pd respectively, and the value of W(p 1 , p) can be determined according to the pixel p 1 and The relationship between pixels p in space and color can be calculated. Specifically, the value of W(p 1 , p) can be obtained according to the following rules: when the distance between pixel p 1 and pixel p is closer, W The larger the value of (p 1 , p) is, at the same time, when the color of pixel p 1 is closer to that of pixel p, the value of W(p 1 , p) is also larger; similarly, the value of W(pd 1 , pd) can be calculated according to pixel The relationship between pd 1 and pixel pd in terms of space and color can be calculated. Specifically, the value of W(pd 1 , pd) can be obtained according to the following rules: when the distance between pixel pd 1 and pixel pd is greater Recently, the larger the value of W(pd 1 , pd), at the same time, the closer the color of the pixel pd 1 to the pixel pd, the larger the value of W(pd 1 , pd).

在代价计算单元10如上计算出像素p(x,y)与像素pd(x-d,y)之间的非参数统计代价函数Ccensus(p,d)之后,所述代价计算单元10来计算参考图像的像素与目标图像的像素之间的彩色平均绝对差值函数。作为示例,可通过等式2来计算像素p(x,y)与像素pd(x-d,y)在R(红)、G(绿)、B(蓝)三色通道上的彩色平均绝对差值函数CAD(p,d):After the cost calculation unit 10 calculates the non-parametric statistical cost function C census (p, d) between the pixel p (x, y) and the pixel pd (xd, y) as above, the cost calculation unit 10 calculates the reference image The color mean absolute difference function between the pixels of and the pixels of the target image. As an example, Equation 2 can be used to calculate the color average absolute difference between the pixel p(x, y) and the pixel pd(xd, y) on the three color channels of R (red), G (green), and B (blue) Function C AD (p, d):

C AD ( p , d ) = min ( 1 3 Σ R , G , B | I ( p ) - I ( pd ) | , λ AD ) 等式2 C AD ( p , d ) = min ( 1 3 Σ R , G , B | I ( p ) - I ( pd ) | , λ AD ) Equation 2

其中,I(p)和I(pd)分别表示像素p(x,y)与像素pd(x-d,y)在对应的R、G、B通道下的强度值,λAD为截断阈值。Among them, I(p) and I(pd) represent the intensity values of pixel p(x, y) and pixel pd(xd, y) under the corresponding R, G, B channels respectively, and λ AD is the truncation threshold.

然后,代价计算单元10将所述非参数统计代价函数Ccensus(p,d)与彩色平均绝对差值函数CAD(p,d)进行线性组合,以将组合的结果作为参考图像的像素与目标图像的像素之间的匹配代价C(p,d)。Then, the cost calculation unit 10 linearly combines the non-parametric statistical cost function C census (p, d) and the color average absolute difference function C AD (p, d) to use the result of the combination as the pixel of the reference image and The matching cost C(p,d) between pixels of the target image.

作为示例,代价计算单元10按照等式3来计算参考图像的像素与目标图像的像素之间的匹配代价C(p,d):As an example, the cost calculation unit 10 calculates the matching cost C(p, d) between the pixels of the reference image and the pixels of the target image according to Equation 3:

C(p,d)=z(p)·Ccensus(p,d)+(1-z(p))·CAD(p,d)等式3C(p,d)=z(p)·C census (p,d)+(1-z(p))·C AD (p,d) Equation 3

其中,z(p)为用于控制非参数统计代价函数Ccensus(p,d)与彩色平均绝对差值函数CAD(p,d)的比例关系的权重,其值在0与1之间。例如,可考虑每个像素p(x,y)的局部纹理、局部梯度和/或匹配置信(matching confidence)信息来自适应性地设置z(p)。Among them, z(p) is the weight used to control the proportional relationship between the non-parametric statistical cost function C census (p, d) and the color average absolute difference function C AD (p, d), and its value is between 0 and 1 . For example, z(p) may be adaptively set in consideration of local texture, local gradient and/or matching confidence information of each pixel p(x,y).

在如上所述得到参考图像中的像素匹配代价函数之后,可在步骤S200,由代价聚集单元20将由代价计算单元10计算出的参考图像的像素与目标图像的像素之间的匹配代价进行聚集,从而产生代价聚集结果Caggr(p,d)。After obtaining the pixel matching cost function in the reference image as described above, in step S200, the cost aggregation unit 20 aggregates the matching costs between the pixels of the reference image calculated by the cost calculation unit 10 and the pixels of the target image, Thus, the cost aggregation result C aggr (p, d) is generated.

这里,作为一种可选方式,可采用根据本发明示例性实施例的优选聚集方案,即,基于每个像素的动态支持域来进行代价聚集。以下,将参照图3来详细描述根据本发明示例性实施例的代价聚集单元的详细结构。Here, as an optional manner, the preferred aggregation scheme according to the exemplary embodiment of the present invention may be adopted, that is, the cost aggregation is performed based on the dynamic support region of each pixel. Hereinafter, the detailed structure of the cost aggregation unit according to the exemplary embodiment of the present invention will be described in detail with reference to FIG. 3 .

图3是示出根据本发明示例性实施例的代价聚集单元的详细结构的框图。参照图3,作为一种示例性结构,代价聚集单元20可包括:像素动态支持域构建单元201,用于针对参考图像中的每个像素构建动态支持域;动态支持域代价聚集单元202,用于针对构建的每个动态支持域,对其中的所有像素进行代价聚集。FIG. 3 is a block diagram illustrating a detailed structure of a cost aggregation unit according to an exemplary embodiment of the present invention. Referring to FIG. 3 , as an exemplary structure, the cost aggregation unit 20 may include: a pixel dynamic support domain construction unit 201 for constructing a dynamic support domain for each pixel in a reference image; a dynamic support domain cost aggregation unit 202 for For each dynamic support domain constructed, cost aggregation is performed on all pixels in it.

如本领域技术人员所知,构建的动态支持域作为每个像素进行代价聚合的基础,其形状和大小对匹配结果的精确性有很大的影响。例如,对于纹理较少的区域需要构建较大的支持域,但是如果所述支持域跨过图像中某对象的边界,则在对象边界处的匹配性能较差。另一方面,较小的支持域对于图像噪声和拐点也比较敏感。因此,综合上述因素,像素动态支持域构建单元201可遵循以下两条准则来构建每个像素的动态支持域:(1)动态支持域中的像素具有与该支持域的中心像素相类似的彩色信息;(2)动态支持域的边界应遵循参考图像中的已有边缘,其中,可利用像素局部的图像梯度信息来检测参考图像中的已有边缘。图4示出根据本发明示例性实施例来构建像素动态支持域的示例。如图4所示,针对像素P和像素Q的动态支持域分别基于上述两条准则被构建。As known to those skilled in the art, the constructed dynamic support domain serves as the basis for cost aggregation for each pixel, and its shape and size have a great influence on the accuracy of the matching result. For example, a larger support domain needs to be constructed for an area with less texture, but if the support domain crosses the boundary of an object in the image, the matching performance at the object boundary will be poor. On the other hand, smaller support domains are also more sensitive to image noise and inflection points. Therefore, considering the above factors, the pixel dynamic support region construction unit 201 can follow the following two criteria to construct the dynamic support region of each pixel: (1) the pixels in the dynamic support region have a color similar to the central pixel of the support region (2) The boundary of the dynamic support domain should follow the existing edges in the reference image, where the local image gradient information of the pixel can be used to detect the existing edges in the reference image. Fig. 4 shows an example of constructing a pixel dynamic support domain according to an exemplary embodiment of the present invention. As shown in FIG. 4 , the dynamic support domains for pixel P and pixel Q are respectively constructed based on the above two criteria.

在像素动态支持域构建单元201构建了各个像素的动态支持域之后,动态支持域代价聚集单元202针对构建的每个动态支持域,对其中的所有像素进行代价聚集。After the pixel dynamic support domain construction unit 201 constructs the dynamic support domain of each pixel, the dynamic support domain cost aggregation unit 202 performs cost aggregation on all the pixels in each constructed dynamic support domain.

作为示例,动态支持域代价聚集单元202可按照图5所示的方式对像素P的动态支持域中的所有像素进行代价聚集。如图5所示,动态支持域代价聚集单元202可按照两个通路对像素P的动态支持域中的所有像素进行代价聚集:首先,沿水平方向分别对每一行的所有像素进行代价聚集,得到并存储水平方向的各个聚集结果,然后,将存储的水平方向的聚焦结果沿垂直方向进行聚合。As an example, the dynamic support region cost aggregation unit 202 may perform cost aggregation on all pixels in the dynamic support region of the pixel P in the manner shown in FIG. 5 . As shown in Figure 5, the dynamic support region cost aggregation unit 202 can perform cost aggregation on all pixels in the dynamic support region of pixel P according to two paths: first, perform cost aggregation on all pixels in each row along the horizontal direction, and obtain Each aggregation result in the horizontal direction is stored, and then the stored focusing results in the horizontal direction are aggregated along the vertical direction.

本领域技术人员应理解,上述通过图3所示的代价聚集单元20基于像素的动态支持域进行代价匹配处理的示例并不依赖于特定的像素匹配代价算法,即,进行聚集的像素匹配代价函数并不受限于由图1的代价计算单元10产生的具体代价函数,而是可以应用于各种像素的匹配代价函数。虽然上述动态域为二维空间域,还可将空间和视差域的三维支持域应用于本发明中。此外,本领域技术人员还应理解,图3所示的像素动态支持域构建单元201和动态支持域代价聚集单元202仅仅是一种示例,这种示例并不构成对代价聚集单元20的结构限制。也就是说,代价聚集单元20完全可以作为单个的实体单元来实现像素动态支持域构建单元201和动态支持域代价聚集单元202整体执行的操作,或者,代价聚集单元20还可被进一步细分为更多的单元来执行所述操作。Those skilled in the art should understand that the above-mentioned example of performing cost matching processing based on the dynamic support domain of pixels by the cost aggregation unit 20 shown in FIG. It is not limited to the specific cost function generated by the cost calculation unit 10 of FIG. 1 , but can be applied to a matching cost function of various pixels. Although the dynamic domain described above is a two-dimensional spatial domain, the three-dimensional support domain of the spatial and disparity domains can also be applied in the present invention. In addition, those skilled in the art should also understand that the pixel dynamic support domain construction unit 201 and the dynamic support domain cost aggregation unit 202 shown in FIG. . That is to say, the cost aggregation unit 20 can be used as a single entity unit to implement the overall operation performed by the pixel dynamic support domain construction unit 201 and the dynamic support domain cost aggregation unit 202, or the cost aggregation unit 20 can be further subdivided into more units to perform the operations.

参照回图2,在如上所述得到代价函数的聚集结果之后,可在步骤S300,由视差确定单元30基于代价聚集单元20对匹配代价进行聚集的结果来确定参考图像的像素到目标图像的视差。具体说来,视差确定单元30可利用能量优化器来计算视差图,即,基于特定的能量运算来确定使得匹配代价最小的视差级别集合

Figure BSA00000297375800091
其中,p表示参考图像中的任一像素,dp为该像素p的视差值。如果直接对聚集的代价结果应用局部WTA(Winner Take All,胜者通吃)算法来选择分别对每个像素p而言具有最小匹配代价的视差,则产生的视差图中会存在很多噪声深度值。因此,作为一种优选的方式,可采用包括对深度平滑性进行附加约束的全局算法,将确定视差的处理构建为确定使全局能量得到最小值的视差级别集合
Figure BSA00000297375800092
作为确定的视差图。Referring back to FIG. 2, after the aggregation result of the cost function is obtained as described above, in step S300, the parallax of the pixels of the reference image to the target image can be determined by the parallax determination unit 30 based on the result of the aggregation of the matching cost by the cost aggregation unit 20 . Specifically, the disparity determining unit 30 can use an energy optimizer to calculate a disparity map, that is, based on a specific energy operation, determine a disparity level set that minimizes the matching cost
Figure BSA00000297375800091
Among them, p represents any pixel in the reference image, and d p is the disparity value of the pixel p. If the local WTA (Winner Take All) algorithm is directly applied to the aggregated cost results to select the disparity with the smallest matching cost for each pixel p, the resulting disparity map will contain many noisy depth values . Therefore, as a preferred method, a global algorithm including additional constraints on depth smoothness can be used, and the process of determining the disparity can be constructed as a set of disparity levels that minimize the global energy
Figure BSA00000297375800092
as a definite disparity map.

作为示例,视差确定单元30可按照等式4来构建附带深度平滑性约束的匹配代价的全局能量,而使得该全局能量取得最小值的视差级别集合则被确定为最终的立体图像匹配视差图:As an example, the disparity determining unit 30 may construct the global energy of the matching cost with a depth smoothness constraint according to Equation 4, and the disparity level set that makes the global energy obtain the minimum value is determined as the final stereoscopic image matching disparity map:

E min ( D ) = Σ p ( C aggr ( p , d p ) + Σ q ∈ N P 1 T [ | d p - d q | = 1 ] + Σ q ∈ N P 2 T [ | d p - d q | > 1 ] ) 等式4 E. min ( D. ) = Σ p ( C aggr ( p , d p ) + Σ q ∈ N P 1 T [ | d p - d q | = 1 ] + Σ q ∈ N P 2 T [ | d p - d q | > 1 ] ) Equation 4

其中,Caggr(p,dp)为匹配代价的聚集结果,N为像素p的邻域,可根据实际需要来确定,q为邻域N中的像素,dq为像素q的视差值,T(A)为真值函数,即,当A等式成立时,T(A)值为1,否则,T(A)值为0,P1和P2为惩罚参数,可根据实际需要来设置。此外,作为优选方式,可采用4向扫描线优化方式来有效地执行上述确定的处理流程。Among them, C aggr (p, d p ) is the aggregation result of the matching cost, N is the neighborhood of pixel p, which can be determined according to actual needs, q is the pixel in the neighborhood N, and d q is the disparity value of pixel q , T(A) is a truth function, that is, when the A equation is established, the value of T(A) is 1, otherwise, the value of T(A) is 0, and P 1 and P 2 are penalty parameters, which can be adjusted according to actual needs to set. In addition, as a preferred method, the 4-direction scan line optimization method can be used to effectively perform the above determination processing flow.

参照回图2,在由视差确定单元30确定了参考图像的像素到目标图像的视差图之后,作为优选但非必要的实施方式,还可额外包括步骤S400,其中,由视差图细化单元40对由视差确定单元30输出的视差图进行细化处理,从而进一步提高立体图像匹配的性能。Referring back to FIG. 2 , after the disparity map from the pixels of the reference image to the target image is determined by the disparity determination unit 30, as a preferred but not necessary implementation, step S400 may be additionally included, wherein the disparity map refinement unit 40 Thinning processing is performed on the disparity map output by the disparity determination unit 30, so as to further improve the performance of stereoscopic image matching.

以下,将参照图6来根据本发明示例性实施例的视差图细化单元40的详细结构。Hereinafter, the detailed structure of the disparity map refinement unit 40 according to an exemplary embodiment of the present invention will be described with reference to FIG. 6 .

图6是示出根据本发明示例性实施例的视差图细化单元的详细结果的框图。参照图6,作为一种示例性结构,视差图细化单元40可包括:遮挡点处理单元401,用于检测视差图中的遮挡点,利用遮挡点周围的图像来估计遮挡点的视差;非连续边界修正单元402,用于在由遮挡点处理单元401处理过的视差图中检测视差值出现跳变的非连续边界,对于非连续边界上的像素,利用在所述非连续边界两侧与所述像素间隔一定距离的像素来重新确定非连续边界上的像素的视差;子像素级差值单元403,用于对由非连续边界修正单元402修正过的视差图进行子像素级别的差值运算。FIG. 6 is a block diagram illustrating detailed results of a disparity map refinement unit according to an exemplary embodiment of the present invention. Referring to FIG. 6, as an exemplary structure, the disparity map refinement unit 40 may include: an occlusion point processing unit 401, configured to detect an occlusion point in the disparity map, and use images around the occlusion point to estimate the disparity of the occlusion point; The continuous boundary correction unit 402 is configured to detect a discontinuous boundary where the disparity value jumps in the disparity map processed by the occlusion point processing unit 401, and for pixels on the discontinuous boundary, use The disparity of the pixels on the discontinuous boundary is re-determined by pixels at a certain distance from the pixel; the sub-pixel level difference unit 403 is used to perform sub-pixel level difference on the disparity map corrected by the discontinuous boundary correction unit 402 Value operations.

具体说来,遮挡点处理单元401通过检查参考图像与目标图像的左右一致性来检测视差图中的遮挡点(即,无法通过目标图像中的相应像素确定其视差值的点)。然后,如图7中的(a)所示,首先利用该遮挡点的动态支持域中的其它像素来估计遮挡点的视差值,例如,可将所述动态支持域的所有像素的视差值中,最为普遍的视差值作为遮挡点的视差值。通过这种方式,可确定某些或全部遮挡点的视差值,对于仍旧无法确定视差值的遮挡点,还可如图7中的(b)所示,从遮挡点出发沿多个方向(例如,16个方向)进行扫描,将沿这些方向扫描到的像素中,色彩与遮挡点最为接近的像素的视差值确定为所述遮挡点的视差值。Specifically, the occlusion point processing unit 401 detects occlusion points in the disparity map (ie, points whose disparity values cannot be determined by corresponding pixels in the target image) by checking the left-right consistency between the reference image and the target image. Then, as shown in (a) in Figure 7, the disparity value of the occlusion point is first estimated by using other pixels in the dynamic support domain of the occlusion point, for example, the disparity values of all pixels in the dynamic support domain can be Among the values, the most common disparity value is used as the disparity value of the occluded point. In this way, the parallax values of some or all occlusion points can be determined. For occlusion points whose parallax values still cannot be determined, as shown in (b) in Figure 7, starting from the occlusion point along multiple directions Scanning is performed in (for example, 16 directions), and the disparity value of the pixel whose color is closest to the occlusion point among the pixels scanned along these directions is determined as the disparity value of the occlusion point.

然后,非连续边界修正单元402在由遮挡点处理单元401处理过的视差图中检测视差值出现跳变(例如,视差值的差>2)的非连续边界,然后,对于非连续边界上的像素,非连续边界修正单元402利用在所述非连续边界两侧与所述像素间隔一定距离(例如,4个像素)的两个像素来重新确定非连续边界上的像素的视差。例如,可将所述两个像素的彩色值分别与非连续边界上的所述像素进行比较,将与非连续边界上的所述像素的颜色较为相近的像素的视差值选为所述边界上的像素的视差值。作为优选方式,可利用像素的动态支持域的彩色信息来代表像素的彩色值,从而减少相应的运算消耗。上述非连续边界修正处理的结果如图8所示,在图8中,左图示出修正前检测出的非连续边界,右图示出修正后的结果。Then, the discontinuous boundary correction unit 402 detects a discontinuous boundary whose disparity value jumps (for example, the difference of the disparity value>2) in the disparity map processed by the occlusion point processing unit 401, and then, for the discontinuous boundary For pixels on the discontinuous boundary, the discontinuous boundary correction unit 402 uses two pixels on both sides of the discontinuous boundary separated from the pixel by a certain distance (for example, 4 pixels) to re-determine the disparity of the pixel on the discontinuous boundary. For example, the color values of the two pixels may be compared with the pixels on the discontinuous boundary, and the disparity value of the pixel whose color is closer to the pixel on the discontinuous boundary is selected as the boundary The disparity value of the pixel on . As a preferred manner, the color information of the dynamic support domain of the pixel can be used to represent the color value of the pixel, thereby reducing the corresponding calculation consumption. The results of the discontinuous boundary correction processing described above are shown in FIG. 8 . In FIG. 8 , the left figure shows the detected discontinuous boundary before correction, and the right figure shows the result after correction.

然后,子像素级差值单元403可对由非连续边界修正单元402修正过的视差图进行子像素级别的差值运算。由于之前的处理均基于整数像素级别,为了进一步对产生的视差图进行细化,例如,可按照等式5进行子像素级别的差值运算:Then, the sub-pixel level difference unit 403 can perform a sub-pixel level difference operation on the disparity map corrected by the discontinuous boundary correction unit 402 . Since the previous processing is based on the integer pixel level, in order to further refine the generated disparity map, for example, the difference operation at the sub-pixel level can be performed according to Equation 5:

d * = d - E ( p , d + 1 ) - E ( p , d - 1 ) 2 ( E ( p , d + 1 ) + E ( p , d - 1 ) - 2 E ( p , d ) ) 等式5 d * = d - E. ( p , d + 1 ) - E. ( p , d - 1 ) 2 ( E. ( p , d + 1 ) + E. ( p , d - 1 ) - 2 E. ( p , d ) ) Equation 5

其中,d*为子像素级别的视差值,E表示通过等式4计算的能量。where d * is the disparity value at the sub-pixel level, and E represents the energy calculated by Equation 4.

本领域技术人员应理解,上述通过图6所示的遮挡点处理单元401、非连续边界修正单元402和子像素级差值单元403仅仅是一种示例,这种示例并不构成对视差图细化单元40的结构限制。也就是说,视差图细化单元40完全可以作为单个的实体单元来实现遮挡点处理单元401、非连续边界修正单元402和子像素级差值单元403整体执行的操作,或者,视差图细化单元40还可被进一步细分为更多的单元来执行所述操作。另外,上述遮挡点处理单元401、非连续边界修正单元402和子像素级差值单元403分别执行了三种独立的视差图细化处理,因此,视差图细化单元40可仅包括上述三个单元中的任意一个或两个,而且三个单元之间并无固定的设置顺序。Those skilled in the art should understand that the above-mentioned occlusion point processing unit 401, discontinuous boundary correction unit 402 and sub-pixel level difference unit 403 shown in FIG. Structural constraints of unit 40. That is to say, the disparity map refinement unit 40 can completely implement the operations performed by the occlusion point processing unit 401, the discontinuous boundary correction unit 402, and the sub-pixel level difference unit 403 as a single entity unit, or the disparity map refinement unit 40 can also be further subdivided into more units to perform the operations. In addition, the above-mentioned occlusion point processing unit 401, discontinuous boundary correction unit 402 and sub-pixel level difference unit 403 respectively perform three independent disparity map refinement processes, therefore, the disparity map refinement unit 40 may only include the above three units Any one or two of them, and there is no fixed order of setting among the three units.

以上示出了根据本发明示例性实施例进行立体图像匹配的设备和方法。图9示出根据本发明示例性实施例的立体图像匹配系统与现有技术相比在性能方面的改进,其中,图9中的(a)和(b)分别示出在Middlebury网站针对误差阈值=1以及误差阈值=0.5的测试结果,如图所示,根据本发明示例性实施例的立体图像匹配方案(在测试结果界面中显示为“YOUR METHOD”)在匹配精度和处理效率方面都明显领先于现有技术中的其它方案。The above shows the device and method for stereo image matching according to the exemplary embodiments of the present invention. Fig. 9 shows the improvement in performance of the stereo image matching system according to an exemplary embodiment of the present invention compared with the prior art, wherein (a) and (b) in Fig. 9 respectively show the error threshold =1 and the test result of error threshold=0.5, as shown in the figure, the stereoscopic image matching scheme (displayed as "YOUR METHOD" in the test result interface) according to the exemplary embodiment of the present invention is all obvious in terms of matching accuracy and processing efficiency ahead of other solutions in the prior art.

根据本发明,能够基于特定的规则计算匹配代价函数,并进一步通过针对每个像素的动态支持域进行代价聚集,在代价聚集的基础上可考虑全局的深度平滑性而产生参考图像的视差图,从而在实现立体图像匹配的过程中同时提高了匹配精度和处理速度。此外,通过对产生的视差图进行细化修正处理,能够进一步提高图像匹配的性能。According to the present invention, the matching cost function can be calculated based on specific rules, and the cost aggregation is further carried out for the dynamic support domain of each pixel. On the basis of the cost aggregation, the global depth smoothness can be considered to generate the disparity map of the reference image. Therefore, the matching accuracy and the processing speed are simultaneously improved in the process of realizing stereoscopic image matching. In addition, the performance of image matching can be further improved by performing thinning and correction processing on the generated disparity map.

应注意,根据本发明示例性实施例的立体图像匹配方法和设备可被包括在用于3D内容的生成装置中,也可被包括在用于3D内容的显示装置中。在上述装置中,除了根据本发明示例性实施例的立体图像匹配设备之外,还包括3D数据输入单元、3D数据分析单元、3D内容产生单元或3D内容显示单元,由于这些单元均属于本发明以外的现有技术,因此,为了避免对本发明的主题造成混淆,在此不进行详细说明。It should be noted that the stereoscopic image matching method and apparatus according to the exemplary embodiments of the present invention may be included in a generating device for 3D content, and may also be included in a display device for 3D content. In the above device, in addition to the stereoscopic image matching device according to the exemplary embodiment of the present invention, it also includes a 3D data input unit, a 3D data analysis unit, a 3D content generation unit or a 3D content display unit, since these units all belong to the present invention Therefore, in order to avoid confusing the subject matter of the present invention, no detailed description is given here.

本发明的以上各个实施例仅仅是示例性的,而本发明并不受限于此。本领域技术人员应该理解:任何分别涉及上述匹配代价处理、像素动态支持域构建、视差图生成和细化来进行立体图像匹配的方式均落入本发明的范围之中。在不脱离本发明的原理和精神的情况下,可对这些实施例进行改变,其中,本发明的范围在权利要求及其等同物中限定。The above respective embodiments of the present invention are merely exemplary, and the present invention is not limited thereto. Those skilled in the art should understand that any method for performing stereo image matching involving the above-mentioned matching cost processing, pixel dynamic support domain construction, disparity map generation and thinning respectively falls within the scope of the present invention. Changes may be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the claims and their equivalents.

Claims (15)

1. stereo-picture matching unit, said equipment comprises:
The cost computing unit; Be used to calculate the coupling cost between the pixel of pixel and target image of reference picture; Wherein, The cost computing unit calculates nonparametric statistics cost function and the colored mean absolute difference value function between the pixel of pixel and target image of reference picture respectively, and said nonparametric statistics cost function and colored mean absolute difference value function are carried out linear combination, with the result that will the make up coupling cost between the pixel of pixel and the target image of image as a reference;
The cost accumulation unit is used for the coupling cost between the pixel of the pixel of the reference picture that is calculated by the cost computing unit and target image is assembled; And
Parallax is confirmed the unit, is used for confirming the parallax of the pixel of reference picture to target image based on the result that the cost accumulation unit is assembled the coupling cost, to realize the stereo-picture coupling.
2. equipment as claimed in claim 1; Wherein, Said cost accumulation unit is confirmed the dynamic support region of this pixel to each pixel in the reference picture; In each dynamic support region of confirming, the coupling cost between the pixel of the reference picture that will be calculated by the cost computing unit and the pixel of target image is assembled.
3. equipment as claimed in claim 1, wherein, said parallax is confirmed the global energy of the coupling cost of the subsidiary degree of depth flatness constraint of cell formation, and the parallax level set that will make global energy obtain minimum value is confirmed as final stereo-picture coupling disparity map.
4. equipment as claimed in claim 3, wherein, said cost accumulation unit comprises:
The dynamic support region construction unit of pixel is used for making up dynamic support region to each pixel of reference picture;
Dynamically support region cost accumulation unit is used for to each the dynamic support region that makes up, and all pixels is wherein carried out cost assemble.
5. equipment as claimed in claim 3 also comprises: disparity map refinement unit is used for the disparity map of being confirmed unit output by parallax is carried out thinning processing.
6. equipment as claimed in claim 5, wherein, said disparity map refinement unit comprises at least one in following:
Block a processing unit, be used for detecting blocking a little of disparity map, utilize the image around blocking a little to estimate to block parallax a little;
Discontinuous border amending unit; Be used for detecting the discontinuous border that saltus step appears in parallax value at disparity map; For discontinuous borderline pixel, the pixel that is utilized in said discontinuous boundaries on either side and said pixel separation certain distance to confirm again the parallax of discontinuous borderline pixel;
The sub-pixel-level difference unit is used for disparity map is carried out other difference computing of sub-pixel-level.
7. equipment as claimed in claim 1; Wherein, In said linear combination, the ratio between said nonparametric statistics cost function and the colored mean absolute difference value function comes adaptively to be provided with based on local grain, partial gradient and/or the coupling confidential information of each pixel.
8. stereo-picture matching process, said method comprises:
Coupling cost between the pixel of calculating reference picture and the pixel of target image; Wherein, Calculate nonparametric statistics cost function and colored mean absolute difference value function between the pixel of pixel and target image of reference picture respectively; And said nonparametric statistics cost function and colored mean absolute difference value function carried out linear combination, with the result that will the make up coupling cost between the pixel of pixel and the target image of image as a reference;
Coupling cost between the pixel of the pixel of the reference picture that calculates and target image is assembled; And
Confirm the parallax of the pixel of reference picture based on the result that the coupling cost is assembled, to realize the stereo-picture coupling to target image.
9. method as claimed in claim 8; Wherein, The step of assembling comprises: the dynamic support region of confirming this pixel to each pixel in the reference picture; In each dynamic support region of confirming, the coupling cost between the pixel of the pixel of the reference picture that calculates and target image is assembled.
10. method as claimed in claim 8; Wherein, The step of confirming parallax comprises: make up the global energy of the coupling cost of subsidiary degree of depth flatness constraint, and the parallax level set that will make global energy obtain minimum value is confirmed as final stereo-picture coupling disparity map.
11. method as claimed in claim 10 also comprises: the stereo-picture coupling disparity map to final carries out thinning processing.
12. method as claimed in claim 11, wherein, at least one during the step that final stereo-picture coupling disparity map is carried out thinning processing may further comprise the steps:
Detect blocking a little in the disparity map, utilize the image around blocking a little to estimate to block parallax a little;
In disparity map, detect the discontinuous border that saltus step appears in parallax value, for discontinuous borderline pixel, the pixel that is utilized in said discontinuous boundaries on either side and said pixel separation certain distance to confirm again the parallax of discontinuous borderline pixel;
Disparity map is carried out other difference computing of sub-pixel-level.
13. method as claimed in claim 8; Wherein, In said linear combination, the ratio between said nonparametric statistics cost function and the colored mean absolute difference value function comes adaptively to be provided with based on local grain, partial gradient and/or the coupling confidential information of each pixel.
14. a stereo-picture matching unit, said equipment comprises:
The cost computing unit is used to calculate the coupling cost between the pixel of pixel and target image of reference picture;
The cost accumulation unit; Be used for confirming the dynamic support region of this pixel to each pixel of reference picture; In each dynamic support region of confirming, the coupling cost between the pixel of the reference picture that will be calculated by the cost computing unit and the pixel of target image is assembled; And
Parallax is confirmed the unit, is used for confirming the parallax of the pixel of reference picture to target image based on the result that the cost accumulation unit is assembled the coupling cost, to realize the stereo-picture coupling.
15. a stereo-picture matching process, said method comprises:
Coupling cost between the pixel of calculating reference picture and the pixel of target image;
Confirm the dynamic support region of this pixel to each pixel in the reference picture, in each dynamic support region of confirming, the coupling cost between the pixel of the reference picture that will be calculated by the cost computing unit and the pixel of target image is assembled; And
Confirm the parallax of the pixel of reference picture based on the result that the coupling cost is assembled, to realize the stereo-picture coupling to target image.
CN2010105025844A 2010-10-08 2010-10-08 Three-dimensional image matching method and equipment thereof Pending CN102447917A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2010105025844A CN102447917A (en) 2010-10-08 2010-10-08 Three-dimensional image matching method and equipment thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2010105025844A CN102447917A (en) 2010-10-08 2010-10-08 Three-dimensional image matching method and equipment thereof

Publications (1)

Publication Number Publication Date
CN102447917A true CN102447917A (en) 2012-05-09

Family

ID=46009941

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2010105025844A Pending CN102447917A (en) 2010-10-08 2010-10-08 Three-dimensional image matching method and equipment thereof

Country Status (1)

Country Link
CN (1) CN102447917A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102831601A (en) * 2012-07-26 2012-12-19 中北大学 Three-dimensional matching method based on union similarity measure and self-adaptive support weighting
CN102881014A (en) * 2012-09-07 2013-01-16 航天恒星科技有限公司 Quick stereo matching method based on graph cut
CN103167306A (en) * 2013-03-22 2013-06-19 上海大学 A device and method for real-time extraction of high-resolution depth maps based on image matching
CN104380338A (en) * 2012-05-22 2015-02-25 索尼电脑娱乐公司 Information processor and information processing method
WO2017067390A1 (en) * 2015-10-20 2017-04-27 努比亚技术有限公司 Method and terminal for obtaining depth information of low-texture regions in image
CN107588721A (en) * 2017-08-28 2018-01-16 武汉科技大学 The measuring method and system of a kind of more sizes of part based on binocular vision
CN107770512A (en) * 2016-08-22 2018-03-06 现代自动车株式会社 Systems and methods for generating disparity maps by matching stereoscopic images
CN108230273A (en) * 2018-01-05 2018-06-29 西南交通大学 A kind of artificial compound eye camera three dimensional image processing method based on geological information
CN109544611A (en) * 2018-11-06 2019-03-29 深圳市爱培科技术股份有限公司 A kind of binocular vision solid matching method and system based on bit feature

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100142828A1 (en) * 2008-12-10 2010-06-10 Electronics And Telecommunications Research Institute Image matching apparatus and method
CN101841730A (en) * 2010-05-28 2010-09-22 浙江大学 Real-time stereoscopic vision implementation method based on FPGA

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100142828A1 (en) * 2008-12-10 2010-06-10 Electronics And Telecommunications Research Institute Image matching apparatus and method
CN101841730A (en) * 2010-05-28 2010-09-22 浙江大学 Real-time stereoscopic vision implementation method based on FPGA

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
KRISTIAN AMBROSCH等: "Parameter Optimization of the SAD-IGMCT forStereo Vision in RGB and HSV Color Spaces", 《ELMAR, 2010 PROCEEDINGS》, 17 September 2010 (2010-09-17) *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10469829B2 (en) 2012-05-22 2019-11-05 Sony Interactive Entertainment Inc. Information processor and information processing method
CN104380338A (en) * 2012-05-22 2015-02-25 索尼电脑娱乐公司 Information processor and information processing method
CN102831601A (en) * 2012-07-26 2012-12-19 中北大学 Three-dimensional matching method based on union similarity measure and self-adaptive support weighting
CN102881014A (en) * 2012-09-07 2013-01-16 航天恒星科技有限公司 Quick stereo matching method based on graph cut
CN102881014B (en) * 2012-09-07 2015-02-11 航天恒星科技有限公司 Quick stereo matching method based on graph cut
CN103167306A (en) * 2013-03-22 2013-06-19 上海大学 A device and method for real-time extraction of high-resolution depth maps based on image matching
WO2017067390A1 (en) * 2015-10-20 2017-04-27 努比亚技术有限公司 Method and terminal for obtaining depth information of low-texture regions in image
CN107770512A (en) * 2016-08-22 2018-03-06 现代自动车株式会社 Systems and methods for generating disparity maps by matching stereoscopic images
CN107770512B (en) * 2016-08-22 2020-11-06 现代自动车株式会社 System and method for generating disparity map by matching stereo images
CN107588721A (en) * 2017-08-28 2018-01-16 武汉科技大学 The measuring method and system of a kind of more sizes of part based on binocular vision
CN108230273B (en) * 2018-01-05 2020-04-07 西南交通大学 Three-dimensional image processing method of artificial compound eye camera based on geometric information
CN108230273A (en) * 2018-01-05 2018-06-29 西南交通大学 A kind of artificial compound eye camera three dimensional image processing method based on geological information
CN109544611A (en) * 2018-11-06 2019-03-29 深圳市爱培科技术股份有限公司 A kind of binocular vision solid matching method and system based on bit feature
CN109544611B (en) * 2018-11-06 2021-05-14 深圳市爱培科技术股份有限公司 Binocular vision stereo matching method and system based on bit characteristics

Similar Documents

Publication Publication Date Title
CN102447917A (en) Three-dimensional image matching method and equipment thereof
CN108596965B (en) Light field image depth estimation method
CN102999913B (en) A kind of sectional perspective matching process based on credible propagation
CN103177451B (en) Based on the self-adapting window of image border and the Stereo Matching Algorithm of weight
US9350969B2 (en) Target region filling involving source regions, depth information, or occlusions
WO2018127007A1 (en) Depth image acquisition method and system
CN106408513B (en) Depth Map Super-Resolution Reconstruction Method
CN109146946B (en) Image non-local stereo matching method
US9380286B2 (en) Stereoscopic target region filling
CN105144234A (en) Generate a depth map for an input image using an example approximate depth map associated with an example similar image
CN106408596B (en) Edge-Based Local Stereo Matching Method
JP2014150521A (en) Method and system for increasing resolution of depth image
CN107133914B (en) Apparatus for generating a three-dimensional color image and method for generating a three-dimensional color image
CN102831601A (en) Three-dimensional matching method based on union similarity measure and self-adaptive support weighting
Shivakumar et al. Real time dense depth estimation by fusing stereo with sparse depth measurements
CN102436671A (en) Virtual viewpoint drawing method based on depth value nonlinear transformation
CN109887008B (en) Method, device and equipment for parallax stereo matching based on forward and backward smoothing and O (1) complexity
CN112991421B (en) A Stereo Matching Method for Robot Vision
CN102129567A (en) Fast stereo matching method based on color partitioning and self-adaptive window
CN104200453A (en) Parallax image correcting method based on image segmentation and credibility
US8884951B2 (en) Depth estimation data generating apparatus, depth estimation data generating method, and depth estimation data generating program, and pseudo three-dimensional image generating apparatus, pseudo three-dimensional image generating method, and pseudo three-dimensional image generating program
CN102857772B (en) Image treatment method and image processor
CN107610070B (en) Free stereo matching method based on three-camera collection
CN118015059A (en) A noise-resistant stereo matching method and system based on improved Census transformation
WO2013100789A1 (en) Method of and apparatus for local optimization texture synthesis 3-d inpainting

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
AD01 Patent right deemed abandoned

Effective date of abandoning: 20160302

C20 Patent right or utility model deemed to be abandoned or is abandoned