[go: up one dir, main page]

CN110060291B - A Calculation Method of Stereo Apparent Distance Considering Human Factors - Google Patents

A Calculation Method of Stereo Apparent Distance Considering Human Factors Download PDF

Info

Publication number
CN110060291B
CN110060291B CN201910273315.6A CN201910273315A CN110060291B CN 110060291 B CN110060291 B CN 110060291B CN 201910273315 A CN201910273315 A CN 201910273315A CN 110060291 B CN110060291 B CN 110060291B
Authority
CN
China
Prior art keywords
stereo
apparent distance
depth
objective
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910273315.6A
Other languages
Chinese (zh)
Other versions
CN110060291A (en
Inventor
权巍
赵云秀
韩成
李华
胡汉平
张超
蒋振刚
杨华民
冯欣
丁莹
姜珊
刘祎
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Changchun University of Science and Technology
Original Assignee
Changchun University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Changchun University of Science and Technology filed Critical Changchun University of Science and Technology
Priority to CN201910273315.6A priority Critical patent/CN110060291B/en
Publication of CN110060291A publication Critical patent/CN110060291A/en
Application granted granted Critical
Publication of CN110060291B publication Critical patent/CN110060291B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/593Depth or shape recovery from multiple images from stereo images
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20228Disparity calculation for image-based rendering

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

本发明涉及一种基于人因的立体视在距离解算模型,其特征在于:提取图像在屏幕上的视差,计算其立体深度;其次,提取图像的感兴趣区域,获取其视差角、宽度角与前后景对比度,根据客观视觉舒适度解算模型计算其舒适度;然后,通过主观实验,获得立体图像主观感知的视在距离值;最后,通过主观测量值,建立解算的立体深度与客观舒适度的关联关系,建立基于立体舒适度的视在距离解算模型;将模型出发点建立在人眼的基础上,使得模型更具客观性,对立体拍摄起到更好的指导作用;结合立体深度、立体视觉舒适度等,综合考量多维因素,使得模型评价结果更为准确、可信度高。

Figure 201910273315

The present invention relates to a stereo apparent distance calculation model based on human factors, which is characterized in that: extracting the parallax of the image on the screen, and calculating its stereo depth; secondly, extracting the region of interest of the image, and obtaining its parallax angle and width angle Compared with the front and back background, the comfort is calculated according to the objective visual comfort calculation model; then, through the subjective experiment, the apparent distance value of the subjective perception of the stereo image is obtained; finally, the calculated stereo depth and the objective distance are established through the subjective measurement value. The relationship between the comfort level and the apparent distance calculation model based on the three-dimensional comfort level are established; Depth, stereo vision comfort, etc., comprehensively consider multi-dimensional factors, making the model evaluation results more accurate and reliable.

Figure 201910273315

Description

一种考虑人因的立体视在距离解算方法A Calculation Method of Stereo Apparent Distance Considering Human Factors

技术领域technical field

本发明涉及一种考虑人因的立体视在距离解算方法,属于计算机视觉、图像处理技术领域。The invention relates to a method for calculating a stereoscopic apparent distance considering human factors, and belongs to the technical fields of computer vision and image processing.

背景技术Background technique

近些年,立体技术、相关硬件技术发展迅速,立体视频和游戏等借助手机、平板、电视、电影、以及各种头盔显示器等媒介全面进入普通大众生活。立体视频观看过程中的立体感是立体显示优于二维显示的最主要特征,因此必须恰当控制立体感的强弱,既要避免没有任何视觉冲击的“假立体”画面,也不能出现令观赏者不适、甚至无法实现立体融合的情况。In recent years, stereoscopic technology and related hardware technologies have developed rapidly. Stereoscopic video and games have fully entered the lives of ordinary people with the help of mobile phones, tablets, TVs, movies, and various head-mounted displays. The three-dimensional effect in the process of watching stereoscopic video is the most important feature that the three-dimensional display is superior to the two-dimensional display. Therefore, the strength of the three-dimensional effect must be properly controlled. It is necessary to avoid the "false three-dimensional" picture without any visual impact, and it must not appear to make people watch. The patient is uncomfortable, or even unable to achieve three-dimensional fusion.

基于左右眼立体图像、拍摄和播放环境参数等众多因素可求解立体深度信息。而实际上用户通过各种媒介观看到的立体效果与求解的立体深度存在差异。现阶段,在立体整个生命周期中,只有观赏者最终观看后才能根据观赏体验效果了解立体效果情况,无法实现及时的立体感控制。Stereoscopic depth information can be calculated based on many factors such as left and right eye stereoscopic images, shooting and playback environment parameters, etc. In fact, there is a difference between the three-dimensional effect viewed by the user through various media and the solved three-dimensional depth. At present, in the entire life cycle of the stereo, only the viewer can understand the stereo effect according to the viewing experience effect after watching it finally, and it is impossible to realize timely stereo control.

立体视在距离即观看者所感知到的立体信息的空间距离,不同于“立体深度”可以直接解算,视在距离的影响因素复杂多样,除“立体深度”解算模型中涉及到的因素外还包含观看者自身因素,即人因。由于立体影像在人眼视网膜上成像不同于人眼观看正常立体实物,在观看过程中会由于“调节-聚合”等问题造成视觉舒适度较低,即观看体验较差,因而会产生一定的疲劳感造成感知的视在距离与解算的立体深度值有一定的出入。因此,必须考虑人因才能更准确的解算立体视在距离。The apparent distance of the stereo is the spatial distance of the stereo information perceived by the viewer. Unlike the "stereo depth", which can be directly calculated, the factors affecting the apparent distance are complex and diverse, except for the factors involved in the "stereo depth" calculation model In addition, it also includes the viewer's own factors, that is, human factors. Since the imaging of stereoscopic images on the retina of the human eye is different from the normal stereoscopic objects viewed by the human eye, the visual comfort will be low due to problems such as "adjustment-aggregation" during the viewing process, that is, the viewing experience is poor, and a certain amount of fatigue will occur The sense of perception causes a certain discrepancy between the perceived apparent distance and the calculated stereo depth value. Therefore, human factors must be considered in order to calculate the stereo apparent distance more accurately.

发明内容Contents of the invention

本发明的目的在于提供一种基于人因的立体视在距离解算模型,其由于人类在空间感知方面,虚拟环境中人感知到的视在距离与实际深度存在显著的误差;虽然现实世界中的视在距离估计精确度约为94%,但在虚拟环境中平均下降到80%左右,即低估或压缩20%,同时感知到的视在距离受视差、颜色及亮度等因素影响。鉴于此,本发明以立体深度解算模型为基础,并以客观舒适度模型解算结果作为人因反馈,建立立体视在距离解算模型,自动完成视在距离的计算,获得了更为准确的视在距离。The object of the present invention is to provide a human-based stereoscopic apparent distance calculation model, because of the human perception of space, there is a significant error between the apparent distance and the actual depth perceived by humans in the virtual environment; although in the real world The accuracy of the apparent distance estimation is about 94%, but it drops to about 80% on average in the virtual environment, that is, it is underestimated or compressed by 20%, and the perceived apparent distance is affected by factors such as parallax, color and brightness. In view of this, the present invention is based on the three-dimensional depth calculation model, and uses the objective comfort model calculation result as human factor feedback to establish a three-dimensional apparent distance calculation model, and automatically completes the calculation of the apparent distance, obtaining a more accurate the apparent distance.

本发明技术方案是这样实现的:一种基于人因的立体视在距离解算模型,其特征在于:提取图像在屏幕上的视差,计算其立体深度;其次,提取图像的感兴趣区域,获取其视差角、宽度角与前后景对比度,根据客观视觉舒适度解算模型计算其舒适度;然后,通过主观实验,获得立体图像主观感知的视在距离值;最后,通过主观测量值,建立解算的立体深度与客观舒适度的关联关系,建立基于立体舒适度的视在距离解算模型;其具体的步骤如下:The technical scheme of the present invention is realized in the following way: a human factor-based stereoscopic apparent distance calculation model is characterized in that: extracting the parallax of the image on the screen, and calculating its stereoscopic depth; secondly, extracting the region of interest of the image, obtaining Its parallax angle, width angle, and front-background contrast are calculated according to the objective visual comfort calculation model; then, through subjective experiments, the apparent distance value of the subjective perception of the stereoscopic image is obtained; finally, the solution is established through the subjective measurement value. Based on the relationship between the calculated stereo depth and objective comfort, an apparent distance calculation model based on stereo comfort is established; the specific steps are as follows:

步骤1、提取立体图像对放映时的屏幕视差,根据下式计算其立体深度:Step 1, extract the screen parallax when the stereoscopic image pair is projected, and calculate its stereoscopic depth according to the following formula:

Figure GDA0003997294090000021
Figure GDA0003997294090000021

其中,SD表示解算的立体深度,e表示观看者的眼瞳距离(一般常取6.5cm),V表示观看者距播放屏幕的距离,Z表示播放时屏幕上立体图像对的视差;Among them, SD represents the stereoscopic depth of the solution, e represents the eye-pupil distance of the viewer (usually 6.5cm), V represents the distance between the viewer and the playback screen, and Z represents the parallax of the stereo image pair on the screen during playback;

步骤2、分别得到立体图像的前景区域视差角、宽度角及前后景区域对比度,根据下式计算其客观视觉舒适度:Step 2. Obtain the foreground area parallax angle, width angle, and foreground area contrast of the stereoscopic image respectively, and calculate its objective visual comfort according to the following formula:

VC(D,w,c)=4.8736-0.7084D+0.1912ln(w)-0.0208Dln(w)+0.0015c2-0.0572cVC(D,w,c)=4.8736-0.7084D+0.1912ln(w)-0.0208Dln(w)+0.0015c 2 -0.0572c

(0.50≤D≤2.00,0.25≤w≤4.00)(0.50≤D≤2.00, 0.25≤w≤4.00)

其中,D表示前景区域的视差角,w表示前景区域的宽度角,c表示前景区域与后景区域的对比度。要求视差角在范围0.5°~2.0°内,宽度角在范围0.25°~4.0°内;VC为客观立体视觉舒适度;Among them, D represents the disparity angle of the foreground area, w represents the width angle of the foreground area, and c represents the contrast between the foreground area and the background area. It is required that the parallax angle is within the range of 0.5°~2.0°, and the width angle is within the range of 0.25°~4.0°; VC is the objective stereoscopic vision comfort;

步骤3、对实拍的21组立体图像进行主观实验,人为地测量实验人员主观感知的视在距离,整理数据,获取数据平均值。Step 3. Conduct a subjective experiment on the 21 sets of stereoscopic images that were actually shot, artificially measure the apparent distance perceived by the experimenters subjectively, organize the data, and obtain the average value of the data.

步骤4、计算每组图像的感知视在距离与立体深度的差值dif,分析该差值与舒适度VC的关系,做曲线拟合得出线性关系如下:Step 4. Calculate the difference dif between the perceived apparent distance and the stereoscopic depth of each group of images, analyze the relationship between the difference and the comfort level VC, and perform curve fitting to obtain a linear relationship as follows:

dif=0.13-0.0353*VCdif=0.13-0.0353*VC

其中,dif为感知视在距离(GTPD)与立体深度(SD)的差值:dif=GTPD-SD,VC为客观立体视觉舒适度;Wherein, dif is the difference between perceived apparent distance (GTPD) and stereo depth (SD): dif=GTPD-SD, VC is objective stereo vision comfort;

步骤5、整理公式,最终得到视在距离、立体深度与立体舒适度的关联关系,得出视在距离的解算模型:Step 5. Arrange the formulas, and finally get the relationship between apparent distance, stereoscopic depth and stereoscopic comfort, and obtain the calculation model of apparent distance:

OPD=SD+0.13-0.0353*VCOPD=SD+0.13-0.0353*VC

式中,OPD表示模型计算的客观视在距离,SD表示模型解算的立体深度,VC表示客观立体视觉舒适度模型解算的舒适度值。In the formula, OPD represents the objective apparent distance calculated by the model, SD represents the stereo depth calculated by the model, and VC represents the comfort value calculated by the objective stereo vision comfort model.

本发明的积极效果是实现了对立体图像视在距离的客观估计,通过分析立体深度解算模型的计算结果、客观视觉舒适度模型解算结果与主观感知的视在距离的关联关系,建立客观立体视在距离的解算模型。能够有效地节约传统人工评价的人力成本,和依赖于评价者的问题;将模型出发点建立在人眼的基础上,使得模型更具客观性,对立体拍摄起到更好的指导作用;结合立体深度、立体视觉舒适度等,综合考量多维因素,使得模型评价结果更为准确、可信度高。The positive effect of the present invention is to realize the objective estimation of the apparent distance of the stereoscopic image, and establish the objective estimation by analyzing the calculation result of the stereoscopic depth calculation model, the calculation result of the objective visual comfort model and the subjectively perceived apparent distance. The solution model for stereo apparent distance. It can effectively save the labor cost of traditional manual evaluation and the problem of relying on the evaluator; the starting point of the model is based on the human eye, which makes the model more objective and plays a better guiding role in stereo shooting; combined with stereo Depth, stereo vision comfort, etc., comprehensively consider multi-dimensional factors, making the model evaluation results more accurate and reliable.

附图说明Description of drawings

图1为本发明流程示意图。Fig. 1 is a schematic flow chart of the present invention.

图2为双目相机系统。Figure 2 is a binocular camera system.

图3为实验组右视图。Figure 3 is the right view of the experimental group.

图4为实验环境。Figure 4 is the experimental environment.

图5为实验组感兴趣区域图。Figure 5 is a map of the region of interest in the experimental group.

图6为立体深度与视在距离的趋势图。Fig. 6 is a trend graph of stereo depth and apparent distance.

图7为视在距离与立体深度的差值与立体舒适度的关系。Fig. 7 shows the relationship between the difference between the apparent distance and the three-dimensional depth and the three-dimensional comfort.

图8为感知的视在距离、立体深度与解算的视在距离三者趋势图。FIG. 8 is a trend diagram of perceived apparent distance, stereoscopic depth and calculated apparent distance.

具体实施方式Detailed ways

下面结合附图和实施例对本发明做进一步的描述:本实施例中,采用大恒工业相机获取21组实拍立体图像对,并使用韩国先进科学院提供的立体图像库对该模型进行验证,流程图如图1所示,具体步骤如下:The present invention will be further described below in conjunction with the accompanying drawings and embodiments: In this embodiment, 21 groups of real-shot stereo image pairs are obtained by using Daheng industrial cameras, and the model is verified using the stereo image library provided by the Korean Academy of Advanced Sciences. The process flow As shown in Figure 1, the specific steps are as follows:

步骤1、使用两个大恒MER-310-12UC工业相机(如图2所示)进行图像采集,相机的分辨率均为为2048×1536。且两台型号相同的相机,其白平衡、增益、模式等基础参数标准一致。在进行数据采集时,首先连接计算机传送信号,将设备置于平稳的水平三角支架,组装后对准校正,通过大恒相机配备的软件在显示器上观看双机画面。将两台机器的光圈、焦距、焦点等随机信息同步一致,以人眼瞳孔距(5.5-7.5cm,本实验选取6.5cm)为拍摄基线进行同步拍摄。本实验在立体图像拍摄过程中,场景布局多分为两层,画面较简洁,通过层次对比凸显出较好的立体效果。如图3所示,为采集到的21组图像的右视图。Step 1. Use two Daheng MER-310-12UC industrial cameras (as shown in Figure 2) for image acquisition, and the resolution of the cameras is 2048×1536. And two cameras with the same model have the same basic parameter standards such as white balance, gain, and mode. When collecting data, first connect the computer to transmit the signal, place the device on a stable horizontal tripod stand, align and correct after assembly, and watch the dual-machine picture on the monitor through the software equipped with Daheng Camera. Synchronize random information such as the aperture, focal length, and focus of the two machines, and use the interpupillary distance of the human eye (5.5-7.5cm, 6.5cm is selected in this experiment) as the shooting baseline for synchronous shooting. In the process of stereo image shooting in this experiment, the scene layout is mostly divided into two layers, the picture is relatively simple, and the better stereo effect is highlighted through layer comparison. As shown in Figure 3, it is the right view of the 21 groups of images collected.

步骤2、选取年龄在19-28岁之间的实验人员15名。确定实验人员的年龄在合理范围,视力正常且无视觉方面的生理疾病,如色盲、色弱等眼部疾病。且在观看过程中佩戴时分式立体眼镜无不良反应,在观看时能准确且稳定的判断出感知到的立体影像位置并确定感知的视在距离。并根据其是否有一定的观测立体影像的经验、是否了解相关专业知识进行区分。由于有相关经验的被测人员能更好的进行主观评测,而其他观测人员由于缺少相关经验,需要先安排一系列的培训,通过前期的训练,能更准确的确定主观评判结果。Step 2. Select 15 experimenters aged between 19 and 28. Make sure that the age of the experimenter is within a reasonable range, with normal eyesight and no visual physical diseases, such as color blindness, color weakness and other eye diseases. In addition, there is no adverse reaction when wearing the time-division stereoscopic glasses during viewing, and the perceived position of the stereoscopic image can be accurately and stably judged and the perceived apparent distance can be determined during viewing. They are also distinguished according to whether they have certain experience in observing stereoscopic images and whether they understand relevant professional knowledge. Since the tested personnel with relevant experience can better perform subjective evaluation, while other observers lack relevant experience, they need to arrange a series of training first. Through the early training, the subjective evaluation results can be determined more accurately.

步骤3、使用英伟达NVIDIA Quadro K620显卡、三星2233RZ 3D显示屏配合英伟达无线立体幻镜--时分式立体眼镜进行观看立体图像。观看距离为1.5米,约为显示器高度5倍,水平(50cm)和垂直(30cm)视角分别为18.96°和8.58°。实验环境基于ITU-R BT.500-11和ITU-R BT.1438推荐设置,为保证观看过程中观看人员的眼部舒适,在进行实验时每两对图像播放间隔可进行眼部休息。在实验观看过程中仿照影院观影环境,将环境光调暗,如图4所示,为暗室下受测人员进行实验。Step 3. Use NVIDIA Quadro K620 graphics card, Samsung 2233RZ 3D display screen and NVIDIA Wireless 3D Vision - time-division stereoscopic glasses to watch stereoscopic images. The viewing distance is 1.5 meters, about 5 times the height of the display, and the horizontal (50cm) and vertical (30cm) viewing angles are 18.96° and 8.58°, respectively. The experimental environment is based on the settings recommended by ITU-R BT.500-11 and ITU-R BT.1438. In order to ensure the eye comfort of the viewers during the viewing process, eye rest can be performed between every two pairs of image playback during the experiment. In the process of watching the experiment, the ambient light is dimmed like the movie viewing environment of the theater, as shown in Figure 4, and the experiment is carried out for the testees in the dark room.

步骤4、为获取感知的视在距离,使用主观评测的方法进行实验,在相关研究中,测量方法主要有3种类型:言语估计,视觉想象动作和感知匹配。词汇估计是指以米为单位直接表示深度估计。视觉想象的动作是指让主体观察物体并想象走向物体。想象时间以他们通常的行走速度记录以进行深度判断。本实施例中实验采用语言估计的方法,直接量化估计感知的视在距离。通过人为的测量来获取观看人员所感知到的各图像的视在距离,将获得到的视在距离取其平均值,整理数据进而得到表1所示的部分实验组图像对的视在距离值。Step 4. In order to obtain the perceived apparent distance, experiments are carried out using subjective evaluation methods. In related research, there are mainly three types of measurement methods: verbal estimation, visual imagination and perception matching. Lexical estimates refer to the direct representation of depth estimates in meters. The act of visual imagination involves having the subject observe an object and imagine walking towards it. Imaginary time is recorded at their usual walking pace for depth judgment. In this embodiment, the method of language estimation is used in the experiment to directly quantify and estimate the perceived apparent distance. Obtain the apparent distance of each image perceived by the viewer through artificial measurement, take the average of the obtained apparent distance, organize the data and obtain the apparent distance of some experimental group image pairs shown in Table 1 value.

表1部分图像对的视在距离Table 1 Apparent distance of some image pairs

Figure GDA0003997294090000041
Figure GDA0003997294090000041

步骤5、通过放映时的视差、观看者瞳距及观看距离等基本参数计算出实验组各图像对所对应的立体深度。根据下式可以解算出影像在播放时某一帧画面的立体深度值:Step 5. Calculate the stereoscopic depth corresponding to each image pair in the experimental group through basic parameters such as parallax during projection, viewer interpupillary distance, and viewing distance. According to the following formula, the stereoscopic depth value of a certain frame of the image during playback can be calculated:

Figure GDA0003997294090000051
Figure GDA0003997294090000051

其中,SD表示解算的立体深度,e表示观看者的眼瞳距离(一般常取6.5cm),V表示观看者距播放屏幕的距离,Z表示播放时屏幕上立体图像对的视差。如表2所示,为实验组图像对(图3所示)所对应的部分立体深度解算值。Among them, SD represents the calculated stereo depth, e represents the eye-pupil distance of the viewer (generally 6.5 cm), V represents the distance between the viewer and the playback screen, and Z represents the parallax of the stereo image pair on the screen during playback. As shown in Table 2, it is the partial stereo depth calculation value corresponding to the image pair of the experimental group (shown in Figure 3).

表2部分图像对的立体深度Table 2 Stereo Depth of Some Image Pairs

Figure GDA0003997294090000052
Figure GDA0003997294090000052

步骤6、获取实验组图像的视差图,并根据GBVS算法计算出各图像的平面显著图。然后,根据下式计算并融合视差图与平面显著图得到立体显著图:Step 6. Obtain the disparity map of the images in the experimental group, and calculate the plane saliency map of each image according to the GBVS algorithm. Then, calculate and fuse the disparity map and the plane saliency map according to the following formula to obtain the stereo saliency map:

IIS(x,y)=ω1ISR(x,y)+ω2DR(x,y)IIS(x,y)=ω 1 IS R (x,y)+ω 2 D R (x,y)

式中ISR(x,y)为平面显著图,DR(x,y)为视差图,ω1、ω2为其权重,本实施例取ω1=ω2=0.5。为得到感兴趣区域,将立体显著图进行阈值分割处理得到掩膜图像IIM(x,y),具体方法如下式:In the formula, IS R (x, y) is a plane saliency map, D R (x, y) is a disparity map, and ω 1 and ω 2 are their weights. In this embodiment, ω 12 =0.5. In order to obtain the region of interest, the stereo saliency map is subjected to threshold segmentation processing to obtain the mask image IIM(x, y). The specific method is as follows:

Figure GDA0003997294090000053
Figure GDA0003997294090000053

式中C(x,y)为(x,y)处的像素值,T为其分割阈值。若C(x,y)>T则该像素属于感兴趣区域,对应掩膜图像中白色区域,否则属于黑色区域。使用预置的感兴趣区掩膜与右视图、视差图相乘,分别得到感兴趣区图像和感兴趣区视差图,将感兴趣区域作为前景区域。对感兴趣区掩膜图像取反,并与右视图相乘可得背景区域。如图5所示,为实验组图像(图3)对应的感兴趣区域图。In the formula, C(x, y) is the pixel value at (x, y), and T is its segmentation threshold. If C(x,y)>T, the pixel belongs to the region of interest, corresponding to the white region in the mask image, otherwise it belongs to the black region. Use the preset ROI mask to multiply the right view and the disparity map to obtain the ROI image and the ROI disparity map respectively, and use the ROI as the foreground area. Invert the mask image of the ROI and multiply it with the right view to obtain the background area. As shown in FIG. 5 , it is a map of the region of interest corresponding to the image of the experimental group ( FIG. 3 ).

步骤7、根据下式可计算实验组图像对应的客观视觉舒适度:Step 7. According to the following formula, the objective visual comfort corresponding to the images of the experimental group can be calculated:

VC(D,w,c)=4.8736-0.7084D+0.1912ln(w)-0.0208Dln(w)+0.0015c2-0.0572cVC(D,w,c)=4.8736-0.7084D+0.1912ln(w)-0.0208Dln(w)+0.0015c 2 -0.0572c

(0.50≤D≤2.00,0.25≤w≤4.00)(0.50≤D≤2.00, 0.25≤w≤4.00)

其中,D表示前景区域的视差角,w表示前景区域的宽度角,c表示前景区域与后景区域的对比度。要求视差角在范围0.5°~2.0°内,宽度角在范围0.25°~4.0°内。D、w、c的计算如下式:Among them, D represents the disparity angle of the foreground area, w represents the width angle of the foreground area, and c represents the contrast between the foreground area and the background area. It is required that the parallax angle is within the range of 0.5°-2.0°, and the width angle is within the range of 0.25°-4.0°. D, w, c are calculated as follows:

Figure GDA0003997294090000061
Figure GDA0003997294090000061

其中,前景区域的平均视差值为Df,of表示前景区域,|of|表示of区域内像素总数,D为前景区域的平均视差角,k为放映放大率,D表示视差,L为观看时距屏幕距离。Among them, the average parallax value of the foreground area is D f , o f represents the foreground area, |o f | represents the total number of pixels in the area of o f , D is the average parallax angle of the foreground area, k is the projection magnification, D represents the parallax, L is the distance from the screen when viewing.

Figure GDA0003997294090000062
Figure GDA0003997294090000062

式中Nf表示前景区域内水平线的数量,

Figure GDA0003997294090000063
表示前景区域内第n条水平线长度。w为宽度角,W表示平均宽度,k为放映放大率,L为观看时距屏幕距离。where N f represents the number of horizontal lines in the foreground area,
Figure GDA0003997294090000063
Indicates the length of the nth horizontal line in the foreground area. w is the width angle, W is the average width, k is the projection magnification, and L is the distance from the screen when viewing.

Figure GDA0003997294090000064
Figure GDA0003997294090000064

Figure GDA0003997294090000065
Figure GDA0003997294090000065

将每个颜色通道量化为16个不同值,将颜色的数量减少了4096倍。然后将RGB空间装换成Lab空间内,获取前后景区域间的颜色距离集合。dr(r1,r2)为区域r1和r2之间的颜色距离。式中f(ck,i)表示第i个颜色ck,i在第k个区域rk的所有颜色nk中出现的频率。d(c1,i,c2,j)表示区域1中第i个颜色与区域2中第j个颜色在Lab空间内的颜色距离。

Figure GDA0003997294090000066
为基于空间加权的区域对比度(即客观舒适度模型中的c),Sr(rk,ri)是区域rk与ri之间的空间距离,σs控制空间权值的强度。σs越大空间权值的影响越小,则背景区域的影响越显著。两区域间的空间距离被定义为区域重心间的欧几里得距离。其中,像素坐标为归一化后结果,取
Figure GDA0003997294090000067
Quantize each color channel to 16 different values, reducing the number of colors by a factor of 4096. Then the RGB space is replaced with the Lab space to obtain the color distance set between the foreground and foreground regions. d r (r 1 , r 2 ) is the color distance between regions r 1 and r 2 . where f(c k,i ) represents the frequency of the i-th color c k,i appearing in all colors nk in the k-th region r k . d(c 1,i ,c 2,j ) represents the color distance between the i-th color in area 1 and the j-th color in area 2 in Lab space.
Figure GDA0003997294090000066
is the regional contrast based on spatial weighting (ie c in the objective comfort model), S r ( rk , ri ) is the spatial distance between region r k and r i , and σ s controls the strength of the spatial weight. The larger the σ s, the smaller the influence of the spatial weight, and the more significant the influence of the background area. The spatial distance between two regions is defined as the Euclidean distance between the centers of gravity of the regions. Among them, the pixel coordinates are the result after normalization, take
Figure GDA0003997294090000067

如表3所示,表中为部分实验图像所对应的视觉舒适度(VC)值。As shown in Table 3, the table shows the visual comfort (VC) values corresponding to some experimental images.

表3各图像的视觉舒适度值Table 3 Visual comfort value of each image

Figure GDA0003997294090000071
Figure GDA0003997294090000071

步骤8、为分析立体深度与视在距离关系,所示的立体深度与视在距离趋势图,从图中可观察得出:人眼感知的视在距离很大程度上受立体深度影响,但其与模型解算得到的立体深度在数值上会存在一定的出入,因此将人因这一因素纳入到视在距离计算的考量范围。而由于立体视觉舒适度直接反应观看者对立体影像的观看感受,同时也是评价立体影像质量的重要准则之一,因此本实施例将立体舒适度作为人因控制环节的反馈因素纳入到视在距离量化的过程中。根据立体深度解算模型与立体舒适度客观评价模型计算得到各图像对的立体深度(VD)及立体视觉舒适度(VC),结合主观评测得到的视在距离(GTPD)的值进行分析。Step 8. In order to analyze the relationship between the stereoscopic depth and the apparent distance, the trend graph of the stereoscopic depth and the apparent distance shown in the figure can be observed: the apparent distance perceived by the human eye is largely affected by the stereoscopic depth, but There will be a certain numerical discrepancy between it and the stereoscopic depth obtained by the model solution, so the human factor is taken into consideration in the calculation of the apparent distance. Since the stereoscopic comfort directly reflects the viewer's viewing experience of the stereoscopic image, and is also one of the important criteria for evaluating the quality of the stereoscopic image, the present embodiment incorporates the stereoscopic comfort into the apparent distance as a feedback factor of the human factor control link. in the process of quantification. According to the stereo depth calculation model and the stereo comfort objective evaluation model, the stereo depth (VD) and stereo visual comfort (VC) of each image pair are calculated, combined with the apparent distance (GTPD) value obtained from the subjective evaluation for analysis.

步骤9、为获取立体深度、立体视觉舒适度与视在距离三者关联关系,首先计算立体深度与视在距离的差异(dif=GTPD-SD)。如图7所示,随着视觉舒适度的增加,人眼实际感知的视在距离与解算的立体深度间的差值在不断减小,可进一步分析出立体舒适度与差异间的关系,通过建立散点图并作拟合曲线得到如下式:Step 9. In order to obtain the relationship among the stereoscopic depth, stereoscopic comfort and apparent distance, first calculate the difference between the stereoscopic depth and the apparent distance (dif=GTPD-SD). As shown in Figure 7, with the increase of visual comfort, the difference between the apparent distance actually perceived by the human eye and the calculated stereo depth is continuously decreasing, and the relationship between the stereo comfort and the difference can be further analyzed. By establishing a scatter plot and making a fitting curve, the following formula is obtained:

dif=0.13-0.0353*VCdif=0.13-0.0353*VC

步骤10、整理公式获得立体视在距离解算模型如下式所示:Step 10, arrange the formula to obtain the stereoscopic apparent distance calculation model as shown in the following formula:

OPD=SD+0.13-0.0353*VCOPD=SD+0.13-0.0353*VC

步骤11、在本实施例中,利用五个常用客观参量作为评价指标对模型估计值与主观评价值的相关性做分析,利用本发明模型计算这些图像的视在距离,然后比较对应的客观感知视在距离、对应的立体深度与主观感知的视在深度测量值间的Pearson相关系数(Pearson Linear Correlation Coefficient,PLCC),Kendall相关系数(Kendall Rank-order Correlation Coefficient,KRCC),SROCC相关系数(SpearmanRank OrderCorrelation Coefficient),平均绝对值误差(Mean Absolute Error,MAE)和均方根误差(Root Mean Squared Error,RMSE)。并在IVY图像库中选取部分视觉舒适度好与视觉舒适度不好的图像作为测试集,比较对应的客观感知视在距离、对应的立体深度与主观感知的视在深度测量值间的Pearson相关系数,Kendall相关系数,SROCC相关系数,平均绝对值误差和均方根误差,结果验证本发明方法得到的立体图像的视在距离解算值较立体深度解算模型结果更接近人眼感知的视在距离,性能较好。Step 11. In this embodiment, use five commonly used objective parameters as evaluation indicators to analyze the correlation between the model estimated value and the subjective evaluation value, use the model of the present invention to calculate the apparent distance of these images, and then compare the corresponding objective perception Pearson Linear Correlation Coefficient (PLCC), Kendall Rank-order Correlation Coefficient (KRCC), SROCC Correlation Coefficient (SpearmanRank OrderCorrelation Coefficient), mean absolute error (Mean Absolute Error, MAE) and root mean square error (Root Mean Squared Error, RMSE). And select some images with good visual comfort and poor visual comfort in the IVY image library as the test set, and compare the corresponding objective perceived apparent distance, the corresponding stereo depth and the Pearson correlation between the subjective perceived apparent depth measurement value coefficient, Kendall correlation coefficient, SROCC correlation coefficient, average absolute value error and root mean square error, the results verify that the apparent distance calculation value of the stereoscopic image obtained by the method of the present invention is closer to the visual perception of the human eye than the stereo depth calculation model result. At distance, the performance is better.

Claims (1)

1. A three-dimensional apparent distance resolving method based on human factors is characterized by comprising the following steps: extracting the parallax of the image on a screen, and calculating the stereo depth of the image; secondly, extracting an interested area of the image, acquiring a parallax angle and a width angle of a foreground area of the interested area and a contrast ratio of the foreground area and a background area, and calculating the comfort level of the interested area according to an objective stereoscopic vision comfort level calculation model; then, obtaining an apparent distance value of the subjective perception of the three-dimensional image through a subjective experiment; finally, establishing a correlation between the resolved stereo depth and the objective stereo vision comfort level through the subjective perception of the apparent distance value, and establishing an apparent distance resolving model based on the objective stereo vision comfort level; the method comprises the following specific steps:
step 1, extracting the screen parallax of the stereo image pair during projection, and calculating the stereo depth according to the following formula:
Figure 957104DEST_PATH_IMAGE001
wherein,SDa calculated stereo depth is represented by the depth of the stereo,eindicating the distance of the viewer's eye pupil, taken at 6.5cm,Vindicating the distance of the viewer from the playback screen,Zrepresenting the disparity of the on-screen stereoscopic image pair when played;
step 2, respectively obtaining a foreground region parallax angle and a width angle of the stereo image and the contrast of the foreground region and a background region, and calculating the objective stereo vision comfort level according to the following formula:
Figure 343086DEST_PATH_IMAGE003
wherein,Dthe parallax angle of the foreground region is represented,wthe width angle of the foreground region is represented,crepresenting the contrast of the foreground region and the background region; requiring the parallax angle of the foreground region to be in range
Figure 857244DEST_PATH_IMAGE004
Inner, the width angle of the foreground region is in the range
Figure 986874DEST_PATH_IMAGE005
Internal;VCcomfort level for objective stereoscopic vision;
step 3, carrying out subjective experiments on the 21 groups of real-shot stereo images, artificially measuring the apparent distance subjectively perceived by experimenters, sorting data, and obtaining a data average value;
step 4, calculating the difference value between the subjective perception apparent distance and the three-dimensional depth of each group of imagesdifAnalyzing the difference and the objective stereoscopic vision comfort levelVCFitting a curve to obtain a linear relation as follows:
Figure 535667DEST_PATH_IMAGE006
wherein,
Figure 76239DEST_PATH_IMAGE007
GTPD isThe apparent distance of the subjective perception is,SD isThe depth of the three-dimensional space is increased,VCcomfort level for objective stereoscopic vision;
step 5, a formula is arranged, the incidence relation of the objective apparent distance, the three-dimensional depth and the objective three-dimensional visual comfort level is finally obtained, and a resolving model of the apparent distance is obtained:
Figure 343272DEST_PATH_IMAGE008
in the formula,OPDrepresenting the objective apparent distance calculated by the model,SDthe depth of the solid of the model solution is represented,VCrepresenting the comfort value resolved by the objective stereoscopic vision comfort model.
CN201910273315.6A 2019-04-04 2019-04-04 A Calculation Method of Stereo Apparent Distance Considering Human Factors Active CN110060291B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910273315.6A CN110060291B (en) 2019-04-04 2019-04-04 A Calculation Method of Stereo Apparent Distance Considering Human Factors

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910273315.6A CN110060291B (en) 2019-04-04 2019-04-04 A Calculation Method of Stereo Apparent Distance Considering Human Factors

Publications (2)

Publication Number Publication Date
CN110060291A CN110060291A (en) 2019-07-26
CN110060291B true CN110060291B (en) 2023-01-31

Family

ID=67318414

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910273315.6A Active CN110060291B (en) 2019-04-04 2019-04-04 A Calculation Method of Stereo Apparent Distance Considering Human Factors

Country Status (1)

Country Link
CN (1) CN110060291B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112153362B (en) * 2020-09-15 2022-03-08 清华大学深圳国际研究生院 Method and system for measuring stereoscopic depth of naked eye 3D display system

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102402005A (en) * 2011-12-06 2012-04-04 北京理工大学 Bifocal-surface monocular stereo helmet-mounted display device with free-form surfaces
CN103118265A (en) * 2011-11-16 2013-05-22 克里斯蒂数字系统美国有限公司 A collimated stereo display system
CN103986925A (en) * 2014-06-05 2014-08-13 吉林大学 Evaluation Method of Visual Comfort of Stereo Video Based on Brightness Compensation
GB201419379D0 (en) * 2014-10-31 2014-12-17 Nokia Corp Method for alignment of low-quality noisy depth map to the high-resolution colour image
CN104887316A (en) * 2015-04-24 2015-09-09 长春理工大学 Virtual three-dimensional endoscope displaying method based on active three-dimensional displaying technology
CN106570900A (en) * 2016-10-11 2017-04-19 宁波大学 Three-dimensional image relocation method
CN109167988A (en) * 2018-08-29 2019-01-08 长春理工大学 A kind of stereo image vision comfort level evaluation method based on D+W model and contrast

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110222757A1 (en) * 2010-03-10 2011-09-15 Gbo 3D Technology Pte. Ltd. Systems and methods for 2D image and spatial data capture for 3D stereo imaging
CN103595990B (en) * 2013-10-30 2015-05-20 清华大学 Method for obtaining binocular stereoscopic video comfort level of motion perception
JP2015154101A (en) * 2014-02-10 2015-08-24 ソニー株式会社 Image processing method, image processor and electronic apparatus

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103118265A (en) * 2011-11-16 2013-05-22 克里斯蒂数字系统美国有限公司 A collimated stereo display system
CN102402005A (en) * 2011-12-06 2012-04-04 北京理工大学 Bifocal-surface monocular stereo helmet-mounted display device with free-form surfaces
CN103986925A (en) * 2014-06-05 2014-08-13 吉林大学 Evaluation Method of Visual Comfort of Stereo Video Based on Brightness Compensation
GB201419379D0 (en) * 2014-10-31 2014-12-17 Nokia Corp Method for alignment of low-quality noisy depth map to the high-resolution colour image
CN104887316A (en) * 2015-04-24 2015-09-09 长春理工大学 Virtual three-dimensional endoscope displaying method based on active three-dimensional displaying technology
CN106570900A (en) * 2016-10-11 2017-04-19 宁波大学 Three-dimensional image relocation method
CN109167988A (en) * 2018-08-29 2019-01-08 长春理工大学 A kind of stereo image vision comfort level evaluation method based on D+W model and contrast

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Measurement of parallax distribution, and its application to the analysis of visual comfort for stereoscopic HDTV;Yuji Nojiri 等;《Stereoscopic Displays and Virtual Reality Systems X》;20030531;195-205 *
多视点虚拟场景的立体效果控制算法与应用;张玉强;《中国优秀硕士学位论文全文数据库 信息科技辑》;20180315;I138-1990 *

Also Published As

Publication number Publication date
CN110060291A (en) 2019-07-26

Similar Documents

Publication Publication Date Title
Mittal et al. Algorithmic assessment of 3D quality of experience for images and videos
CN103986925B (en) based on the stereoscopic video visual comfort evaluation method of luminance compensation
CN104853185A (en) Stereo video comfort evaluation method combining multiple parallaxes with motion
CN207589060U (en) A naked-eye three-dimensional display device combined with visual fatigue detection
Choi et al. Visual fatigue modeling and analysis for stereoscopic video
WO2015096346A1 (en) Test method and test device for 3d display crosstalk
CN103780895A (en) Stereoscopic video quality evaluation method
Chen et al. Study of subject agreement on stereoscopic video quality
CN104581141A (en) Three-dimensional picture visual comfort evaluation method
WO2019041352A1 (en) Panum's area measurement method and apparatus, and wearable display device
Opozda et al. The survey of subjective and objective methods for quality assessment of 2D and 3D images
CN110139095B (en) Naked eye 3D display module detection method and system and readable storage medium
CN110060291B (en) A Calculation Method of Stereo Apparent Distance Considering Human Factors
CN106791802B (en) A kind of the stereoscopic display visual comfort detection method and system of difference viewing angle and parallax
Hanhart et al. Subjective evaluation of two stereoscopic imaging systems exploiting visual attention to improve 3D quality of experience
CN108259888A (en) The test method and system of stereo display effect
CN105915894A (en) Naked eye stereotelevision 3D crosstalk value measuring method
CN107018398B (en) A method of for the quantization calibration of light field Three-dimensional Display
Cho et al. The measurement of eyestrain caused from diverse binocular disparities, viewing time and display sizes in watching stereoscopic 3D content
Kim et al. Perceptual crosstalk prediction on autostereoscopic 3D display
CN206650798U (en) The test system of stereo display effect
CN105430370B (en) Stereoscopic image visual comfort evaluation method based on ranking learning
CN206341331U (en) Detection device and stereoscopic display device tracking effect evaluation system
CN106331677B (en) The evaluation method and system of bore hole stereoscopic display intermediate-resolution index
Barkowsky et al. Is visual fatigue changing the perceived depth accuracy on an autostereoscopic display?

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant