CN110060291B - A Calculation Method of Stereo Apparent Distance Considering Human Factors - Google Patents
A Calculation Method of Stereo Apparent Distance Considering Human Factors Download PDFInfo
- Publication number
- CN110060291B CN110060291B CN201910273315.6A CN201910273315A CN110060291B CN 110060291 B CN110060291 B CN 110060291B CN 201910273315 A CN201910273315 A CN 201910273315A CN 110060291 B CN110060291 B CN 110060291B
- Authority
- CN
- China
- Prior art keywords
- stereo
- apparent distance
- depth
- objective
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20228—Disparity calculation for image-based rendering
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
本发明涉及一种基于人因的立体视在距离解算模型,其特征在于:提取图像在屏幕上的视差,计算其立体深度;其次,提取图像的感兴趣区域,获取其视差角、宽度角与前后景对比度,根据客观视觉舒适度解算模型计算其舒适度;然后,通过主观实验,获得立体图像主观感知的视在距离值;最后,通过主观测量值,建立解算的立体深度与客观舒适度的关联关系,建立基于立体舒适度的视在距离解算模型;将模型出发点建立在人眼的基础上,使得模型更具客观性,对立体拍摄起到更好的指导作用;结合立体深度、立体视觉舒适度等,综合考量多维因素,使得模型评价结果更为准确、可信度高。
The present invention relates to a stereo apparent distance calculation model based on human factors, which is characterized in that: extracting the parallax of the image on the screen, and calculating its stereo depth; secondly, extracting the region of interest of the image, and obtaining its parallax angle and width angle Compared with the front and back background, the comfort is calculated according to the objective visual comfort calculation model; then, through the subjective experiment, the apparent distance value of the subjective perception of the stereo image is obtained; finally, the calculated stereo depth and the objective distance are established through the subjective measurement value. The relationship between the comfort level and the apparent distance calculation model based on the three-dimensional comfort level are established; Depth, stereo vision comfort, etc., comprehensively consider multi-dimensional factors, making the model evaluation results more accurate and reliable.
Description
技术领域technical field
本发明涉及一种考虑人因的立体视在距离解算方法,属于计算机视觉、图像处理技术领域。The invention relates to a method for calculating a stereoscopic apparent distance considering human factors, and belongs to the technical fields of computer vision and image processing.
背景技术Background technique
近些年,立体技术、相关硬件技术发展迅速,立体视频和游戏等借助手机、平板、电视、电影、以及各种头盔显示器等媒介全面进入普通大众生活。立体视频观看过程中的立体感是立体显示优于二维显示的最主要特征,因此必须恰当控制立体感的强弱,既要避免没有任何视觉冲击的“假立体”画面,也不能出现令观赏者不适、甚至无法实现立体融合的情况。In recent years, stereoscopic technology and related hardware technologies have developed rapidly. Stereoscopic video and games have fully entered the lives of ordinary people with the help of mobile phones, tablets, TVs, movies, and various head-mounted displays. The three-dimensional effect in the process of watching stereoscopic video is the most important feature that the three-dimensional display is superior to the two-dimensional display. Therefore, the strength of the three-dimensional effect must be properly controlled. It is necessary to avoid the "false three-dimensional" picture without any visual impact, and it must not appear to make people watch. The patient is uncomfortable, or even unable to achieve three-dimensional fusion.
基于左右眼立体图像、拍摄和播放环境参数等众多因素可求解立体深度信息。而实际上用户通过各种媒介观看到的立体效果与求解的立体深度存在差异。现阶段,在立体整个生命周期中,只有观赏者最终观看后才能根据观赏体验效果了解立体效果情况,无法实现及时的立体感控制。Stereoscopic depth information can be calculated based on many factors such as left and right eye stereoscopic images, shooting and playback environment parameters, etc. In fact, there is a difference between the three-dimensional effect viewed by the user through various media and the solved three-dimensional depth. At present, in the entire life cycle of the stereo, only the viewer can understand the stereo effect according to the viewing experience effect after watching it finally, and it is impossible to realize timely stereo control.
立体视在距离即观看者所感知到的立体信息的空间距离,不同于“立体深度”可以直接解算,视在距离的影响因素复杂多样,除“立体深度”解算模型中涉及到的因素外还包含观看者自身因素,即人因。由于立体影像在人眼视网膜上成像不同于人眼观看正常立体实物,在观看过程中会由于“调节-聚合”等问题造成视觉舒适度较低,即观看体验较差,因而会产生一定的疲劳感造成感知的视在距离与解算的立体深度值有一定的出入。因此,必须考虑人因才能更准确的解算立体视在距离。The apparent distance of the stereo is the spatial distance of the stereo information perceived by the viewer. Unlike the "stereo depth", which can be directly calculated, the factors affecting the apparent distance are complex and diverse, except for the factors involved in the "stereo depth" calculation model In addition, it also includes the viewer's own factors, that is, human factors. Since the imaging of stereoscopic images on the retina of the human eye is different from the normal stereoscopic objects viewed by the human eye, the visual comfort will be low due to problems such as "adjustment-aggregation" during the viewing process, that is, the viewing experience is poor, and a certain amount of fatigue will occur The sense of perception causes a certain discrepancy between the perceived apparent distance and the calculated stereo depth value. Therefore, human factors must be considered in order to calculate the stereo apparent distance more accurately.
发明内容Contents of the invention
本发明的目的在于提供一种基于人因的立体视在距离解算模型,其由于人类在空间感知方面,虚拟环境中人感知到的视在距离与实际深度存在显著的误差;虽然现实世界中的视在距离估计精确度约为94%,但在虚拟环境中平均下降到80%左右,即低估或压缩20%,同时感知到的视在距离受视差、颜色及亮度等因素影响。鉴于此,本发明以立体深度解算模型为基础,并以客观舒适度模型解算结果作为人因反馈,建立立体视在距离解算模型,自动完成视在距离的计算,获得了更为准确的视在距离。The object of the present invention is to provide a human-based stereoscopic apparent distance calculation model, because of the human perception of space, there is a significant error between the apparent distance and the actual depth perceived by humans in the virtual environment; although in the real world The accuracy of the apparent distance estimation is about 94%, but it drops to about 80% on average in the virtual environment, that is, it is underestimated or compressed by 20%, and the perceived apparent distance is affected by factors such as parallax, color and brightness. In view of this, the present invention is based on the three-dimensional depth calculation model, and uses the objective comfort model calculation result as human factor feedback to establish a three-dimensional apparent distance calculation model, and automatically completes the calculation of the apparent distance, obtaining a more accurate the apparent distance.
本发明技术方案是这样实现的:一种基于人因的立体视在距离解算模型,其特征在于:提取图像在屏幕上的视差,计算其立体深度;其次,提取图像的感兴趣区域,获取其视差角、宽度角与前后景对比度,根据客观视觉舒适度解算模型计算其舒适度;然后,通过主观实验,获得立体图像主观感知的视在距离值;最后,通过主观测量值,建立解算的立体深度与客观舒适度的关联关系,建立基于立体舒适度的视在距离解算模型;其具体的步骤如下:The technical scheme of the present invention is realized in the following way: a human factor-based stereoscopic apparent distance calculation model is characterized in that: extracting the parallax of the image on the screen, and calculating its stereoscopic depth; secondly, extracting the region of interest of the image, obtaining Its parallax angle, width angle, and front-background contrast are calculated according to the objective visual comfort calculation model; then, through subjective experiments, the apparent distance value of the subjective perception of the stereoscopic image is obtained; finally, the solution is established through the subjective measurement value. Based on the relationship between the calculated stereo depth and objective comfort, an apparent distance calculation model based on stereo comfort is established; the specific steps are as follows:
步骤1、提取立体图像对放映时的屏幕视差,根据下式计算其立体深度:
其中,SD表示解算的立体深度,e表示观看者的眼瞳距离(一般常取6.5cm),V表示观看者距播放屏幕的距离,Z表示播放时屏幕上立体图像对的视差;Among them, SD represents the stereoscopic depth of the solution, e represents the eye-pupil distance of the viewer (usually 6.5cm), V represents the distance between the viewer and the playback screen, and Z represents the parallax of the stereo image pair on the screen during playback;
步骤2、分别得到立体图像的前景区域视差角、宽度角及前后景区域对比度,根据下式计算其客观视觉舒适度:
VC(D,w,c)=4.8736-0.7084D+0.1912ln(w)-0.0208Dln(w)+0.0015c2-0.0572cVC(D,w,c)=4.8736-0.7084D+0.1912ln(w)-0.0208Dln(w)+0.0015c 2 -0.0572c
(0.50≤D≤2.00,0.25≤w≤4.00)(0.50≤D≤2.00, 0.25≤w≤4.00)
其中,D表示前景区域的视差角,w表示前景区域的宽度角,c表示前景区域与后景区域的对比度。要求视差角在范围0.5°~2.0°内,宽度角在范围0.25°~4.0°内;VC为客观立体视觉舒适度;Among them, D represents the disparity angle of the foreground area, w represents the width angle of the foreground area, and c represents the contrast between the foreground area and the background area. It is required that the parallax angle is within the range of 0.5°~2.0°, and the width angle is within the range of 0.25°~4.0°; VC is the objective stereoscopic vision comfort;
步骤3、对实拍的21组立体图像进行主观实验,人为地测量实验人员主观感知的视在距离,整理数据,获取数据平均值。
步骤4、计算每组图像的感知视在距离与立体深度的差值dif,分析该差值与舒适度VC的关系,做曲线拟合得出线性关系如下:
dif=0.13-0.0353*VCdif=0.13-0.0353*VC
其中,dif为感知视在距离(GTPD)与立体深度(SD)的差值:dif=GTPD-SD,VC为客观立体视觉舒适度;Wherein, dif is the difference between perceived apparent distance (GTPD) and stereo depth (SD): dif=GTPD-SD, VC is objective stereo vision comfort;
步骤5、整理公式,最终得到视在距离、立体深度与立体舒适度的关联关系,得出视在距离的解算模型:
OPD=SD+0.13-0.0353*VCOPD=SD+0.13-0.0353*VC
式中,OPD表示模型计算的客观视在距离,SD表示模型解算的立体深度,VC表示客观立体视觉舒适度模型解算的舒适度值。In the formula, OPD represents the objective apparent distance calculated by the model, SD represents the stereo depth calculated by the model, and VC represents the comfort value calculated by the objective stereo vision comfort model.
本发明的积极效果是实现了对立体图像视在距离的客观估计,通过分析立体深度解算模型的计算结果、客观视觉舒适度模型解算结果与主观感知的视在距离的关联关系,建立客观立体视在距离的解算模型。能够有效地节约传统人工评价的人力成本,和依赖于评价者的问题;将模型出发点建立在人眼的基础上,使得模型更具客观性,对立体拍摄起到更好的指导作用;结合立体深度、立体视觉舒适度等,综合考量多维因素,使得模型评价结果更为准确、可信度高。The positive effect of the present invention is to realize the objective estimation of the apparent distance of the stereoscopic image, and establish the objective estimation by analyzing the calculation result of the stereoscopic depth calculation model, the calculation result of the objective visual comfort model and the subjectively perceived apparent distance. The solution model for stereo apparent distance. It can effectively save the labor cost of traditional manual evaluation and the problem of relying on the evaluator; the starting point of the model is based on the human eye, which makes the model more objective and plays a better guiding role in stereo shooting; combined with stereo Depth, stereo vision comfort, etc., comprehensively consider multi-dimensional factors, making the model evaluation results more accurate and reliable.
附图说明Description of drawings
图1为本发明流程示意图。Fig. 1 is a schematic flow chart of the present invention.
图2为双目相机系统。Figure 2 is a binocular camera system.
图3为实验组右视图。Figure 3 is the right view of the experimental group.
图4为实验环境。Figure 4 is the experimental environment.
图5为实验组感兴趣区域图。Figure 5 is a map of the region of interest in the experimental group.
图6为立体深度与视在距离的趋势图。Fig. 6 is a trend graph of stereo depth and apparent distance.
图7为视在距离与立体深度的差值与立体舒适度的关系。Fig. 7 shows the relationship between the difference between the apparent distance and the three-dimensional depth and the three-dimensional comfort.
图8为感知的视在距离、立体深度与解算的视在距离三者趋势图。FIG. 8 is a trend diagram of perceived apparent distance, stereoscopic depth and calculated apparent distance.
具体实施方式Detailed ways
下面结合附图和实施例对本发明做进一步的描述:本实施例中,采用大恒工业相机获取21组实拍立体图像对,并使用韩国先进科学院提供的立体图像库对该模型进行验证,流程图如图1所示,具体步骤如下:The present invention will be further described below in conjunction with the accompanying drawings and embodiments: In this embodiment, 21 groups of real-shot stereo image pairs are obtained by using Daheng industrial cameras, and the model is verified using the stereo image library provided by the Korean Academy of Advanced Sciences. The process flow As shown in Figure 1, the specific steps are as follows:
步骤1、使用两个大恒MER-310-12UC工业相机(如图2所示)进行图像采集,相机的分辨率均为为2048×1536。且两台型号相同的相机,其白平衡、增益、模式等基础参数标准一致。在进行数据采集时,首先连接计算机传送信号,将设备置于平稳的水平三角支架,组装后对准校正,通过大恒相机配备的软件在显示器上观看双机画面。将两台机器的光圈、焦距、焦点等随机信息同步一致,以人眼瞳孔距(5.5-7.5cm,本实验选取6.5cm)为拍摄基线进行同步拍摄。本实验在立体图像拍摄过程中,场景布局多分为两层,画面较简洁,通过层次对比凸显出较好的立体效果。如图3所示,为采集到的21组图像的右视图。
步骤2、选取年龄在19-28岁之间的实验人员15名。确定实验人员的年龄在合理范围,视力正常且无视觉方面的生理疾病,如色盲、色弱等眼部疾病。且在观看过程中佩戴时分式立体眼镜无不良反应,在观看时能准确且稳定的判断出感知到的立体影像位置并确定感知的视在距离。并根据其是否有一定的观测立体影像的经验、是否了解相关专业知识进行区分。由于有相关经验的被测人员能更好的进行主观评测,而其他观测人员由于缺少相关经验,需要先安排一系列的培训,通过前期的训练,能更准确的确定主观评判结果。
步骤3、使用英伟达NVIDIA Quadro K620显卡、三星2233RZ 3D显示屏配合英伟达无线立体幻镜--时分式立体眼镜进行观看立体图像。观看距离为1.5米,约为显示器高度5倍,水平(50cm)和垂直(30cm)视角分别为18.96°和8.58°。实验环境基于ITU-R BT.500-11和ITU-R BT.1438推荐设置,为保证观看过程中观看人员的眼部舒适,在进行实验时每两对图像播放间隔可进行眼部休息。在实验观看过程中仿照影院观影环境,将环境光调暗,如图4所示,为暗室下受测人员进行实验。
步骤4、为获取感知的视在距离,使用主观评测的方法进行实验,在相关研究中,测量方法主要有3种类型:言语估计,视觉想象动作和感知匹配。词汇估计是指以米为单位直接表示深度估计。视觉想象的动作是指让主体观察物体并想象走向物体。想象时间以他们通常的行走速度记录以进行深度判断。本实施例中实验采用语言估计的方法,直接量化估计感知的视在距离。通过人为的测量来获取观看人员所感知到的各图像的视在距离,将获得到的视在距离取其平均值,整理数据进而得到表1所示的部分实验组图像对的视在距离值。
表1部分图像对的视在距离Table 1 Apparent distance of some image pairs
步骤5、通过放映时的视差、观看者瞳距及观看距离等基本参数计算出实验组各图像对所对应的立体深度。根据下式可以解算出影像在播放时某一帧画面的立体深度值:
其中,SD表示解算的立体深度,e表示观看者的眼瞳距离(一般常取6.5cm),V表示观看者距播放屏幕的距离,Z表示播放时屏幕上立体图像对的视差。如表2所示,为实验组图像对(图3所示)所对应的部分立体深度解算值。Among them, SD represents the calculated stereo depth, e represents the eye-pupil distance of the viewer (generally 6.5 cm), V represents the distance between the viewer and the playback screen, and Z represents the parallax of the stereo image pair on the screen during playback. As shown in Table 2, it is the partial stereo depth calculation value corresponding to the image pair of the experimental group (shown in Figure 3).
表2部分图像对的立体深度Table 2 Stereo Depth of Some Image Pairs
步骤6、获取实验组图像的视差图,并根据GBVS算法计算出各图像的平面显著图。然后,根据下式计算并融合视差图与平面显著图得到立体显著图:
IIS(x,y)=ω1ISR(x,y)+ω2DR(x,y)IIS(x,y)=ω 1 IS R (x,y)+ω 2 D R (x,y)
式中ISR(x,y)为平面显著图,DR(x,y)为视差图,ω1、ω2为其权重,本实施例取ω1=ω2=0.5。为得到感兴趣区域,将立体显著图进行阈值分割处理得到掩膜图像IIM(x,y),具体方法如下式:In the formula, IS R (x, y) is a plane saliency map, D R (x, y) is a disparity map, and ω 1 and ω 2 are their weights. In this embodiment, ω 1 =ω 2 =0.5. In order to obtain the region of interest, the stereo saliency map is subjected to threshold segmentation processing to obtain the mask image IIM(x, y). The specific method is as follows:
式中C(x,y)为(x,y)处的像素值,T为其分割阈值。若C(x,y)>T则该像素属于感兴趣区域,对应掩膜图像中白色区域,否则属于黑色区域。使用预置的感兴趣区掩膜与右视图、视差图相乘,分别得到感兴趣区图像和感兴趣区视差图,将感兴趣区域作为前景区域。对感兴趣区掩膜图像取反,并与右视图相乘可得背景区域。如图5所示,为实验组图像(图3)对应的感兴趣区域图。In the formula, C(x, y) is the pixel value at (x, y), and T is its segmentation threshold. If C(x,y)>T, the pixel belongs to the region of interest, corresponding to the white region in the mask image, otherwise it belongs to the black region. Use the preset ROI mask to multiply the right view and the disparity map to obtain the ROI image and the ROI disparity map respectively, and use the ROI as the foreground area. Invert the mask image of the ROI and multiply it with the right view to obtain the background area. As shown in FIG. 5 , it is a map of the region of interest corresponding to the image of the experimental group ( FIG. 3 ).
步骤7、根据下式可计算实验组图像对应的客观视觉舒适度:
VC(D,w,c)=4.8736-0.7084D+0.1912ln(w)-0.0208Dln(w)+0.0015c2-0.0572cVC(D,w,c)=4.8736-0.7084D+0.1912ln(w)-0.0208Dln(w)+0.0015c 2 -0.0572c
(0.50≤D≤2.00,0.25≤w≤4.00)(0.50≤D≤2.00, 0.25≤w≤4.00)
其中,D表示前景区域的视差角,w表示前景区域的宽度角,c表示前景区域与后景区域的对比度。要求视差角在范围0.5°~2.0°内,宽度角在范围0.25°~4.0°内。D、w、c的计算如下式:Among them, D represents the disparity angle of the foreground area, w represents the width angle of the foreground area, and c represents the contrast between the foreground area and the background area. It is required that the parallax angle is within the range of 0.5°-2.0°, and the width angle is within the range of 0.25°-4.0°. D, w, c are calculated as follows:
其中,前景区域的平均视差值为Df,of表示前景区域,|of|表示of区域内像素总数,D为前景区域的平均视差角,k为放映放大率,D表示视差,L为观看时距屏幕距离。Among them, the average parallax value of the foreground area is D f , o f represents the foreground area, |o f | represents the total number of pixels in the area of o f , D is the average parallax angle of the foreground area, k is the projection magnification, D represents the parallax, L is the distance from the screen when viewing.
式中Nf表示前景区域内水平线的数量,表示前景区域内第n条水平线长度。w为宽度角,W表示平均宽度,k为放映放大率,L为观看时距屏幕距离。where N f represents the number of horizontal lines in the foreground area, Indicates the length of the nth horizontal line in the foreground area. w is the width angle, W is the average width, k is the projection magnification, and L is the distance from the screen when viewing.
将每个颜色通道量化为16个不同值,将颜色的数量减少了4096倍。然后将RGB空间装换成Lab空间内,获取前后景区域间的颜色距离集合。dr(r1,r2)为区域r1和r2之间的颜色距离。式中f(ck,i)表示第i个颜色ck,i在第k个区域rk的所有颜色nk中出现的频率。d(c1,i,c2,j)表示区域1中第i个颜色与区域2中第j个颜色在Lab空间内的颜色距离。为基于空间加权的区域对比度(即客观舒适度模型中的c),Sr(rk,ri)是区域rk与ri之间的空间距离,σs控制空间权值的强度。σs越大空间权值的影响越小,则背景区域的影响越显著。两区域间的空间距离被定义为区域重心间的欧几里得距离。其中,像素坐标为归一化后结果,取 Quantize each color channel to 16 different values, reducing the number of colors by a factor of 4096. Then the RGB space is replaced with the Lab space to obtain the color distance set between the foreground and foreground regions. d r (r 1 , r 2 ) is the color distance between regions r 1 and r 2 . where f(c k,i ) represents the frequency of the i-th color c k,i appearing in all colors nk in the k-th region r k . d(c 1,i ,c 2,j ) represents the color distance between the i-th color in
如表3所示,表中为部分实验图像所对应的视觉舒适度(VC)值。As shown in Table 3, the table shows the visual comfort (VC) values corresponding to some experimental images.
表3各图像的视觉舒适度值Table 3 Visual comfort value of each image
步骤8、为分析立体深度与视在距离关系,所示的立体深度与视在距离趋势图,从图中可观察得出:人眼感知的视在距离很大程度上受立体深度影响,但其与模型解算得到的立体深度在数值上会存在一定的出入,因此将人因这一因素纳入到视在距离计算的考量范围。而由于立体视觉舒适度直接反应观看者对立体影像的观看感受,同时也是评价立体影像质量的重要准则之一,因此本实施例将立体舒适度作为人因控制环节的反馈因素纳入到视在距离量化的过程中。根据立体深度解算模型与立体舒适度客观评价模型计算得到各图像对的立体深度(VD)及立体视觉舒适度(VC),结合主观评测得到的视在距离(GTPD)的值进行分析。
步骤9、为获取立体深度、立体视觉舒适度与视在距离三者关联关系,首先计算立体深度与视在距离的差异(dif=GTPD-SD)。如图7所示,随着视觉舒适度的增加,人眼实际感知的视在距离与解算的立体深度间的差值在不断减小,可进一步分析出立体舒适度与差异间的关系,通过建立散点图并作拟合曲线得到如下式:
dif=0.13-0.0353*VCdif=0.13-0.0353*VC
步骤10、整理公式获得立体视在距离解算模型如下式所示:
OPD=SD+0.13-0.0353*VCOPD=SD+0.13-0.0353*VC
步骤11、在本实施例中,利用五个常用客观参量作为评价指标对模型估计值与主观评价值的相关性做分析,利用本发明模型计算这些图像的视在距离,然后比较对应的客观感知视在距离、对应的立体深度与主观感知的视在深度测量值间的Pearson相关系数(Pearson Linear Correlation Coefficient,PLCC),Kendall相关系数(Kendall Rank-order Correlation Coefficient,KRCC),SROCC相关系数(SpearmanRank OrderCorrelation Coefficient),平均绝对值误差(Mean Absolute Error,MAE)和均方根误差(Root Mean Squared Error,RMSE)。并在IVY图像库中选取部分视觉舒适度好与视觉舒适度不好的图像作为测试集,比较对应的客观感知视在距离、对应的立体深度与主观感知的视在深度测量值间的Pearson相关系数,Kendall相关系数,SROCC相关系数,平均绝对值误差和均方根误差,结果验证本发明方法得到的立体图像的视在距离解算值较立体深度解算模型结果更接近人眼感知的视在距离,性能较好。
Claims (1)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910273315.6A CN110060291B (en) | 2019-04-04 | 2019-04-04 | A Calculation Method of Stereo Apparent Distance Considering Human Factors |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910273315.6A CN110060291B (en) | 2019-04-04 | 2019-04-04 | A Calculation Method of Stereo Apparent Distance Considering Human Factors |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN110060291A CN110060291A (en) | 2019-07-26 |
| CN110060291B true CN110060291B (en) | 2023-01-31 |
Family
ID=67318414
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201910273315.6A Active CN110060291B (en) | 2019-04-04 | 2019-04-04 | A Calculation Method of Stereo Apparent Distance Considering Human Factors |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN110060291B (en) |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN112153362B (en) * | 2020-09-15 | 2022-03-08 | 清华大学深圳国际研究生院 | Method and system for measuring stereoscopic depth of naked eye 3D display system |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102402005A (en) * | 2011-12-06 | 2012-04-04 | 北京理工大学 | Bifocal-surface monocular stereo helmet-mounted display device with free-form surfaces |
| CN103118265A (en) * | 2011-11-16 | 2013-05-22 | 克里斯蒂数字系统美国有限公司 | A collimated stereo display system |
| CN103986925A (en) * | 2014-06-05 | 2014-08-13 | 吉林大学 | Evaluation Method of Visual Comfort of Stereo Video Based on Brightness Compensation |
| GB201419379D0 (en) * | 2014-10-31 | 2014-12-17 | Nokia Corp | Method for alignment of low-quality noisy depth map to the high-resolution colour image |
| CN104887316A (en) * | 2015-04-24 | 2015-09-09 | 长春理工大学 | Virtual three-dimensional endoscope displaying method based on active three-dimensional displaying technology |
| CN106570900A (en) * | 2016-10-11 | 2017-04-19 | 宁波大学 | Three-dimensional image relocation method |
| CN109167988A (en) * | 2018-08-29 | 2019-01-08 | 长春理工大学 | A kind of stereo image vision comfort level evaluation method based on D+W model and contrast |
Family Cites Families (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20110222757A1 (en) * | 2010-03-10 | 2011-09-15 | Gbo 3D Technology Pte. Ltd. | Systems and methods for 2D image and spatial data capture for 3D stereo imaging |
| CN103595990B (en) * | 2013-10-30 | 2015-05-20 | 清华大学 | Method for obtaining binocular stereoscopic video comfort level of motion perception |
| JP2015154101A (en) * | 2014-02-10 | 2015-08-24 | ソニー株式会社 | Image processing method, image processor and electronic apparatus |
-
2019
- 2019-04-04 CN CN201910273315.6A patent/CN110060291B/en active Active
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN103118265A (en) * | 2011-11-16 | 2013-05-22 | 克里斯蒂数字系统美国有限公司 | A collimated stereo display system |
| CN102402005A (en) * | 2011-12-06 | 2012-04-04 | 北京理工大学 | Bifocal-surface monocular stereo helmet-mounted display device with free-form surfaces |
| CN103986925A (en) * | 2014-06-05 | 2014-08-13 | 吉林大学 | Evaluation Method of Visual Comfort of Stereo Video Based on Brightness Compensation |
| GB201419379D0 (en) * | 2014-10-31 | 2014-12-17 | Nokia Corp | Method for alignment of low-quality noisy depth map to the high-resolution colour image |
| CN104887316A (en) * | 2015-04-24 | 2015-09-09 | 长春理工大学 | Virtual three-dimensional endoscope displaying method based on active three-dimensional displaying technology |
| CN106570900A (en) * | 2016-10-11 | 2017-04-19 | 宁波大学 | Three-dimensional image relocation method |
| CN109167988A (en) * | 2018-08-29 | 2019-01-08 | 长春理工大学 | A kind of stereo image vision comfort level evaluation method based on D+W model and contrast |
Non-Patent Citations (2)
| Title |
|---|
| Measurement of parallax distribution, and its application to the analysis of visual comfort for stereoscopic HDTV;Yuji Nojiri 等;《Stereoscopic Displays and Virtual Reality Systems X》;20030531;195-205 * |
| 多视点虚拟场景的立体效果控制算法与应用;张玉强;《中国优秀硕士学位论文全文数据库 信息科技辑》;20180315;I138-1990 * |
Also Published As
| Publication number | Publication date |
|---|---|
| CN110060291A (en) | 2019-07-26 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Mittal et al. | Algorithmic assessment of 3D quality of experience for images and videos | |
| CN103986925B (en) | based on the stereoscopic video visual comfort evaluation method of luminance compensation | |
| CN104853185A (en) | Stereo video comfort evaluation method combining multiple parallaxes with motion | |
| CN207589060U (en) | A naked-eye three-dimensional display device combined with visual fatigue detection | |
| Choi et al. | Visual fatigue modeling and analysis for stereoscopic video | |
| WO2015096346A1 (en) | Test method and test device for 3d display crosstalk | |
| CN103780895A (en) | Stereoscopic video quality evaluation method | |
| Chen et al. | Study of subject agreement on stereoscopic video quality | |
| CN104581141A (en) | Three-dimensional picture visual comfort evaluation method | |
| WO2019041352A1 (en) | Panum's area measurement method and apparatus, and wearable display device | |
| Opozda et al. | The survey of subjective and objective methods for quality assessment of 2D and 3D images | |
| CN110139095B (en) | Naked eye 3D display module detection method and system and readable storage medium | |
| CN110060291B (en) | A Calculation Method of Stereo Apparent Distance Considering Human Factors | |
| CN106791802B (en) | A kind of the stereoscopic display visual comfort detection method and system of difference viewing angle and parallax | |
| Hanhart et al. | Subjective evaluation of two stereoscopic imaging systems exploiting visual attention to improve 3D quality of experience | |
| CN108259888A (en) | The test method and system of stereo display effect | |
| CN105915894A (en) | Naked eye stereotelevision 3D crosstalk value measuring method | |
| CN107018398B (en) | A method of for the quantization calibration of light field Three-dimensional Display | |
| Cho et al. | The measurement of eyestrain caused from diverse binocular disparities, viewing time and display sizes in watching stereoscopic 3D content | |
| Kim et al. | Perceptual crosstalk prediction on autostereoscopic 3D display | |
| CN206650798U (en) | The test system of stereo display effect | |
| CN105430370B (en) | Stereoscopic image visual comfort evaluation method based on ranking learning | |
| CN206341331U (en) | Detection device and stereoscopic display device tracking effect evaluation system | |
| CN106331677B (en) | The evaluation method and system of bore hole stereoscopic display intermediate-resolution index | |
| Barkowsky et al. | Is visual fatigue changing the perceived depth accuracy on an autostereoscopic display? |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |