CN106203269A - A kind of based on can the human face super-resolution processing method of deformation localized mass and system - Google Patents
A kind of based on can the human face super-resolution processing method of deformation localized mass and system Download PDFInfo
- Publication number
- CN106203269A CN106203269A CN201610494253.8A CN201610494253A CN106203269A CN 106203269 A CN106203269 A CN 106203269A CN 201610494253 A CN201610494253 A CN 201610494253A CN 106203269 A CN106203269 A CN 106203269A
- Authority
- CN
- China
- Prior art keywords
- resolution
- image
- block
- low
- face
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
- G06F18/24147—Distances to closest patterns, e.g. nearest neighbour classification
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
本发明公开一种基于可形变局部块的人脸超分辨率处理方法及系统,本方法主要关注基于局部模型可变的人脸超分辨率方法,在人脸位置先验的基础上,通过引入SIFT Flow特征对库样本图像块进行形变,扩充样本库图像块模式,增强已有图像块的表达能力,使得重建结果具有更高的精确性,进一步的挖掘样本库中人脸图像块与输入人脸图像块的关系,优化了人脸超分辨率算法的结果;本发明可显著提高恢复图像的视觉感受,特别适用于低质量监控环境下人脸图像的恢复。
The invention discloses a face super-resolution processing method and system based on deformable local blocks. This method mainly focuses on the face super-resolution method based on variable local models. The SIFT Flow feature deforms the sample image blocks of the library, expands the image block mode of the sample library, enhances the expressive ability of the existing image blocks, makes the reconstruction result more accurate, and further mines the face image blocks in the sample library and the input person. The relationship between the face image blocks optimizes the result of the face super-resolution algorithm; the invention can significantly improve the visual experience of the restored image, and is especially suitable for the restoration of the face image in a low-quality monitoring environment.
Description
技术领域technical field
本发明涉及图像处理和图像恢复技术领域,具体涉及一种基于可形变局部块的人脸超分辨率处理方法及系统。The invention relates to the technical field of image processing and image restoration, in particular to a face super-resolution processing method and system based on deformable local blocks.
背景技术Background technique
人脸超分辨率技术是通过辅助训练库,学习高低分辨率对应关系,进而达到从已有的低分辨率人脸图像中估计出高分辨率人脸图像的目的。人脸超分辨率现在被广泛应用于多个领域,其中最具代表性的领域之一就是监控录像中的人脸图像增强。随着监控系统的广泛普及,监控视频在刑事取证和刑侦调查过程中发挥着越来越重要的作用。而人脸图像作为直接证据之一,在案件分析和法庭取证中占据着重要的位置。然而,由于现有条件下,目标嫌疑人与摄像头距离相对较远,捕捉到的监控人脸可用像素非常少,兼之真实情况下由于恶劣天气(例如:雨雾)、光照(例如:光照过强、过暗、明暗不均)、器件等因素对捕获的图像引发的严重损毁(例如:严重的模糊和噪声),图像恢复、放大和辨识往往受到严重的干扰。这就需要用到人脸超分辨率技术提升图像分辨率,从低分辨率图像恢复到高分辨率图像。Face super-resolution technology is to learn the corresponding relationship between high and low resolution through the auxiliary training library, and then achieve the purpose of estimating high-resolution face images from existing low-resolution face images. Face super-resolution is now widely used in many fields, one of the most representative fields is face image enhancement in surveillance video. With the widespread popularization of surveillance systems, surveillance video is playing an increasingly important role in the process of criminal evidence collection and criminal investigation. Face images, as one of the direct evidence, occupy an important position in case analysis and court evidence collection. However, due to the existing conditions, the distance between the target suspect and the camera is relatively far away, and the available pixels of the captured surveillance face are very few. Too dark, uneven light and shade), devices and other factors cause serious damage to the captured image (for example: severe blur and noise), image restoration, amplification and identification are often seriously disturbed. This requires the use of face super-resolution technology to improve image resolution and restore from low-resolution images to high-resolution images.
人脸超分辨率技术便是解决这一难题的技术,它可以通过一副或者多幅低分辨率的人脸图像重构出高分辨率的清晰人脸图像。自从该技术由Baker等人在2000年首次提出之后,这一领域就获得了学者的广泛关注,并且产生一系列优秀的研究成果,其中以基于学习的人脸超分辨率算法最受学者重视。近年来,流形学习逐渐成为了人脸超分辨率的主流方法。这类方法的核心思想是:描述低分辨率图像的流形空间关系,寻找出每个低分辨率图像数据点周围的局部性质,然后将低分辨率图像的流形非线性地映射到高分辨率图像的流形空间中,在高分辨率对应空间上做投影,从而合成高分辨图像。具有代表性的有以下几种方法:2004年,Chang[1]等首次将流形学习法引入图像超分辨率重构中,提出了一种邻域嵌入的图像超分辨率重构法。Sung Won Park[2]提出一种基于局部保持投影的自适应流形学习方法,从局部子流形分析人脸的内在特征,重构出低分辨率图像缺失的高频成分。2005年,Wang[3]提出一种基于PCA(Principal component analysis,主成分分析)分解的方法,把低分辨率待处理图像用低分辨率空间的主成分的线性组合表示,投影系数到对应的高分辨率主成分空间获得最终结果。该方法对早上具有较好的鲁棒性,但是仍然在结果图像的边缘存在鬼影、混叠的现象。2010年,Lan[5]针对监控环境下严重的模糊和噪声导致的图像像素损毁严重的问题,提出一种基于形状约束的人脸超分辨率方法,在传统PCA架构中添加形状约束作为相似度度量准则,利用人眼睛识别形状时对干扰的鲁棒性来人工添加形状特征点作为约束,优化低质量图像的重建结果。2014年,Dong[4]提出基于局部特征转换的方法,进一步解决了这个问题。Face super-resolution technology is a technology to solve this problem. It can reconstruct a high-resolution clear face image from one or more low-resolution face images. Since the technology was first proposed by Baker et al. in 2000, this field has received extensive attention from scholars and produced a series of excellent research results, among which the learning-based face super-resolution algorithm has attracted the most attention from scholars. In recent years, manifold learning has gradually become the mainstream method for face super-resolution. The core idea of this type of method is to describe the manifold spatial relationship of the low-resolution image, find out the local properties around each low-resolution image data point, and then map the manifold of the low-resolution image nonlinearly to the high-resolution image. In the manifold space of high-resolution images, projection is made on the high-resolution corresponding space to synthesize high-resolution images. There are several representative methods: In 2004, Chang [1] et al first introduced the manifold learning method into image super-resolution reconstruction, and proposed a neighborhood-embedded image super-resolution reconstruction method. Sung Won Park [2] proposed an adaptive manifold learning method based on local projection, which analyzed the intrinsic features of the face from the local sub-manifold, and reconstructed the high-frequency components missing from the low-resolution image. In 2005, Wang [3] proposed a method based on PCA (Principal component analysis, principal component analysis) decomposition, which represented the low-resolution image to be processed by a linear combination of the principal components of the low-resolution space, and projected the coefficients to the corresponding High-resolution principal component space to obtain the final result. This method is more robust to the morning, but there are still ghosting and aliasing at the edges of the resulting image. In 2010, Lan [5] proposed a face super-resolution method based on shape constraints to address the problem of serious image pixel damage caused by severe blur and noise in the monitoring environment, adding shape constraints to the traditional PCA architecture as similarity The metric criterion uses the robustness of the human eye to recognize shapes to interfere with artificially adding shape feature points as constraints to optimize the reconstruction results of low-quality images. In 2014, Dong [4] proposed a method based on local feature transformation to further solve this problem.
但是,基于学习的人脸超分辨率技术想要得到更好的复原结果,必须用规模更大的样本库来覆盖更多的图像块模式,解决在流形空间中训练样本不够稠密(有限的训练库与表示人脸特征信息高维流形空间相比)的问题。然而,建立人脸训练样本库是个繁重且复杂的工程。此外,训练库样本越多,人脸进行超分辨率重构运算复杂度越高。因此,如何增强现有的训练样本库人脸图像的表达能力,使其在拟合低分辨率的图像块时给出精确的表达,成为了当前人脸超分辨率技术研究一个亟需解决的问题。However, if the learning-based face super-resolution technology wants to get better restoration results, it must use a larger sample library to cover more image block patterns, and solve the problem that the training samples in the manifold space are not dense enough (limited The training library is compared with the problem of representing the high-dimensional manifold space of facial feature information). However, establishing a face training sample library is a heavy and complicated project. In addition, the more samples in the training library, the higher the computational complexity of face super-resolution reconstruction. Therefore, how to enhance the expressive ability of the face image in the existing training sample library so that it can give an accurate expression when fitting low-resolution image blocks has become an urgent problem to be solved in the current face super-resolution technology research. question.
综上所述,针对上述问题,本文提出了基于局部可形变模型的人脸超分辨率算法,在人脸位置先验的基础上,通过引入SIFT Flow[6]特征对在库样本图像块进行形变,扩充可用图像块模式,增强已有图像块的表达能力,使得重建结果具有更高的精确性,进一步的挖掘样本库中人脸与输入人脸图像的关系,从而实现优化了人脸超分辨率算法结果的目标。在CAS-PEAL-R1人脸数据库中,对下采样四倍和有噪声情况下的测试样本进行实验证明,我们提出的算法客观评价指标和主观重建结果均优于目前最好的算法,并且对噪声具有鲁棒性。To sum up, in view of the above problems, this paper proposes a face super-resolution algorithm based on a locally deformable model. On the basis of the face position prior, by introducing the SIFT Flow [6] feature, the sample image block in the library is processed. Deformation, expand the available image block mode, enhance the expressive ability of the existing image block, make the reconstruction result have higher accuracy, and further mine the relationship between the face in the sample database and the input face image, so as to realize the optimization of face super Target for resolution algorithm results. In the CAS-PEAL-R1 face database, the experiments on the test samples under the condition of four times downsampling and noise prove that the objective evaluation index and subjective reconstruction results of the proposed algorithm are better than the current best algorithms, and the Noise is robust.
综上所述,通过引入SIFT Flow特征对样本库图像块进行形变,增强已经存在的图像块的表达能力,覆盖更多的图像块模式,进一步优化了人脸超分辨率算法。To sum up, by introducing SIFT Flow features to deform the image blocks of the sample library, the expression ability of existing image blocks is enhanced, and more image block modes are covered, which further optimizes the face super-resolution algorithm.
文中涉及如下参考文献:The following references are involved in the article:
[1]H.Chang,D.-Y.Yeung,and Y.Xiong,“Super-resolution through neighborembedding,”in Proc.IEEE Conf.Comput.Vis.Pattern Recog.,Jul.2004,pp.275–282.[1] H. Chang, D.-Y. Yeung, and Y. Xiong, “Super-resolution through neighbor embedding,” in Proc.IEEE Conf.Comput.Vis.Pattern Recog.,Jul.2004,pp.275–282 .
[2]Sung Won Park,Savvides,M."Breaking the Limitation of ManifoldAnalysis for Super-Resolution of Facial Images",ICASSP,pp:573-576,2007.[2]Sung Won Park, Savvides, M."Breaking the Limitation of Manifold Analysis for Super-Resolution of Facial Images", ICASSP, pp:573-576, 2007.
[3]Xiaogang Wang and Xiaoou Tang,“Hallucinating face byeigentransformation,”Systems,Man,and Cybernetics,Part C:Applications andReviews,IEEE Transactions on,vol.35,no.3,pp.425–434,2005.[3] Xiaogang Wang and Xiaoou Tang, "Hallucinating face byeigen transformation," Systems, Man, and Cybernetics, Part C: Applications and Reviews, IEEE Transactions on, vol.35, no.3, pp.425–434, 2005.
[4]董小慧,高戈,陈亮等.2014.数据驱动局部特征转换的噪声人脸幻构[J].计算机应用,34(12):3576-3579.[4] Dong Xiaohui, Gao Ge, Chen Liang, etc. 2014. Noisy face illusion based on data-driven local feature transformation [J]. Computer Applications, 34(12): 3576-3579.
[5]C Lan,R Hu,Z Han,A face super-resolution approach using shapesemantic mode regularization.IEEE International Conference on ImageProcessing(ICIP),2021–2024,26-29Sept.2010.[5] C Lan, R Hu, Z Han, A face super-resolution approach using shapesemantic mode regularization. IEEE International Conference on Image Processing (ICIP), 2021–2024, 26-29Sept.2010.
[6]Ce Liu,Jenny Yuen,Antonio Torralba,et al.2011.SIFT Flow:DenseCorrespondence across Scenes and its Applications[J],IEEE Transactions onPattern Analysis and Machine Intelligence,33(5):978-994.[6] Ce Liu, Jenny Yuen, Antonio Torralba, et al. 2011. SIFT Flow: Dense Correspondence across Scenes and its Applications [J], IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(5): 978-994.
发明内容Contents of the invention
针对现有技术存在的问题,本发明提供了一种基于可形变局部块的人脸超分辨率处理方法及系统,尤其适用于低质量监控视频中人脸图像的恢复。Aiming at the problems existing in the prior art, the present invention provides a face super-resolution processing method and system based on deformable local blocks, especially suitable for restoration of face images in low-quality surveillance videos.
为了解决上述技术问题,本发明采用如下的技术方案;In order to solve the above technical problems, the present invention adopts the following technical solutions;
一种基于可形变局部块的人脸超分辨率处理方法,包括以下步骤:A face super-resolution processing method based on deformable local blocks, comprising the following steps:
S1:构建包含高分辨率人脸图像库及其对应的低分辨率人脸图像库的训练库;S1: Construct a training library containing a high-resolution face image library and its corresponding low-resolution face image library;
S2:采用相同的分块方式将待处理低分辨率人脸图像和训练库中图像划分为具交叠部分的图像块,所述的图像块为边长为psize的正方形图像块;S2: Divide the low-resolution face image to be processed and the images in the training library into image blocks with overlapping parts in the same block-dividing manner, and the image blocks are square image blocks with a side length of psize;
S3:对待处理低分辨率人脸图像每一块,在对应位置的低分辨率训练块集合中查找其近邻块;S3: For each block of the low-resolution face image to be processed, search for its neighbor blocks in the low-resolution training block set at the corresponding position;
S4:计算待处理低分辨率人脸图像块,每一块到近邻的形变场矩阵;S4: Calculate the low-resolution face image block to be processed, and the deformation field matrix from each block to the neighbor;
S5:根据形变场矩阵,计算每一个近邻块到对应待处理低分辨率人脸图像块的形变块;S5: According to the deformation field matrix, calculate each adjacent block to the deformation block corresponding to the low-resolution face image block to be processed;
S6:计算待处理低分辨率人脸图像块和其近邻形变块之间的权重系数;S6: Calculate the weight coefficient between the low-resolution face image block to be processed and its adjacent deformation block;
S7:将权重投影到高分辨率空间上,根据重建系数恢复图像块获得其对应的高分辨率人脸图像块 S7: Project the weights onto the high-resolution space and restore the image blocks according to the reconstruction coefficients Obtain its corresponding high-resolution face image block
S8:拼接高分辨率人脸图像块得高分辨率人脸图像。S8: Stitching high-resolution face image blocks high-resolution face images.
进一步的,所述的S1中;将高分辨率人脸图像库中高分辨率人脸图像位置对齐,并进行降质处理,得到对应的低分辨率人脸图像库,高分辨率人脸图像库和低分辨率人脸图像库构成训练库;Further, in said S1; aligning the positions of the high-resolution face images in the high-resolution face image library, and performing degrading processing to obtain the corresponding low-resolution face image library, and the high-resolution face image library and the low-resolution face image library to form a training library;
在S2之前,使待处理低分辨率人脸图像与训练库中图像大小相同,且位置对齐。Before S2, the size of the low-resolution face image to be processed is the same as that of the image in the training library, and the positions are aligned.
进一步的,所述的位置对齐采用仿射变换法将进行位置对齐;具体五个位置包括:两个眼角、一个鼻尖、两个嘴角;Further, the position alignment adopts the affine transformation method to perform position alignment; the specific five positions include: two corners of the eyes, one tip of the nose, and two corners of the mouth;
仿射变换法为,将高分辨率人脸图像库中所有人脸图像相加并除以样本数,得平均脸;设(x′i,y′i)为平均脸上第i个特征点坐标,(xi,yi)为待对齐的高分辨率人脸图像上对应的第i个特征点坐标;设仿射矩阵其中a、b、c、d、e、f为仿射变换系数,表示平均脸和待对齐的高分辨率人脸图像上第i个特征点坐标(x′i,y′i)和(xi,yi)间的关系,采用直接线性变换法求解仿射变换矩阵M;待对齐的高分辨率人脸图像所有坐标点与仿射矩阵M相乘得到的坐标即对齐后的高分辨率人脸图像坐标。The affine transformation method is to add all the face images in the high-resolution face image library and divide them by the number of samples to obtain the average face; let (x′ i ,y′ i ) be the i-th feature point on the average face Coordinates, (x i , y i ) are the coordinates of the i-th feature point corresponding to the high-resolution face image to be aligned; set the affine matrix Where a, b, c, d, e, f are affine transformation coefficients, Represents the relationship between the average face and the i-th feature point coordinates (x′ i , y′ i ) and ( xi , y i ) on the high-resolution face image to be aligned, and uses the direct linear transformation method to solve the affine transformation Matrix M; the coordinates obtained by multiplying all coordinate points of the high-resolution face image to be aligned with the affine matrix M are the coordinates of the high-resolution face image after alignment.
进一步的,所述的S3中,对于待处理低分辨率图像xin,假设在位置i上的块为低分辨率图像库设为X,X上在位置i的所以图像块记为Xi;在Xi上的K个近邻块,通过和Xi的每一个图像块差值的绝对值一一对比获得,差值绝对值最小的K个低分辨率图像块,作为的的近邻,记为 Further, in the above S3, for the low-resolution image x in to be processed, it is assumed that the block at position i is The low-resolution image library is set to X, and the image block at position i on X is recorded as X i ; The K nearest neighbor blocks on Xi, by Compared with the absolute value of each image block difference of X i one by one, the K low-resolution image blocks with the smallest absolute value of the difference are obtained as neighbors of
进一步的,所述的S4中,每一个图像块的形变场矩阵[u,v]可通过待处理图像与之间通过以下方法计算得到;Further, in the S4, each image block The deformation field matrix [u, v] of the image to be processed can be and It is calculated by the following method;
设p=(x,y)为图像上像素p的坐标,设w(p)=(u(p),v(p))是像素p点的流向量,只准许u(p)和v(p)是整数集;使s1和s2表示与的SIFT特征;W(p)为在p的像素点的流向量,其中u(p)和v(p)分别代表该像素点在水平和垂直两个方向的位移场;得到能量函数E(w):Let p=(x,y) be the coordinates of pixel p on the image, let w(p)=(u(p),v(p)) be the flow vector of pixel p, only u(p) and v( p) is the set of integers; let s 1 and s 2 denote and SIFT features; W(p) is the flow vector at the pixel point of p, where u(p) and v(p) represent the displacement field of the pixel point in the horizontal and vertical directions respectively; the energy function E(w ):
其中,第一行是数据项,表示的是让SIFT特征沿着像素p点的流向量w(p)来匹配;第二项是位移项,表示的是当没有其他信息可利用时,让流向量w(p)尽可能的小;第三项是平滑项,让相邻像素相似;在这个目标函数中,L1准则被用于第一行数据项和第三行平滑项当中,来计算匹配离群和流向量不连续,t和d分别作为阈值;η、α是平衡参数,根据经验赋值;Among them, the first line is a data item, which means that the SIFT feature is matched along the flow vector w(p) of the pixel point p; the second item is a displacement item, which means that when no other information is available, let the flow direction The amount w(p) is as small as possible; the third item is a smoothing item, which makes adjacent pixels similar; in this objective function, the L1 criterion is used in the first row of data items and the third row of smoothing items to calculate the matching Outliers and flow vectors are discontinuous, t and d are used as thresholds respectively; η and α are balance parameters, which are assigned according to experience;
通过最小化能量函数E(w),得到形变场矩阵[u,v]。By minimizing the energy function E(w), the deformation field matrix [u,v] is obtained.
进一步的,所述的S5中,根据形变场矩阵,计算每一个近邻块到对应待处理低分辨率人脸图像块的形变块,具体过程为:Further, in said S5, according to the deformation field matrix, calculate each adjacent block to the deformation block corresponding to the low-resolution face image block to be processed, the specific process is:
通过引入[u,v],让匹配低分辨率图像块获得到样本库中低分辨率图像块的映射关系,将这个关系定义为特征算子sf(·),通过特征算子sf(·),将匹配到样本库图像块形变至与输入图像块相似,则可得形变场[u,v]的表达式:By introducing [u,v], let Match low-resolution image patches get to the low-resolution image patch in the sample library The mapping relationship is defined as the feature operator sf( ), through the feature operator sf( ), the image block matched to the sample library is deformed to be similar to the input image block, and the deformation field [u,v ] expression:
根据高、低分辨率的图像块在流形空间中具有相似的一致性,通过形变矩阵[u,v],得到形变图像块Rk(i,j),定义Deform(·)表示将的形变过程,则According to the similar consistency of high and low resolution image blocks in the manifold space, the deformed image block R k (i, j) is obtained through the deformation matrix [u, v], and Deform( ) is defined to represent deformation process, then
对于训练样本库中的低分辨率图像块形变后的图像块表示为类似的,低分辨率图像块所对应的高分辨率图像块,形变后的高分辨率图像块表示为 For low-resolution image patches in the training sample library The deformed image block is expressed as Similarly, for the high-resolution image block corresponding to the low-resolution image block, the deformed high-resolution image block is expressed as
根据引入SIFT Flow特征算子sf(·),得到形变矩阵[u,v];利用形变矩阵[u,v]形变图像块,得到与输入图像块更相似的形变图像块,使得样本库图像块的表达更加准确,增强样本库中样本的表达能力。According to the introduction of the SIFT Flow feature operator sf( ), the deformation matrix [u, v] is obtained; the deformation matrix [u, v] is used to deform the image block to obtain a deformed image block that is more similar to the input image block, so that the sample library image block The expression is more accurate, and the expression ability of the samples in the sample library is enhanced.
进一步的,所述的S6中,计算待处理图像块和近邻形变块之间的权重系数,具体过程如下:Further, in said S6, the weight coefficient between the image block to be processed and the adjacent deformation block is calculated, and the specific process is as follows:
首先引入欧氏距离dk(i,j):First introduce the Euclidean distance d k (i,j):
即是输入低分辨率率图像块与训练样本中的低分辨率图像块的欧式距离;然后选取欧式距离近的K个高分辨率图像块进行形变,因此此时形变图像块可表示为:That is, the input low-resolution image block with the low-resolution image patches in the training samples The Euclidean distance; then select K high-resolution image blocks with the closest Euclidean distance for deformation, so the deformed image block can be expressed as:
输入的低分辨率人脸图像块由样本训练集中K个距离最近的低分辨率形变图像块表示的最优表示权重w*(i,j)为:Input low-resolution face image patch From the sample training set K nearest low-resolution deformed image blocks The optimal representation weight w * (i,j) expressed is:
其中,是用样本库中的低分辨率形变图像块线性表示输入的低分辨率图像块的权重系数,wi是元素为的向量;τ是平衡参数,根据经验赋值。in, is to use the low-resolution warped image patch from the sample library Linearly represents the weight coefficient of the input low-resolution image block, w i is the element The vector; τ is a balance parameter, assigned according to experience.
进一步的,所述的S7中,对于重构出的目标高分辨率图像块它可由最近邻K个高分辨率形变图像块与它最优权重合成:Further, in said S7, for the reconstructed target high-resolution image block It can be deformed by K nearest neighbor high-resolution image patches with its optimal weight synthesis:
其中, 为对应的高分辨率样本。in, for Corresponding high-resolution samples.
一种基于可形变局部块的人脸超分辨率处理系统,包括:A face super-resolution processing system based on deformable local blocks, including:
训练库构建模型,用于构建包含高分辨率人脸图像库及其对应的低分辨率人脸图像库的训练库;The training library construction model is used to build a training library comprising a high-resolution face image library and its corresponding low-resolution face image library;
分块模块,用于采用相同的分块方式将待处理低分辨率人脸图像和训练库中图像划分为具交叠部分的图像块,所述的图像块为边长为psize的正方形图像块;The block module is used to divide the low-resolution face image to be processed and the image in the training library into image blocks with overlapping parts in the same block mode, and the image blocks are square image blocks with a side length of psize ;
近邻获取模块,用于对待处理低分辨率人脸图像每一块,在对应位置的低分辨率训练块集合中查找其近邻块;The neighbor acquisition module is used to search for each block of the low-resolution face image to be processed in the low-resolution training block set of the corresponding position;
形变场矩阵计算模块,用于计算待处理低分辨率人脸图像块,每一块到近邻的形变场矩阵;The deformation field matrix calculation module is used to calculate the low-resolution face image block to be processed, and the deformation field matrix from each block to the nearest neighbor;
形变模块,用于根据形变场矩阵,计算每一个近邻块到对应待处理低分辨率人脸图像块的形变块;The deformation module is used to calculate each adjacent block to the deformation block corresponding to the low-resolution face image block to be processed according to the deformation field matrix;
权重系数获取模块,用于计算待处理低分辨率人脸图像块和其近邻形变块之间的权重系数;The weight coefficient acquisition module is used to calculate the weight coefficient between the low-resolution face image block to be processed and its adjacent deformation block;
高分辨率图像块生成模块,用于将权重投影到高分辨率空间上,根据重建系数恢复图像块获得其对应的高分辨率人脸图像块 A high-resolution image patch generation module for projecting weights onto a high-resolution space and recovering image patches based on reconstruction coefficients Obtain its corresponding high-resolution face image block
拼接模块,用于根据位置i拼接高分辨率人脸图像块得高分辨率人脸图像。Stitching module for stitching high-resolution face image blocks according to position i high-resolution face images.
和现有技术相比,本发明具有以下优点和积极效果:Compared with the prior art, the present invention has the following advantages and positive effects:
本方法基于学习人脸超分辨率技术如果想要复原出更好的结果,必须用规模更大的样本库来覆盖更多的图像块模式,解决在流形空间中训练样本不够稠密(有限的训练库与表示人脸特征信息高维流形空间相比)的问题。然而,建立人脸训练样本库是个繁重且复杂的工程,此外,样本库图像越多,人脸进行超分辨率重构运算复杂度越高。因此,如何增强现有的训练样本库人脸图像的表达能力,成为了当前人脸超分辨率研究一个亟需解决的问题。This method is based on learning face super-resolution technology. If you want to restore better results, you must use a larger sample library to cover more image block patterns, and solve the problem that the training samples in the manifold space are not dense enough (limited The training library is compared with the problem of representing the high-dimensional manifold space of facial feature information). However, establishing a face training sample library is a heavy and complicated project. In addition, the more images in the sample library, the higher the computational complexity of face super-resolution reconstruction. Therefore, how to enhance the expression ability of the existing training sample database face images has become an urgent problem to be solved in the current face super-resolution research.
本方法主要关注基于局部模型可变的人脸超分辨率方法,在人脸位置先验的基础上,通过引入SIFT Flow特征对库样本图像块进行形变,扩充样本库图像块模式,增强已有图像块的表达能力,使得重建结果具有更高的精确性,进一步的挖掘样本库中人脸图像块与输入人脸图像块的关系,优化了人脸超分辨率算法的结果。This method mainly focuses on the face super-resolution method based on the variable local model. On the basis of the face position prior, by introducing the SIFT Flow feature to deform the image block of the library sample, expand the image block mode of the sample library, and enhance the existing The expressive ability of the image block makes the reconstruction result more accurate, further mining the relationship between the face image block in the sample library and the input face image block, and optimizing the result of the face super-resolution algorithm.
附图说明Description of drawings
图1是本发明实施例的流程示意图;Fig. 1 is a schematic flow chart of an embodiment of the present invention;
图2是本发明实施例的人脸图像基于位置分块示意图。Fig. 2 is a schematic diagram of face image segmentation based on location according to an embodiment of the present invention.
具体实施方式detailed description
本发明在人脸位置先验的基础上,通过引入SIFT Flow特征对库样本图像块进行形变,扩充样本库图像块模式,增强已有图像块的表达能力,使得重建结果具有更高的精确性,进一步的挖掘样本库中人脸图像块与输入人脸图像块的关系,利用多重表达的一致性作为约束,增强图像块表征的一致性和噪声鲁棒性,提升恢复结果的客观质量和相似度。On the basis of the face position prior, the present invention introduces the SIFT Flow feature to deform the library sample image block, expands the sample library image block mode, and enhances the expressive ability of the existing image block, so that the reconstruction result has higher accuracy , to further mine the relationship between the face image blocks in the sample library and the input face image blocks, use the consistency of multiple expressions as a constraint, enhance the consistency and noise robustness of the image block representation, and improve the objective quality and similarity of the restoration results. Spend.
下面将结合具体实施例和附图对本发明做进一步说明。The present invention will be further described below in conjunction with specific embodiments and accompanying drawings.
本发明面向监控环境下的极低质量人脸图像,采用双层流形假设和一致性约束图像块的表征。具体实施时,本发明技术方案可采用计算机软件技术实现自动运行流程。The invention is oriented to extremely low-quality face images in a monitoring environment, and adopts a double-manifold assumption and a representation of image blocks constrained by consistency. During specific implementation, the technical solution of the present invention can use computer software technology to realize the automatic operation process.
参见图1,本发明具体步骤如下:Referring to Fig. 1, the concrete steps of the present invention are as follows:
S1:构建包含高分辨率人脸图像库及其对应的低分辨率人脸图像库的训练库;S1: Construct a training library containing a high-resolution face image library and its corresponding low-resolution face image library;
将高分辨率人脸图像库Y中高分辨率人脸图像位置对齐,对高分辨率人脸图像进行降质处理得对应的低分辨率人脸图像,从而获得低分辨率人脸图像库X。The positions of the high-resolution face images in the high-resolution face image library Y are aligned, and the high-resolution face images are degraded to obtain corresponding low-resolution face images, thereby obtaining the low-resolution face image library X.
在S2之前,使待处理低分辨率人脸图像与训练库中图像大小相同,且位置对齐。Before S2, the size of the low-resolution face image to be processed is the same as that of the image in the training library, and the positions are aligned.
具体实施中,首先,将高分辨率人脸图像的眼睛和嘴巴位置对齐;然后,对高分辨率人脸图像依次进行下采样、模糊窗过滤、上采样,得到与高分辨率人脸图像对应的低分辨率人脸图像。In the specific implementation, first, the eyes and mouth positions of the high-resolution face image are aligned; then, the high-resolution face image is sequentially down-sampled, fuzzy window filtered, and up-sampled to obtain the corresponding high-resolution face image. low-resolution face images.
为便于实施参考,下面将提供采用仿射变换法实现人脸图像对齐的具体过程:For the convenience of implementation reference, the following will provide the specific process of using the affine transformation method to achieve face image alignment:
所述的位置对齐采用仿射变换法将进行位置对齐;具体五个位置包括:两个眼角、一个鼻尖、两个嘴角。The position alignment is performed by affine transformation method; the specific five positions include: two corners of the eyes, one tip of the nose, and two corners of the mouth.
对高分辨率人脸图像进行特征点标注,特征点为五官边缘点,例如眼角、鼻尖、嘴角等;然后,采用仿射变换法对齐特征点。Mark the feature points on the high-resolution face image, and the feature points are the edge points of the facial features, such as the corners of the eyes, the tip of the nose, the corners of the mouth, etc.; then, use the affine transformation method to align the feature points.
仿射变换法具体为:The affine transformation method is specifically:
将高分辨率人脸图像库Y中所有人脸图像相加并除以样本数,得平均脸。设(x′i,y′i)为平均脸上第i个特征点坐标,(xi,yi)为待对齐的高分辨率人脸图像上对应的第i个特征点坐标。设仿射矩阵其中a、b、c、d、e、f为仿射变换系数,表示平均脸和待对齐的高分辨率人脸图像上第i个特征点坐标(x′i,y′i)和(xi,yi)间的关系,采用直接线性变换法求解仿射变换矩阵M。待对齐的高分辨率人脸图像所有坐标点与仿射矩阵M相乘得到的坐标即对齐后的高分辨率人脸图像坐标。Add all face images in the high-resolution face image library Y and divide by the number of samples to get the average face. Let (x′ i , y′ i ) be the coordinates of the i-th feature point on the average face, and ( xi , y i ) be the coordinates of the corresponding i-th feature point on the high-resolution face image to be aligned. Let the affine matrix Where a, b, c, d, e, f are affine transformation coefficients, Represents the relationship between the average face and the i-th feature point coordinates (x′ i , y′ i ) and ( xi , y i ) on the high-resolution face image to be aligned, and uses the direct linear transformation method to solve the affine transformation Matrix M. The coordinates obtained by multiplying all the coordinate points of the high-resolution face image to be aligned with the affine matrix M are the coordinates of the high-resolution face image after alignment.
对对齐后的高分辨率人脸图像做降质处理,例如,依次对高分辨率人脸图像下采样4倍、模糊窗过滤3*3、上采样4倍,得到与高分辨率人脸图像对应的低分辨率人脸图像,从而获得低分辨率人脸图像库X。Degrade the aligned high-resolution face image, for example, downsample the high-resolution face image by 4 times, blur window filter by 3*3, and upsample by 4 times to obtain a high-resolution face image Corresponding low-resolution face images, so as to obtain the low-resolution face image library X.
高分辨率人脸图像库Y和低分辨率人脸图像库X中人脸图像一一对应,构成高低分辨率人脸图像对。高分辨率人脸图像库Y和低分辨率人脸图像库X构成训练库。There is a one-to-one correspondence between the face images in the high-resolution face image database Y and the low-resolution face image database X, forming a pair of high-resolution and low-resolution face images. The high-resolution face image library Y and the low-resolution face image library X constitute the training library.
使待处理低分辨率人脸图像与训练库中图像大小相同,且位置对齐。Make the low-resolution face image to be processed the same size as the image in the training library, and the positions are aligned.
本发明是要对待处理低分辨率人脸图像xin进行处理,估计出其对应的高分辨率人脸图像,将估计出的高分辨率人脸图像记为待估高分辨率人脸图像yout。The present invention is to process the low-resolution human face image x in to be processed, estimate its corresponding high-resolution human face image, and record the estimated high-resolution human face image as the high-resolution human face image to be estimated y out .
待处理低分辨率人脸图像xin通常是在含噪严重环境获得的低分辨率人脸图像。对于作为输入的待处理低分辨率人脸图像,一般要经过预处理,包括剪切出符合统一规定的人脸部分,即将待处理低分辨率人脸图像xin进行上采样,使其与训练库中人脸图像大小相同。对待处理低分辨率人脸图像xin进行特征点标注,最后采用S1中记载的仿射变换法使待处理低分辨率人脸图像xin与平均脸位置对齐。这样,使得训练库中人脸图像和待处理低分辨率人脸图像xin在尺寸、眉毛高度处于相同的水平。若待处理低分辨率人脸图像xin采集时光线不足,则可对位置对齐后的待处理低分辨率人脸图像xin进行自动亮度对比度调整,使其与训练库中低分辨率人脸图像处于相近亮度水平。The low-resolution face image x in to be processed is usually a low-resolution face image obtained in a noisy environment. For the low-resolution face image to be processed as input, it generally needs to be pre-processed, including cutting out the face part that meets the unified regulations, and upsampling the low-resolution face image x in to be processed to make it consistent with the training The face images in the library are of the same size. Mark the feature points of the low-resolution face image x in to be processed, and finally use the affine transformation method described in S1 to align the low-resolution face image x in to be processed with the average face position. In this way, the face image in the training library and the low-resolution face image x in to be processed are at the same level in size and eyebrow height. If there is insufficient light when the low-resolution face image x in to be processed is collected, automatic brightness and contrast adjustment can be performed on the low-resolution face image x in to be processed after position alignment, so that it is consistent with the low-resolution face in the training library Images are at similar brightness levels.
S2:采用相同的分块方式将待处理低分辨率人脸图像、训练库中图像划分为具交叠部分的正方形图像块;所述的图像块为边长为psize的正方形图像块。S2: Divide the low-resolution face image to be processed and the images in the training library into square image blocks with overlapping parts by using the same block division method; the image blocks are square image blocks with a side length of psize.
本步骤中,将训练库中各图像均划分为N个正方形图像块;同时,将待处理低分辨率人脸图像xin也划分为N个图像块。采用图像块集表示相应的人脸图像,待估高分辨率人脸图像yout将通过对待处理低分辨率人脸图像xin的图像块恢复获得。将待处理低分辨率人脸图像xin、待估高分辨率人脸图像yout、训练库中低分辨率人脸图像X、训练库中高分辨率人脸图像Y的图像块集分别记为i表示图像块编号,Xi、Yi分别表示待处理低分辨率人脸图像xin、待估计高分辨率人脸图像yout、训练库中低分辨率人脸图像X、训练库中高分辨率人脸图像Y中的第i个图像块。In this step, each image in the training library is divided into N square image blocks; at the same time, the low-resolution face image x in to be processed is also divided into N image blocks. The image block set is used to represent the corresponding face image, and the high-resolution face image y out to be estimated will be obtained by recovering the image blocks of the low-resolution face image x in to be processed. The image block sets of the low-resolution face image x in to be processed, the high-resolution face image y out to be estimated, the low-resolution face image X in the training library, and the high-resolution face image Y in the training library are recorded as i represents the image block number, X i and Y i represent the low-resolution face image x in to be processed, the high-resolution face image y out to be estimated, the low-resolution face image X in the training library, and the high-resolution face image Y in the training library The i-th image block of .
见图2,对人脸图像进行分块的主要依据是局部流形的思想,即人脸图像是一类特殊图像,这些图像具有特定的结构意义,比如在某个位置上所有的小块都是眼睛、或者某个位置上都是鼻子,也就是说图像中每一个位置的局部小块都处于一个特定的局部几何流形当中。为保证这个局部流形,需要将图像分为若干正方形的图像块。图像块的大小需要有合适尺寸,若分块太大,则会由于微小的对齐问题引起重影现象;若分块太小,会模糊、淡化每个小块的位置特征。此外,还需要选择图像块之间交叠块的尺寸。因为如果简单的将图像分为不含交叠块的若干正方形小块,那么这些正方形块与块之间会因为不兼容问题出现网格效应。而且人脸图像并不总是正方形,那么交叠块的尺寸选择需要注意使得图像尽可能充分的分块。As shown in Figure 2, the main basis for dividing face images into blocks is the idea of local manifolds, that is, face images are a special type of image, and these images have specific structural meanings, such as all small blocks at a certain position. It is the eyes, or the nose at a certain position, that is to say, the local small blocks at each position in the image are in a specific local geometric manifold. In order to ensure this local manifold, the image needs to be divided into several square image blocks. The size of the image block needs to have an appropriate size. If the block is too large, it will cause ghosting due to slight alignment problems; if the block is too small, the positional characteristics of each small block will be blurred and diluted. In addition, the size of overlapping blocks between image blocks also needs to be selected. Because if the image is simply divided into several small square blocks without overlapping blocks, then there will be grid effects between these square blocks due to incompatibility problems. Moreover, face images are not always square, so the size selection of overlapping blocks needs to pay attention to make the image as fully divided as possible.
将图像块尺寸记为psize×psize,相邻图像块间交叠部分的宽记为D,将图像块所在位置表示为i,i=1,2,...N,则有: The size of the image block is recorded as psize×psize, the width of the overlapping part between adjacent image blocks is recorded as D, and the position of the image block is expressed as i, i=1,2,...N, then:
其中,height和width分别为人脸图像的高和宽。实施例中,psize取12,D取8。Among them, height and width are the height and width of the face image respectively. In the embodiment, psize is 12, and D is 8.
S3:对待处理低分辨率人脸图像每一块,在对应位置的低分辨率训练块集合中查找其近邻块;S3: For each block of the low-resolution face image to be processed, search for its neighbor blocks in the low-resolution training block set at the corresponding position;
对于待处理低分辨率图像xin,假设在位置i上的块为低分辨率图像库设为X,X上在位置i的所以图像块记为Xi。在Xi上的K个近邻块,通过和Xi的每一个图像块差值的绝对值一一对比获得,差值绝对值最小的K个低分辨率图像块,作为的的近邻,记为 For the low-resolution image x in to be processed, suppose the block at position i is The low-resolution image library is set to X, and the image block at position i on X is denoted as X i . The K nearest neighbor blocks on Xi, by Compared with the absolute value of each image block difference of X i one by one, the K low-resolution image blocks with the smallest absolute value of the difference are obtained as neighbors of
S4:计算待处理低分辨率人脸图像块,每一块到近邻的形变场矩阵;S4: Calculate the low-resolution face image block to be processed, and the deformation field matrix from each block to the neighbor;
S4中,每一个图像块的形变场矩阵[u,v]可通过待处理图像与之间通过以下方法计算得到:In S4, each image block The deformation field matrix [u, v] of the image to be processed can be and Calculated by the following method:
我们设p=(x,y)为图像上像素p的坐标,设w(p)=(u(p),v(p))是像素p点的流向量,我们只准许u(p)和v(p)是整数集。W(p)为在p的像素点的流向量,其中u(p)和v(p)分别代表该像素点在水平和垂直两个方向的位移场;让s1和s2表示与的SIFT特征。那么得到能量函数E(w):We set p=(x,y) as the coordinates of pixel p on the image, let w(p)=(u(p),v(p)) be the flow vector of pixel p, we only allow u(p) and v(p) is the set of integers. W(p) is the flow vector at the pixel point of p, where u(p) and v(p) represent the displacement field of the pixel point in the horizontal and vertical directions respectively; let s 1 and s 2 denote and SIFT features. Then get the energy function E(w):
其中,第一行是数据项,表示的是让SIFT特征沿着像素p点的流向量w(p)来匹配;第二项是位移项,表示的是当没有其他信息可利用时,让流向量w(p)尽可能的小;第三项是平滑项,让相邻像素相似。在这个目标函数中,L1准则(绝对值相加)被用于第一行数据项和第三行平滑项当中,来计算匹配离群和流向量不连续,t和d分别作为阈值;η、α是平衡参数,根据经验赋值。Among them, the first line is a data item, which means that the SIFT feature is matched along the flow vector w(p) of the pixel point p; the second item is a displacement item, which means that when no other information is available, let the flow direction The amount w(p) is as small as possible; the third term is a smooth term, which makes adjacent pixels similar. In this objective function, the L1 criterion (absolute value addition) is used in the first row of data items and the third row of smoothing items to calculate matching outliers and flow vector discontinuities, and t and d are used as thresholds respectively; η, α is a balance parameter, assigned according to experience.
通过最小化能量函数E(w),我们可以得到形变场矩阵[u,v]。By minimizing the energy function E(w), we can obtain the deformation field matrix [u,v].
S5:根据形变场矩阵,计算每一个近邻块到对应待处理低分辨率人脸图像块的形变块;S5: According to the deformation field matrix, calculate each adjacent block to the deformation block corresponding to the low-resolution face image block to be processed;
S5中,根据形变场矩阵,计算每一个近邻块到对应待处理低分辨率人脸图像块的形变块,具体过程为:In S5, according to the deformation field matrix, calculate each adjacent block to the deformation block corresponding to the low-resolution face image block to be processed, the specific process is:
我们通过引入[u,v],让匹配低分辨率图像块获得到样本库中低分辨率图像块的映射关系,将这个关系定义为特征算子sf(·),通过特征算子sf(·),我们可以将匹配到样本库图像块形变至与输入图像块相似,则可得形变场[u,v]的表达式:We introduce [u,v], let Match low-resolution image patches get to the low-resolution image patch in the sample library The mapping relationship is defined as the feature operator sf( ), through the feature operator sf( ), we can deform the image block matched to the sample library to be similar to the input image block, then the deformation field [u ,v] expression:
我们得到的形变矩阵[u,v],根据高、低分辨率的图像块在流形空间中具有相似的一致性,那么我们则可以通过形变矩阵[u,v],得到形变图像块Rk(i,j),我们定义Deform(·)表示将的形变过程,则The deformation matrix [u, v] we get, according to the high and low resolution image blocks have similar consistency in the manifold space, then we can get the deformation image block R k through the deformation matrix [u, v] (i,j), we define Deform( ) to represent the deformation process, then
对于训练样本库中的低分辨率图像块它形变后的图像块可表示为类似的,低分辨率图像块所对应的高分辨率图像块,它形变后的高分辨率图像块可表示为 For low-resolution image patches in the training sample library Its deformed image block can be expressed as Similarly, for the high-resolution image block corresponding to the low-resolution image block, the deformed high-resolution image block can be expressed as
因此,本文引入SIFT Flow特征算子sf(·),得到形变矩阵[u,v]。利用形变矩阵[u,v]形变图像块,可以得到与输入图像块更相似的形变图像块,使得样本库图像块的表达更加准确,增强样本库中样本的表达能力。Therefore, this paper introduces the SIFT Flow feature operator sf( ) to obtain the deformation matrix [u,v]. Using the deformation matrix [u, v] to deform the image block, a deformed image block that is more similar to the input image block can be obtained, which makes the expression of the image block in the sample library more accurate and enhances the expressive ability of the samples in the sample library.
S6:计算待处理图像块和近邻形变块之间的权重系数;具体的是低分辨率人脸图像的每一个图像块的每一个近邻图像块的权重系数S6: Calculate the weight coefficient between the image block to be processed and the adjacent deformation block; specifically, the weight coefficient of each adjacent image block of each image block of the low-resolution face image
步骤S6中,计算待处理图像块和近邻形变块之间的权重系数,具体过程如下:In step S6, the weight coefficient between the image block to be processed and the adjacent deformed block is calculated, and the specific process is as follows:
首先引入欧氏距离dk(i,j):First introduce the Euclidean distance d k (i,j):
即是输入低分辨率率图像块与训练样本中的低分辨率图像块的欧式距离。然后选取欧式距离近的K个高分辨率图像块进行形变,因此此时形变图像块可表示为:That is, the input low-resolution image block with the low-resolution image patches in the training samples Euclidean distance of . Then select K high-resolution image blocks with the closest Euclidean distance for deformation, so the deformed image block can be expressed as:
输入的低分辨率人脸图像块由样本训练集中K个距离最近的低分辨率形变图像块表示的最优表示权重w*(i,j)为:Input low-resolution face image patch From the sample training set K nearest low-resolution deformed image blocks The optimal representation weight w * (i,j) expressed is:
其中,是用样本库中的低分辨率形变图像块线性表示输入的低分辨率图像块的权重系数;wi是元素为的向量;τ是平衡参数,根据经验赋值。in, is to use the low-resolution warped image patch from the sample library Linearly represents the weight coefficient of the input low-resolution image block; w i is the element The vector; τ is a balance parameter, assigned according to experience.
S7:将权重投影到高分辨率空间上,根据重建系数恢复图像块获得其对应的高分辨率人脸图像块 S7: Project the weights onto the high-resolution space and restore the image blocks according to the reconstruction coefficients Obtain its corresponding high-resolution face image block
步骤S7中,并将权重投影到高分辨率空间上,根据重建系数恢复图像块获得其对应的高分辨率人脸图像块具体过程为:In step S7, the weights are projected onto the high-resolution space, and the image blocks are restored according to the reconstruction coefficients Obtain its corresponding high-resolution face image block The specific process is:
对于重构出的目标高分辨率图像块它可由最近邻K个高分辨率形变图像块与它最优权重合成:For the reconstructed target high-resolution image block It can be deformed by K nearest neighbor high-resolution image patches with its optimal weight synthesis:
其中, 为对应的高分辨率样本。。in, for Corresponding high-resolution samples. .
S8:拼接高分辨率人脸图像块得高分辨率人脸图像。S8: Stitching high-resolution face image blocks high-resolution face images.
为验证本发明技术效果,使用中国人脸数据库CAS-PEAL进行验证。从中选择1040个人脸样本,分辨率是112*96,用仿射变换法对齐人脸。从人脸样本中选取40幅图像下采样4倍(分辨率为24*28)后加上0.015的高斯噪声后作为测试图像。将人脸样本剩余图像作为训练库,分别采用传统局部脸人脸超分辨率方法(方法1)、.数据驱动局部特征转换的噪声人脸幻构方法(方法2)、基于轮廓先验的鲁棒性人脸超分辨率处理方法(方法3)得到主观图像。In order to verify the technical effect of the present invention, the Chinese face database CAS-PEAL is used for verification. Select 1040 face samples from them, the resolution is 112*96, and use the affine transformation method to align the faces. Select 40 images from the face samples and downsample them by 4 times (resolution 24*28) and add Gaussian noise of 0.015 as test images. Using the remaining images of face samples as the training library, the traditional local face super-resolution method (method 1), the noise face illusion method based on data-driven local feature conversion (method 2), and the contour prior-based robust Sticky face super-resolution processing method (method 3) obtains subjective images.
从实验结果可知,方法1~3虽然比插值方法在分辨率上有所提升,但出现了较严重误差,与原始图像的相似度很低。方法2中的结果由于是全局脸架构,基于全局的方法往往具有细节恢复上的短板,所以在这方面稍逊于本发明方法。本发明方法所恢复图像的质量相比于方法1~3和双三次插值方法都有显著提高。It can be seen from the experimental results that although the resolution of methods 1 to 3 is improved compared with the interpolation method, there are serious errors and the similarity with the original image is very low. The result in method 2 is a global face architecture, and the global-based method often has shortcomings in detail restoration, so it is slightly inferior to the method of the present invention in this respect. Compared with the methods 1-3 and the bicubic interpolation method, the quality of the image restored by the method of the present invention is significantly improved.
表1展示了各图像对应的客观质量,包括PSNR(峰值信噪比)和SSIM值(结构相似性准则)。从表1中可以看出,本发明方法在恢复图像的客观质量上,也有较为明显的稳定提升。Table 1 shows the corresponding objective quality of each image, including PSNR (peak signal-to-noise ratio) and SSIM value (structural similarity criterion). It can be seen from Table 1 that the method of the present invention also has a relatively obvious and stable improvement in the objective quality of the restored image.
表1恢复图像客观质量的对比Table 1 Comparison of objective quality of recovered images
本发明方法通过从原始低分辨率人脸图像中自动提取的大尺度边缘数据与原始尺度的图像特征进行组合,对低质量人脸图像进行恢复。实验结果从主观质量到客观质量均证明了本发明的有效性,即边缘数据的引入有效减弱了严重噪声对超分辨率重建的影响,自动提取的特征避免了人工干预带来的负面效果(如处理结果不稳定、不精确等问题),从而提升了人脸超分辨率处理结果。The method of the invention restores the low-quality human face image by combining the large-scale edge data automatically extracted from the original low-resolution human face image with the image features of the original scale. The experimental results have proved the effectiveness of the present invention from subjective quality to objective quality, that is, the introduction of edge data effectively weakens the influence of severe noise on super-resolution reconstruction, and the automatically extracted features avoid the negative effects caused by manual intervention (such as The processing results are unstable, inaccurate, etc.), thus improving the results of face super-resolution processing.
本文中所描述的具体实施例仅仅是对本发明精神作举例说明。本发明所属技术领域的技术人员可以对所描述的具体实施例做各种各样的修改或补充或采用类似的方式替代,但并不会偏离本发明的精神或者超越所附权利要求书所定义的范围。The specific embodiments described herein are merely illustrative of the spirit of the invention. Those skilled in the art to which the present invention belongs can make various modifications or supplements to the described specific embodiments or adopt similar methods to replace them, but they will not deviate from the spirit of the present invention or go beyond the definition of the appended claims range.
Claims (9)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610494253.8A CN106203269A (en) | 2016-06-29 | 2016-06-29 | A kind of based on can the human face super-resolution processing method of deformation localized mass and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610494253.8A CN106203269A (en) | 2016-06-29 | 2016-06-29 | A kind of based on can the human face super-resolution processing method of deformation localized mass and system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106203269A true CN106203269A (en) | 2016-12-07 |
Family
ID=57462572
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610494253.8A Pending CN106203269A (en) | 2016-06-29 | 2016-06-29 | A kind of based on can the human face super-resolution processing method of deformation localized mass and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106203269A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107194872A (en) * | 2017-05-02 | 2017-09-22 | 武汉大学 | Remote sensed image super-resolution reconstruction method based on perception of content deep learning network |
CN107545567A (en) * | 2017-07-31 | 2018-01-05 | 中国科学院自动化研究所 | The method for registering and device of biological tissue's sequence section micro-image |
CN107705249A (en) * | 2017-07-19 | 2018-02-16 | 苏州闻捷传感技术有限公司 | Image super-resolution method based on depth measure study |
CN109035144A (en) * | 2018-07-19 | 2018-12-18 | 广东工业大学 | super-resolution image construction method |
CN112991191A (en) * | 2019-12-13 | 2021-06-18 | 北京金山云网络技术有限公司 | Face image enhancement method and device and electronic equipment |
CN114970801A (en) * | 2021-02-23 | 2022-08-30 | 武汉Tcl集团工业研究院有限公司 | Video super-resolution method and device, terminal equipment and storage medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102243711A (en) * | 2011-06-24 | 2011-11-16 | 南京航空航天大学 | Neighbor embedding-based image super-resolution reconstruction method |
US8139899B2 (en) * | 2007-10-24 | 2012-03-20 | Motorola Mobility, Inc. | Increasing resolution of video images |
CN102521810A (en) * | 2011-12-16 | 2012-06-27 | 武汉大学 | Face super-resolution reconstruction method based on local constraint representation |
CN103824272A (en) * | 2014-03-03 | 2014-05-28 | 武汉大学 | Face super-resolution reconstruction method based on K-neighboring re-recognition |
CN104091320A (en) * | 2014-07-16 | 2014-10-08 | 武汉大学 | Noise human face super-resolution reconstruction method based on data-driven local feature conversion |
CN105701515A (en) * | 2016-01-18 | 2016-06-22 | 武汉大学 | Face super-resolution processing method and system based on double-layer manifold constraint |
-
2016
- 2016-06-29 CN CN201610494253.8A patent/CN106203269A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8139899B2 (en) * | 2007-10-24 | 2012-03-20 | Motorola Mobility, Inc. | Increasing resolution of video images |
CN102243711A (en) * | 2011-06-24 | 2011-11-16 | 南京航空航天大学 | Neighbor embedding-based image super-resolution reconstruction method |
CN102521810A (en) * | 2011-12-16 | 2012-06-27 | 武汉大学 | Face super-resolution reconstruction method based on local constraint representation |
CN103824272A (en) * | 2014-03-03 | 2014-05-28 | 武汉大学 | Face super-resolution reconstruction method based on K-neighboring re-recognition |
CN104091320A (en) * | 2014-07-16 | 2014-10-08 | 武汉大学 | Noise human face super-resolution reconstruction method based on data-driven local feature conversion |
CN105701515A (en) * | 2016-01-18 | 2016-06-22 | 武汉大学 | Face super-resolution processing method and system based on double-layer manifold constraint |
Non-Patent Citations (3)
Title |
---|
CE LIU等: "SIFT Flow: Dense Correspondence across Scenes and Its Applications", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 * |
兰诚栋等: "采用后验信息构建稀疏原子库的超分辨率人脸重建", 《北京工业大学学报》 * |
王婷婷等: "基于分块的肺 4D-CT 图像超分辨率重建", 《中国生物医学工程学报》 * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107194872A (en) * | 2017-05-02 | 2017-09-22 | 武汉大学 | Remote sensed image super-resolution reconstruction method based on perception of content deep learning network |
CN107194872B (en) * | 2017-05-02 | 2019-08-20 | 武汉大学 | Super-resolution reconstruction method of remote sensing images based on content-aware deep learning network |
CN107705249A (en) * | 2017-07-19 | 2018-02-16 | 苏州闻捷传感技术有限公司 | Image super-resolution method based on depth measure study |
CN107545567A (en) * | 2017-07-31 | 2018-01-05 | 中国科学院自动化研究所 | The method for registering and device of biological tissue's sequence section micro-image |
CN107545567B (en) * | 2017-07-31 | 2020-05-19 | 中国科学院自动化研究所 | Registration method and device for biological tissue sequence section microscopic image |
CN109035144A (en) * | 2018-07-19 | 2018-12-18 | 广东工业大学 | super-resolution image construction method |
CN109035144B (en) * | 2018-07-19 | 2023-03-17 | 广东工业大学 | Super-resolution image construction method |
CN112991191A (en) * | 2019-12-13 | 2021-06-18 | 北京金山云网络技术有限公司 | Face image enhancement method and device and electronic equipment |
CN114970801A (en) * | 2021-02-23 | 2022-08-30 | 武汉Tcl集团工业研究院有限公司 | Video super-resolution method and device, terminal equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110570353B (en) | Densely connected generative adversarial network single image super-resolution reconstruction method | |
CN110689482B (en) | Face super-resolution method based on supervised pixel-by-pixel generation countermeasure network | |
CN107423698B (en) | A Gesture Estimation Method Based on Parallel Convolutional Neural Network | |
CN110706157A (en) | A face super-resolution reconstruction method based on identity prior generative adversarial network | |
CN106203269A (en) | A kind of based on can the human face super-resolution processing method of deformation localized mass and system | |
CN110490796B (en) | A face super-resolution processing method and system based on fusion of high and low frequency components | |
CN101976435A (en) | Combination learning super-resolution method based on dual constraint | |
CN103020898B (en) | Sequence iris image super resolution ratio reconstruction method | |
CN110310228A (en) | A face super-resolution processing method and system based on closed-link data re-expression | |
CN102982520A (en) | Robustness face super-resolution processing method based on contour inspection | |
CN105701770B (en) | A kind of human face super-resolution processing method and system based on context linear model | |
CN108280804A (en) | A kind of multi-frame image super-resolution reconstruction method | |
CN105335930B (en) | The robustness human face super-resolution processing method and system of edge data driving | |
CN112785629A (en) | Aurora motion characterization method based on unsupervised deep optical flow network | |
CN116664952A (en) | Image direction identification method integrating convolution and ViT | |
CN116563916A (en) | Attention fusion-based cyclic face super-resolution method and system | |
CN114663880A (en) | Three-dimensional target detection method based on multi-level cross-modal self-attention mechanism | |
Zhang et al. | Catmullrom splines-based regression for image forgery localization | |
CN105701515A (en) | Face super-resolution processing method and system based on double-layer manifold constraint | |
CN118154984A (en) | Unsupervised neighborhood classification superpixel generation method and system based on fusion guided filtering | |
CN116863285A (en) | Infrared and visible light image fusion method of multi-scale generative adversarial network | |
CN115588220B (en) | Two-stage multi-scale adaptive low-resolution face recognition method and its application | |
Luanyuan et al. | Mgnet: Learning correspondences via multiple graphs | |
CN114612798B (en) | Satellite image tampering detection method based on Flow model | |
CN109934193B (en) | Global context prior constraint anti-occlusion face super-resolution method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20161207 |