[go: up one dir, main page]

CN114003878A - Double-factor authentication method, device and system based on voice recognition - Google Patents

Double-factor authentication method, device and system based on voice recognition Download PDF

Info

Publication number
CN114003878A
CN114003878A CN202111241027.6A CN202111241027A CN114003878A CN 114003878 A CN114003878 A CN 114003878A CN 202111241027 A CN202111241027 A CN 202111241027A CN 114003878 A CN114003878 A CN 114003878A
Authority
CN
China
Prior art keywords
layer
user
spectrogram
voice
voice data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111241027.6A
Other languages
Chinese (zh)
Inventor
潘超
杨立宾
曾强
王志男
王磊
姜海阔
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guodian Nanjing Automation Co Ltd
Original Assignee
Guodian Nanjing Automation Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guodian Nanjing Automation Co Ltd filed Critical Guodian Nanjing Automation Co Ltd
Priority to CN202111241027.6A priority Critical patent/CN114003878A/en
Publication of CN114003878A publication Critical patent/CN114003878A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • G06F21/32User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/08Network architectures or network communication protocols for network security for authentication of entities
    • H04L63/083Network architectures or network communication protocols for network security for authentication of entities using passwords
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/08Network architectures or network communication protocols for network security for authentication of entities
    • H04L63/0861Network architectures or network communication protocols for network security for authentication of entities using biometrical features, e.g. fingerprint, retina-scan

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Security & Cryptography (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

本发明公开了一种基于语音识别的双因子认证方法、装置及系统,所述方法包括获取用户输入的语音数据;调用预设的语音识别模型,判断用户输入的语音数据是否为已定义词语,以及是否为目标客户的语音;获取用户输入的密码;调用密码识别模型,判断用户输入的密码是否满足预设的要求;当用户输入的语音数据为已定义词语,且为目标客户的语音;同时满足用户输入的密码满足预设的要求,则完成认证。本发明可以识别已定义的特定词语,通过对语音特征提取分类,区分是系统包含的语音或未包含的未知语音,从而对特定用户的语音进行身份认证。

Figure 202111241027

The invention discloses a voice recognition-based two-factor authentication method, device and system. The method includes acquiring voice data input by a user; calling a preset voice recognition model to determine whether the voice data input by the user is a defined word, and whether it is the voice of the target customer; obtain the password input by the user; call the password recognition model to determine whether the password input by the user meets the preset requirements; when the voice data input by the user is a defined word and is the voice of the target customer; If the password entered by the user meets the preset requirements, the authentication is completed. The present invention can identify the defined specific words, and by extracting and classifying the speech features, distinguish the speech included in the system or the unknown speech not included, so as to perform identity authentication on the speech of a specific user.

Figure 202111241027

Description

Double-factor authentication method, device and system based on voice recognition
Technical Field
The invention belongs to the technical field of voice recognition, and particularly relates to a voice recognition-based double-factor authentication method, device and system.
Background
Two-factor authentication (TFA) refers to the use of Two different types of authentication to protect a user's account, i.e., Two of the following three elements must be provided:
1. some secret known to the person being authenticated, such as: password information or answers to security questions;
2. something that the authenticated person owns, such as: holding a smartphone or other physical device;
3. the characteristics inherent in the person being authenticated, such as: fingerprint, iris, or voice.
Authentication by means of any two combinations of the above may be referred to as two-factor authentication. The currently common two-factor authentication implementation types are mainly the following six types:
mode 1: safety issues; in the method, when the account is created, one or more security questions can be selected, and corresponding answers are set for each question in advance. Subsequently, when logging in using the account, correct answers must be provided for each question, and the party can complete authentication and access authorization.
Mode 2: short messages or e-mails; when an account is created, a certain commonly used cell phone number is provided. When logging in again, the system sends the verification code by short message or email. This is a temporary verification code that expires in a short amount of time (e.g., 1-5 minutes). Therefore, it must be entered and submitted in a timely manner to complete the login.
Mode 3: a Time-based One-Time Password (OTP); if conditional, an identity authentication application may be used to scan a two-dimensional code containing a secret key. Accordingly, the key is loaded into the application, a temporary password that changes periodically is generated, and authentication required for login can be completed by inputting the password into the page.
Mode 4: a U2F key; universal two Factor (U2 nd Factor, U2F) is an open standard that can be used in USB devices, NFC devices, and smart cards. In the process of identity authentication, only a USB secret key needs to be inserted, and the NFC equipment is touched or the smart card is slid.
Mode 5: pushing a notification; on some platforms, after entering the password, a push message containing a notification of the relevant login attempt is received on the device. The request can be responded to simply by clicking on the device for approval or denial.
Mode 6: biological recognition; face recognition, voice recognition and fingerprint scanning are all types of biometric identification. They are often deployed in certain areas where security checks are required to determine the true identity of a guest.
The performance of a speech recognition system depends to a large extent on the speech representation, whose quality can be influenced by many variations, such as pitch, loudness and timbre. It makes sense to design a system that is sufficiently discriminative to recognize distinct speech. The voice recognition is mainly divided into text-independent and text-dependent, and a text-independent recognition system can recognize the voiceprint by only collecting a relatively short section of voice, but from the perspective of user experience and cost, a text-independent model often needs several hours or more of training data of a target user.
Disclosure of Invention
In view of the above problems, the present invention provides a method, an apparatus, and a system for dual-factor authentication based on speech recognition, which can recognize a defined specific word, and distinguish a speech included in the system or an unknown speech not included in the system by extracting and classifying speech features, thereby performing identity authentication on a speech of a specific user.
In order to achieve the technical purpose and achieve the technical effects, the invention is realized by the following technical scheme:
in a first aspect, the present invention provides a two-factor authentication method based on voice recognition, including:
acquiring voice data input by a user;
calling a preset voice recognition model, and judging whether voice data input by a user is a defined word or not and whether the voice data is the voice of a target client or not;
acquiring a password input by a user;
calling a password identification model, and judging whether a password input by a user meets a preset requirement;
when the voice data input by the user is the defined words and is the voice of the target customer; and if the password input by the user meets the preset requirement, the authentication is finished.
Optionally, the speech recognition model is obtained by:
collecting voice data of appointed words and non-appointed words to form a first sample set, extracting features, and converting the sample set into a first spectrogram;
collecting appointed words of a target user and appointed word voice data of a non-target user to form a second sample set, extracting characteristics, and converting the second sample set into a second spectrogram;
respectively sending the first spectrogram and the second spectrogram into a preset convolutional neural network model, wherein the convolutional neural network model comprises a first convolutional layer, a first pooling layer, a second convolutional layer, a second pooling layer, a first full-connection layer, a Dropout layer and a second full-connection layer which are sequentially connected; the first spectrogram and the second spectrogram are processed by a first convolution layer, a first pooling layer, a second convolution layer and a second pooling layer and then are sent to a first full-connection layer;
and continuously adjusting parameters of the convolutional neural network model to generate a final voice recognition model.
Optionally, the first convolution layer and the second convolution layer both use a 3 × 3 convolution kernel as the feature extraction filter.
Optionally, the first pooling layer and the second pooling layer both adopt a maximum value calculation method.
In a second aspect, the present invention provides a two-factor authentication device based on voice recognition, including:
the first acquisition module is used for acquiring voice data input by a user;
the first judgment module is used for calling a preset voice recognition model, judging whether voice data input by a user is a defined word or not and judging whether the voice data is the voice of a target client or not;
the second acquisition module is used for acquiring the password input by the user;
the second judgment module calls the password identification model and judges whether the password input by the user meets the preset requirement or not;
the authentication module is used for determining whether the voice data input by the user is the defined word or not and is the voice of the target client; and if the password input by the user meets the preset requirement, the authentication is finished.
Optionally, the speech recognition model is obtained by:
collecting voice data of appointed words and non-appointed words to form a first sample set, extracting features, and converting the sample set into a first spectrogram;
collecting appointed words of a target user and appointed word voice data of a non-target user to form a second sample set, extracting characteristics, and converting the second sample set into a second spectrogram;
respectively sending the first spectrogram and the second spectrogram into a preset convolutional neural network model, wherein the convolutional neural network model comprises a first convolutional layer, a first pooling layer, a second convolutional layer, a second pooling layer, a first full-connection layer, a Dropout layer and a second full-connection layer which are sequentially connected; the first spectrogram and the second spectrogram are processed by a first convolution layer, a first pooling layer, a second convolution layer and a second pooling layer and then are sent to a first full-connection layer;
and continuously adjusting parameters of the convolutional neural network model to generate a final voice recognition model.
Optionally, the first convolution layer and the second convolution layer both use a 3 × 3 convolution kernel as the feature extraction filter.
Optionally, the first pooling layer and the second pooling layer both adopt a maximum value calculation method.
In a third aspect, the present invention provides a two-factor authentication method based on voice recognition, including:
a processor;
a memory having stored thereon a computer program operable on the processor;
wherein the computer program realizes the method according to any one of the first aspect when executed by the processor.
Compared with the prior art, the invention has the beneficial effects that:
in the invention, in the process of logging in the intranet terminal host of the power system by a user, the identity authentication is safer by using the password and voice double factors, the biological identification technology is extremely difficult to break, and the biological characteristics are basically lifelong for adults compared with the changeable and modified digital passwords. The ability to resist abnormal login attacks is more secure and trained only for specific persons.
The invention adopts the text-related recognition, reduces the complexity of the system and the training data of the target user by limiting the selectable dictionary information, thereby reducing the cost and improving the efficiency.
The invention adopts the multilayer convolutional neural network, solves the problems that in the training process of the traditional feedforward neural network structure, weight distribution depends on the occurrence frequency of samples, the samples need to be provided are all the more, otherwise, the samples are easy to be unidentified, and has better matching rate only aiming at the training of specific people.
Drawings
In order that the present disclosure may be more readily and clearly understood, reference is now made to the following detailed description of the present disclosure taken in conjunction with the accompanying drawings, in which:
FIG. 1 is a flow chart of a method for two-factor authentication based on voice recognition according to an embodiment of the present invention;
FIG. 2 is a flow chart of a convolutional neural network construction according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail with reference to the following embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and do not limit the scope of the invention.
The following detailed description of the principles of the invention is provided in connection with the accompanying drawings.
Example 1
The embodiment of the invention provides a double-factor authentication method based on voice recognition, which comprises the following steps:
acquiring voice data input by a user;
calling a preset voice recognition model, and judging whether voice data input by a user is a defined word or not and whether the voice data is the voice of a target client or not;
acquiring a password input by a user;
calling a password identification model, and judging whether a password input by a user meets a preset requirement;
when the voice data input by the user is the defined words and is the voice of the target customer; and if the password input by the user meets the preset requirement, the authentication is finished.
In a specific implementation manner of the embodiment of the present invention, as shown in fig. 1, the speech recognition model is obtained by the following method:
collecting appointed voice materials, namely collecting appointed words and non-appointed word voice data to form a first sample set, extracting characteristics and converting the sample set into a first spectrogram; collecting appointed words of a target user and appointed word voice data of a non-target user to form a second sample set, extracting features, converting the sample set into a second spectrogram, and converting a oscillogram into the spectrogram; in the specific implementation process, the audio files of the WAV codes are uniformly adopted, the number of sound channels is single channel, the sampling frequency is 16kHz, the sample data obtained by sampling each time is 16 bits, the average number of bytes per second is the sampling frequency channel number, the sample data bits of sampling each time is divided by 8 to be 32000, and the collected WAV coded audio is imported through an API of TensorFlow; the imported WAV waveforms are converted into spectrograms. The invention adopts STFT short-time Fourier transform to convert audio frequency into time-frequency domain, which can be expressed as two-dimensional image and output two-dimensional tensor; and carrying out manual classification, attaching labels, forming a combination of a waveform data set, a spectrogram image and corresponding labels thereof, and realizing the classification of the voice files through the relationship between the frequency domain image and the audio content.
Respectively sending the first spectrogram and the second spectrogram into a preset convolutional neural network model, wherein the convolutional neural network model comprises a first convolutional layer, a first pooling layer, a second convolutional layer, a second pooling layer, a first full-connection layer, a Dropout layer and a second full-connection layer which are sequentially connected; the first spectrogram and the second spectrogram are processed by a first convolution layer, a first pooling layer, a second convolution layer and a second pooling layer and then are sent to a first full-connection layer, specifically referring to fig. 2; in a specific implementation, the first fully-connected layer and the second fully-connected layer each include 1024 neurons, and the first fully-connected layer changes the output into one dimension to correspond to the tag. And the Dropout layer and the second fully-connected layer are used for filtering output information, so that repeated data is prevented from causing 'overfitting' of the result. And finally, training the configured model script by using TensorFlow to complete model construction, and obtaining the voice recognition model.
Continuously adjusting parameters of the convolutional neural network model until the model test accuracy reaches the standard, and generating a final voice recognition model; in the specific implementation process, the method specifically comprises the following steps: and compiling the model, and observing the change process of the loss value loss of the training set and the loss value val _ loss of the test set. The loss of the training set is continuously reduced, and the loss of the testing set is continuously reduced, which shows that the network is still learning. And finally, accuracuracy accuracy tends to be stable, and a voice recognition model is generated.
In a specific implementation manner of the embodiment of the present invention, each of the first convolution layer and the second convolution layer uses a 3 × 3 convolution kernel as a feature extraction filter; the first pooling layer and the second pooling layer both adopt a mode of solving the maximum value, the purpose of the pooling layers is to reduce unnecessary redundant information, the result of re-pooling of the first convolution is used as the input of the second convolution, and specific words which accord with the target user are screened out.
The following describes the speech recognition processing procedure in the embodiment of the present invention in detail with reference to fig. 1.
Starting user operation;
initializing a system recording function, and recording a user instruction to complete the acquisition of voice data for input;
and calling a preset voice recognition model, judging whether the voice data input by the user is a defined word or not and whether the voice data is the voice of the target client or not, and outputting a recognition result.
Example 2
Based on the same inventive concept as embodiment 1, an embodiment of the present invention provides a two-factor authentication apparatus based on voice recognition, including:
the first acquisition module is used for acquiring voice data input by a user;
the first judgment module is used for calling a preset voice recognition model, judging whether voice data input by a user is a defined word or not and judging whether the voice data is the voice of a target client or not;
the second acquisition module is used for acquiring the password input by the user;
the second judgment module calls the password identification model and judges whether the password input by the user meets the preset requirement or not;
the authentication module is used for determining whether the voice data input by the user is the defined word or not and is the voice of the target client; and if the password input by the user meets the preset requirement, the authentication is finished.
In a specific implementation manner of the embodiment of the present invention, the speech recognition model is obtained by the following method:
collecting voice data of appointed words and non-appointed words to form a first sample set, extracting features, and converting the sample set into a first spectrogram;
collecting appointed words of a target user and appointed word voice data of a non-target user to form a second sample set, extracting characteristics, and converting the second sample set into a second spectrogram;
respectively sending the first spectrogram and the second spectrogram into a preset convolutional neural network model, wherein the convolutional neural network model comprises a first convolutional layer, a first pooling layer, a second convolutional layer, a second pooling layer, a first full-connection layer, a Dropout layer and a second full-connection layer which are sequentially connected; the first spectrogram and the second spectrogram are processed by a first convolution layer, a first pooling layer, a second convolution layer and a second pooling layer and then are sent to a first full-connection layer;
and continuously adjusting parameters of the convolutional neural network model to generate a final voice recognition model.
In a specific implementation manner of the embodiment of the present invention, each of the first convolution layer and the second convolution layer uses a 3 × 3 convolution kernel as a feature extraction filter.
In a specific implementation manner of the embodiment of the present invention, the first pooling layer and the second pooling layer both adopt a maximum value calculation method.
Example 3
Based on the same inventive concept as embodiment 1, the embodiment of the present invention provides a two-factor authentication method based on voice recognition, including:
a processor;
a memory having stored thereon a computer program operable on the processor;
wherein the computer program realizes the method according to any one of embodiment 1 when being executed by the processor
The foregoing shows and describes the general principles and broad features of the present invention and advantages thereof. It will be understood by those skilled in the art that the present invention is not limited to the embodiments described above, which are described in the specification and illustrated only to illustrate the principle of the present invention, but that various changes and modifications may be made therein without departing from the spirit and scope of the present invention, which fall within the scope of the invention as claimed. The scope of the invention is defined by the appended claims and equivalents thereof.

Claims (9)

1.一种基于语音识别的双因子认证方法,其特征在于,包括:1. a two-factor authentication method based on speech recognition, is characterized in that, comprises: 获取用户输入的语音数据;Obtain the voice data input by the user; 调用预设的语音识别模型,判断用户输入的语音数据是否为已定义词语,以及是否为目标客户的语音;Call the preset voice recognition model to determine whether the voice data input by the user is a defined word and whether it is the voice of the target customer; 获取用户输入的密码;Get the password entered by the user; 调用密码识别模型,判断用户输入的密码是否满足预设的要求;Call the password recognition model to determine whether the password entered by the user meets the preset requirements; 当用户输入的语音数据为已定义词语,且为目标客户的语音;同时满足用户输入的密码满足预设的要求,则完成认证。When the voice data input by the user is a defined word and is the voice of the target customer; and the password input by the user meets the preset requirements, the authentication is completed. 2.根据权利要求1所述的一种基于语音识别的双因子认证方法,其特征在于:所述语音识别模型通过以下方法获得:2. a kind of two-factor authentication method based on speech recognition according to claim 1, is characterized in that: described speech recognition model is obtained by the following method: 采集指定词语和非指定词语语音数据,形成第一个样本集,并进行特征提取,转换为第一频谱图;Collect the voice data of specified words and non-specified words, form the first sample set, and perform feature extraction to convert it into the first spectrogram; 采集目标用户的指定词语和非目标用户的指定词语语音数据,形成第二个样本集,并进行特征提取,转换为第二频谱图;Collect the voice data of the specified words of the target user and the specified words of the non-target users, form a second sample set, and perform feature extraction, and convert it into a second spectrogram; 将所述第一频谱图和第二频谱图分别送入预设的卷积神经网络模型,所述卷积神经网络模型包括顺次相连的第一卷积层、第一池化层、第二卷积层、第二池化层、第一全连接层、Dropout层和第二全连接层;所述第一频谱图和第二频谱图均经第一卷积层、第一池化层、第二卷积层、第二池化层处理后送入第一全连接层;The first spectrogram and the second spectrogram are respectively sent to a preset convolutional neural network model, and the convolutional neural network model includes a first convolutional layer, a first pooling layer, and a second convolutional layer that are connected in sequence. Convolution layer, second pooling layer, first fully connected layer, Dropout layer and second fully connected layer; the first spectrogram and the second spectrogram are processed by the first convolution layer, the first pooling layer, The second convolution layer and the second pooling layer are processed and sent to the first fully connected layer; 不断调节卷积神经网络模型的参数,生成最终的语音识别模型。The parameters of the convolutional neural network model are continuously adjusted to generate the final speech recognition model. 3.根据权利要求2所述的一种基于语音识别的双因子认证方法,其特征在于:所述第一卷积层和第二卷积层均采用3*3卷积核作为特征提取滤波器。3. a kind of two-factor authentication method based on speech recognition according to claim 2, is characterized in that: described first convolution layer and second convolution layer all adopt 3*3 convolution kernel as feature extraction filter . 4.根据权利要求2所述的一种基于语音识别的双因子认证方法,其特征在于:所述第一池化层和第二池化层均采用求最大值的方式。4 . The two-factor authentication method based on speech recognition according to claim 2 , wherein the first pooling layer and the second pooling layer both adopt the method of finding the maximum value. 5 . 5.一种基于语音识别的双因子认证装置,其特征在于,包括:5. A two-factor authentication device based on voice recognition, characterized in that, comprising: 第一获取模块,用于获取用户输入的语音数据;The first acquisition module is used for acquiring the voice data input by the user; 第一判断模块,用于调用预设的语音识别模型,判断用户输入的语音数据是否为已定义词语,以及是否为目标客户的语音;The first judgment module is used to call a preset voice recognition model to judge whether the voice data input by the user is a defined word, and whether it is the voice of the target customer; 第二获取模块,获取用户输入的密码;The second obtaining module obtains the password entered by the user; 第二判断模块,调用密码识别模型,判断用户输入的密码是否满足预设的要求;The second judgment module calls the password recognition model to judge whether the password input by the user meets the preset requirements; 认证模块,用于当用户输入的语音数据为已定义词语,且为目标客户的语音;同时满足用户输入的密码满足预设的要求,则完成认证。The authentication module is used to complete the authentication when the voice data input by the user is a defined word and is the voice of the target customer; at the same time, the password input by the user meets the preset requirements. 6.根据权利要求5所述的一种基于语音识别的双因子认证方法,其特征在于:所述语音识别模型通过以下方法获得:6. a kind of two-factor authentication method based on speech recognition according to claim 5, is characterized in that: described speech recognition model is obtained by the following method: 采集指定词语和非指定词语语音数据,形成第一个样本集,并进行特征提取,转换为第一频谱图;Collect the voice data of specified words and non-specified words, form the first sample set, and perform feature extraction to convert it into the first spectrogram; 采集目标用户的指定词语和非目标用户的指定词语语音数据,形成第二个样本集,并进行特征提取,转换为第二频谱图;Collect the specified words of the target user and the voice data of the specified words of the non-target users, form a second sample set, perform feature extraction, and convert it into a second spectrogram; 将所述第一频谱图和第二频谱图分别送入预设的卷积神经网络模型,所述卷积神经网络模型包括顺次相连的第一卷积层、第一池化层、第二卷积层、第二池化层、第一全连接层、Dropout层和第二全连接层;所述第一频谱图和第二频谱图均经第一卷积层、第一池化层、第二卷积层、第二池化层处理后送入第一全连接层;The first spectrogram and the second spectrogram are respectively sent to a preset convolutional neural network model, and the convolutional neural network model includes a first convolutional layer, a first pooling layer, and a second convolutional layer that are connected in sequence. The convolution layer, the second pooling layer, the first fully connected layer, the Dropout layer and the second fully connected layer; the first spectrogram and the second spectrogram are processed by the first convolution layer, the first pooling layer, The second convolution layer and the second pooling layer are processed and sent to the first fully connected layer; 不断调节卷积神经网络模型的参数,生成最终的语音识别模型。The parameters of the convolutional neural network model are continuously adjusted to generate the final speech recognition model. 7.根据权利要求6所述的一种基于语音识别的双因子认证方法,其特征在于:所述第一卷积层和第二卷积层均采用3*3卷积核作为特征提取滤波器。7. a kind of two-factor authentication method based on speech recognition according to claim 6, is characterized in that: described first convolution layer and second convolution layer all adopt 3*3 convolution kernel as feature extraction filter . 8.根据权利要求6所述的一种基于语音识别的双因子认证方法,其特征在于:所述第一池化层和第二池化层均采用求最大值的方式。8 . The two-factor authentication method based on speech recognition according to claim 6 , wherein the first pooling layer and the second pooling layer both adopt the method of finding the maximum value. 9 . 9.一种基于语音识别的双因子认证方法,其特征在于,包括:9. A two-factor authentication method based on speech recognition, characterized in that, comprising: 处理器;processor; 存储器,其上存储有可在所述处理器上运行的计算机程序;a memory having stored thereon a computer program executable on the processor; 其中,所述计算机程序被所述处理器执行时实现如权利要求1至4中任一项所述的方法。Wherein, the computer program implements the method according to any one of claims 1 to 4 when executed by the processor.
CN202111241027.6A 2021-10-25 2021-10-25 Double-factor authentication method, device and system based on voice recognition Pending CN114003878A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111241027.6A CN114003878A (en) 2021-10-25 2021-10-25 Double-factor authentication method, device and system based on voice recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111241027.6A CN114003878A (en) 2021-10-25 2021-10-25 Double-factor authentication method, device and system based on voice recognition

Publications (1)

Publication Number Publication Date
CN114003878A true CN114003878A (en) 2022-02-01

Family

ID=79923762

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111241027.6A Pending CN114003878A (en) 2021-10-25 2021-10-25 Double-factor authentication method, device and system based on voice recognition

Country Status (1)

Country Link
CN (1) CN114003878A (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102543084A (en) * 2010-12-29 2012-07-04 盛乐信息技术(上海)有限公司 Online voiceprint recognition system and implementation method thereof
CN104765996A (en) * 2014-01-06 2015-07-08 讯飞智元信息科技有限公司 Voiceprint authentication method and system
CN106920545A (en) * 2017-03-21 2017-07-04 百度在线网络技术(北京)有限公司 Speech Feature Extraction and device based on artificial intelligence
US10210860B1 (en) * 2018-07-27 2019-02-19 Deepgram, Inc. Augmented generalized deep learning with special vocabulary
CN112435672A (en) * 2020-10-15 2021-03-02 讯飞智元信息科技有限公司 Voiceprint recognition method, device, equipment and storage medium
CN112751838A (en) * 2020-12-25 2021-05-04 中国人民解放军陆军装甲兵学院 Identity authentication method, device and system

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102543084A (en) * 2010-12-29 2012-07-04 盛乐信息技术(上海)有限公司 Online voiceprint recognition system and implementation method thereof
CN104765996A (en) * 2014-01-06 2015-07-08 讯飞智元信息科技有限公司 Voiceprint authentication method and system
CN106920545A (en) * 2017-03-21 2017-07-04 百度在线网络技术(北京)有限公司 Speech Feature Extraction and device based on artificial intelligence
US10210860B1 (en) * 2018-07-27 2019-02-19 Deepgram, Inc. Augmented generalized deep learning with special vocabulary
CN112435672A (en) * 2020-10-15 2021-03-02 讯飞智元信息科技有限公司 Voiceprint recognition method, device, equipment and storage medium
CN112751838A (en) * 2020-12-25 2021-05-04 中国人民解放军陆军装甲兵学院 Identity authentication method, device and system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
成刚: "《一本书读懂智能家居核心技术》", 31 January 2020, pages: 239 - 243 *

Similar Documents

Publication Publication Date Title
US6810480B1 (en) Verification of identity and continued presence of computer users
TWI592820B (en) Man-machine recognition method and system
EP2817601B1 (en) System and method for speaker recognition on mobile devices
Liu et al. An MFCC‐based text‐independent speaker identification system for access control
KR100297833B1 (en) Speaker verification system using continuous digits with flexible figures and method thereof
WO2008083571A1 (en) A random voice print cipher certification system, random voice print cipher lock and generating method thereof
CN1841402A (en) System, method, and computer program product for determining whether to accept a subject for registration
CN101231737A (en) A system and method for enhancing the security of online banking transactions
CN109785834B (en) Voice data sample acquisition system and method based on verification code
Duraibi Voice biometric identity authentication model for IoT devices
CN103078828A (en) Cloud-model voice authentication system
Beigi Challenges of LargeScale Speaker Recognition
Chang et al. My voiceprint is my authenticator: A two-layer authentication approach using voiceprint for voice assistants
CN111710340A (en) Method, device, server and storage medium for identifying user identity based on voice
CN111312258A (en) User identity authentication method, device, server and storage medium
CN114003878A (en) Double-factor authentication method, device and system based on voice recognition
Al-Shayea et al. Speaker identification: A novel fusion samples approach
CN119675941A (en) Two-factor identity authentication method and system based on password and key resonance biometrics
Pawade et al. Voice Based Authentication Using Mel-Frequency Cepstral Coefficients and Gaussian Mixture Model
CN114003883A (en) Portable digital identity authentication equipment and identity authentication method
Yelure et al. Machine Learning based Voice Authentication and Identification
Priyadharshini et al. Convolutional Neural Network for Speaker Recognition Embedding with Biometric System
CN1655501A (en) Identification apparatus and method employing biological statistic data
Kalamani et al. Advancement and Applications in Biometric Techniques: A Comprehensive Study
Aloufi et al. On-Device Voice Authentication with Paralinguistic Privacy

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20220201

RJ01 Rejection of invention patent application after publication