CN110970018B - Speech recognition method and device - Google Patents
Speech recognition method and device Download PDFInfo
- Publication number
- CN110970018B CN110970018B CN201811143178.6A CN201811143178A CN110970018B CN 110970018 B CN110970018 B CN 110970018B CN 201811143178 A CN201811143178 A CN 201811143178A CN 110970018 B CN110970018 B CN 110970018B
- Authority
- CN
- China
- Prior art keywords
- language
- voice information
- information
- voice
- candidate
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 39
- 238000012549 training Methods 0.000 claims abstract description 45
- 238000006243 chemical reaction Methods 0.000 claims abstract description 26
- 238000013527 convolutional neural network Methods 0.000 claims description 17
- 238000012545 processing Methods 0.000 claims description 10
- 230000008569 process Effects 0.000 abstract description 8
- 238000013518 transcription Methods 0.000 abstract description 6
- 230000035897 transcription Effects 0.000 abstract description 6
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000003745 diagnosis Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000005034 decoration Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 241001672694 Citrus reticulata Species 0.000 description 1
- 206010011878 Deafness Diseases 0.000 description 1
- 238000004378 air conditioning Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/005—Language recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/04—Training, enrolment or model building
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Quality & Reliability (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Machine Translation (AREA)
Abstract
The invention discloses a voice recognition method and a voice recognition device. Wherein, the method comprises the following steps: acquiring voice information; based on the voice information, determining the language to which the voice information belongs according to a pre-acquired language judgment model, wherein the language judgment model is obtained by training according to training data, and the training data comprises: voice information of a plurality of languages and tags for indicating the languages to which the voice information belongs; and calling the corresponding language conversion module according to the language to which the language information belongs, and converting the voice information into corresponding character information. The invention solves the technical problem that the process of voice transcription in the prior art can only run in a fixed language mode, so that the intelligence degree is lower.
Description
Technical Field
The invention relates to the field of language processing, in particular to a voice recognition method and a voice recognition device.
Background
The artificial intelligence technology is rapidly developed, the influence of smart home on the life of a user is more and more, the application convenience is gradually upgraded, and some problems needing improvement still exist.
For example, the existing smart home system has a function of receiving a voice command, but can only receive a fixed voice command, and if the language used for sending the voice command is changed, the equipment is difficult to correctly recognize when performing the operation of transcribing characters by voice; if the voice command comprises voices of a plurality of languages, all voice information can still be converted into one voice when the equipment carries out voice character transcription operation, so that the intelligence degree is low, and the user experience is poor.
Aiming at the problem that the process of voice transcription in the prior art can only run in a fixed language mode, which results in lower intelligence degree, an effective solution is not provided at present.
Disclosure of Invention
The embodiment of the invention provides a voice recognition method and a voice recognition device, which at least solve the technical problem that the intelligent degree is lower because the voice transcription process in the prior art can only run in a fixed language mode.
According to an aspect of an embodiment of the present invention, there is provided a speech recognition method including: acquiring voice information; based on the voice information, determining the language to which the voice information belongs according to a pre-acquired language judgment model, wherein the language judgment model is obtained by training according to training data, and the training data comprises: voice information of a plurality of languages and tags for indicating the languages to which the voice information belongs; and calling the corresponding language conversion module according to the language to which the language information belongs, and converting the voice information into corresponding character information.
Further, acquiring collected sound information; and denoising the sound information to obtain the voice information of the target object.
Further, inputting the voice information into a language judgment model, wherein the language judgment model outputs the probability that the voice information belongs to each candidate language; and determining the language to which the voice information belongs according to the probability corresponding to each candidate language.
Further, acquiring preset candidate languages and weights corresponding to the candidate languages; and determining the language to which the voice information belongs according to the probability corresponding to each candidate language and the weight of each candidate language.
Further, calling a corresponding language conversion module according to the language to which the language information belongs, converting the voice information into corresponding text information, and then acquiring a preset text display type; and displaying the text information according to the text display type.
Further, before obtaining the voice information, obtaining a language judgment model, wherein obtaining the language judgment model includes: acquiring training data and an initial convolutional neural network model, wherein the initial convolutional neural network model has initial network parameters; and training the initial convolutional neural network model by using the training data to obtain target network parameters, wherein the target network parameters are used for forming a language judgment model.
According to another aspect of the embodiments of the present invention, there is also provided a speech recognition apparatus, including: the acquisition module is used for acquiring voice information; a determining module, configured to determine, based on the voice information, a language to which the voice information belongs according to a language judgment model obtained in advance, where the language judgment model is obtained by training according to training data, and the training data includes: the voice information of a plurality of languages and a label used for expressing the language to which the voice information belongs; and the conversion module is used for calling the corresponding language conversion module according to the language to which the language information belongs and converting the voice information into corresponding character information.
Further, the acquisition module includes: the acquisition submodule is used for acquiring the acquired sound information; and the processing submodule is used for carrying out denoising processing on the sound information to obtain the voice information of the target object.
According to another aspect of the embodiments of the present invention, there is also provided a storage medium, where the storage medium includes a stored program, and when the program runs, the apparatus on which the storage medium is located is controlled to execute the above-mentioned speech recognition method.
According to another aspect of the embodiments of the present invention, there is also provided a processor, wherein the processor is configured to execute a program, and wherein the program executes to perform the above-mentioned speech recognition method.
In the embodiment of the invention, the voice information is acquired, the language to which the voice information belongs is determined according to the pre-acquired language judgment model based on the voice information, and the corresponding language conversion module is called according to the language to which the language information belongs, so that the voice information is converted into the corresponding text information. According to the scheme, the language of the voice information is judged through the language diagnosis model, the condition that the voice recognition is inaccurate due to the fact that the languages of various languages are subjected to voice recognition according to one language is avoided, and the technical problem that the intelligent degree is low due to the fact that the voice transcription process can only be operated in a fixed language mode in the prior art is solved.
Drawings
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the invention and together with the description serve to explain the invention without limiting the invention. In the drawings:
FIG. 1 is a flow diagram of a speech recognition method according to an embodiment of the present invention; and
fig. 2 is a schematic diagram of a speech recognition apparatus according to an embodiment of the present invention.
Detailed Description
In order to make the technical solutions of the present invention better understood, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that the terms "first," "second," and the like in the description and claims of the present invention and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the invention described herein are capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
Example 1
In accordance with an embodiment of the present invention, there is provided an embodiment of a speech recognition method, it being noted that the steps illustrated in the flowchart of the drawings may be performed in a computer system such as a set of computer-executable instructions and that, although a logical order is illustrated in the flowchart, in some cases, the steps illustrated or described may be performed in an order different than here.
Fig. 1 is a flowchart of a speech recognition method according to an embodiment of the present invention, as shown in fig. 1, the method including the steps of:
step S102, voice information is acquired.
Specifically, the voice information may be a voice uttered by the user, or a voice uttered by the user through controlling other devices.
The voice recognition method can be applied to various scenes, for example, a scene of an intelligent home, can be applied to air conditioning equipment, enables a user to control an air conditioner through voice, and can be applied to control of an intelligent terminal, for example, the scene of the intelligent terminal.
In an alternative embodiment, taking the application in an intelligent terminal as an example, the intelligent terminal may pick up voice information of a user through a microphone; taking the smart home system as an example, the smart home can also pick up the voice information of the user through the microphone.
Step S104, based on the voice information, determining the language to which the voice information belongs according to a language judgment model obtained in advance, wherein the language judgment model is obtained by training according to training data, and the training data comprises: the voice information of a plurality of languages and a label for indicating the language to which the voice information belongs.
In the foregoing solution, the language judgment model may be obtained by pre-training according to training data, and the speech information is judged according to the pre-obtained language judgment model, so that corresponding speech recognition can be performed according to a language judgment result of the speech information.
In an optional embodiment, after receiving the voice information, the device does not input the voice information to the voice recognition module, but first inputs the voice information to the language type determination module to determine the language type of the voice.
The training data comprises voice information of a plurality of languages and a label used for representing the language to which the voice information belongs, the label and the language have a preset corresponding relation, and the label is used for determining the language to which the voice information in the training data belongs when the language judgment model is trained.
It should be noted here that the speech information received by the device may be that the whole sentence belongs to one language, for example: "television on the TV" belongs to chinese language, "turn on the TV" belongs to english language, but it is also possible that the whole sentence contains speech information in multiple languages, for example: and playing the my heart will go on. Therefore, when the language judgment model judges the language to which the speech information belongs, the language to which the whole sentence belongs can be judged, and each word can be judged after the sentence in the speech information is segmented.
Step S106, calling a corresponding language conversion module according to the language to which the language information belongs, and converting the voice information into corresponding character information.
In an alternative embodiment, a plurality of speech conversion modules may be provided in the speech recognition system, for example: the system comprises a Chinese voice conversion module, an English voice conversion module, a French voice conversion module and the like, wherein in the voice recognition process, the corresponding voice conversion module is called to recognize according to the language type of the determined voice information.
In another alternative embodiment, the chinese speech conversion module may further classify the chinese speech according to the type of the chinese, for example: the Guangdong language voice conversion module, the Henan speech conversion module, the Mandarin language voice conversion module and the like can further determine the category of the voice information in order to further improve the accuracy of voice information identification under the condition that the language to which the voice information belongs is Chinese, so that the voice conversion module can be called more accurately.
In the prior art, equipment can always recognize according to the type of voice recognition specified by a user, and if the type of voice recognition specified by the user is Chinese, the equipment can still recognize according to the Chinese even if the user sends out voice information containing foreign language. For example, the user instructs the smart audio to: and playing the my heart will go on. The device can only recognize "play" and it is difficult to recognize "my heart will go on", so it is only acoustically difficult to correctly execute the instruction of the user.
In the above scheme, the device may first determine the language to which the voice information belongs, and then recognize according to the language, and still take the voice instruction "play my heart will go on" as an example, after the intelligent sound recognizes that "play" is chinese and "my heart will go on" is english, may call the chinese voice recognition module to recognize "play", and call the english recognition module to recognize "my heart will go on" so as to determine the instruction of the user and execute it accurately.
Therefore, in the embodiment of the present application, the voice information is obtained, the language to which the voice information belongs is determined according to the pre-obtained language judgment model based on the voice information, and the corresponding language conversion module is called according to the language to which the language information belongs, so as to convert the voice information into the corresponding text information. According to the scheme, the language of the voice information is judged through the language diagnosis model, the condition that the voice recognition is inaccurate due to the fact that the languages of various languages are subjected to voice recognition according to one language is avoided, and the technical problem that the intelligent degree is low due to the fact that the voice transcription process can only be operated in a fixed language mode in the prior art is solved.
As an alternative embodiment, acquiring the voice information includes: acquiring collected sound information; and denoising the sound information to obtain the voice information of the target object.
When the device collects the voice information, since there may be other interfering sounds in the environment, for example, in an indoor environment, besides the sound of the voice information emitted by the user, there may also be sounds emitted by other devices such as a television and a sound, so that it is difficult to obtain an accurate voice recognition result, after the device collects the voice information, the device may perform denoising processing on the voice information in a manner such as filtering, so as to extract the voice information from the voice information.
In an alternative embodiment, the sound information may be denoised by means of wavelet transform. Decomposing the signal into multiple scales by using wavelet transform, performing threshold processing on each layer of wavelet coefficient, separating the noise wavelet coefficient from the wavelet coefficient of the sound signal, and recovering the signal by using a wavelet reconstruction algorithm, thereby achieving the purpose of noise reduction.
As an alternative embodiment, determining, based on the speech information, a language to which the speech information belongs according to a language judgment model obtained in advance includes: inputting the voice information into a language judgment model, wherein the language judgment model outputs the probability that the voice information belongs to each candidate language; and determining the language to which the voice information belongs according to the probability corresponding to each candidate language.
Specifically, the language candidates may be selected by the user in advance, or may be default by the device in advance. The language judgment model predicts the probability of each candidate language, which is used for representing the possibility that the voice information belongs to the language, and determines the language to which the voice information belongs according to the prediction result.
In an alternative embodiment, the language candidate with the highest probability may be determined as the language to which the speech information belongs. For example, the candidate languages are chinese and english, and after the device receives the speech information, the language judgment model predicts the language of the speech information to obtain that the probability that the speech information is chinese is 98% and the probability that the speech information is english is 2%, so that the chinese can be determined to be the language to which the sentence belongs.
As an alternative embodiment, determining the language to which the speech information belongs according to the probability corresponding to each candidate language includes: acquiring preset candidate languages and weights corresponding to the candidate languages; and determining the language to which the voice information belongs according to the probability corresponding to each candidate language and the weight of each candidate language.
Specifically, the candidate languages and the weights of the candidate languages may be set by the user according to the requirement. Taking an intelligent air conditioner in an intelligent home scene as an example, if a family includes not only members using Chinese as a native language but also members using English as a native language, the candidate languages can be set as Chinese and English, and because the number of the members using Chinese as a native language is large, the weight of Chinese can be set to be greater than that of English.
After the probability result output by the language judgment model is obtained, the probability of each candidate language after combining the weight can be obtained by combining the weight corresponding to each candidate language, and the language to which the voice information belongs can be determined by combining the probability after combining the weight.
In an alternative embodiment, the candidate languages are set to Chinese and English, and the probability of Chinese corresponding to 65% and the probability of English corresponding to 35% are set. The prediction result of the language diagnosis model on the voice information is as follows: 0.6 Chinese and 0.4 English. The probability 65% corresponding to Chinese is multiplied by the weight 0.65 corresponding to Chinese to obtain 0.39, the probability 35% corresponding to English is multiplied by the weight 0.4 corresponding to English to obtain 0.14, and the speech information is determined to belong to Chinese due to the fact that 0.39 is greater than 0.14.
In application scenarios such as smart home, the voice information received by the device is usually a voice instruction of the user to the device, and therefore, in a case where the device recognizes the voice information to obtain the voice instruction, the device can execute the voice instruction, but in other scenarios, the recognition result of the voice information needs to be displayed, for example, in a scenario of voice input, the voice recognition result needs to be displayed in a display interface of the terminal; for example, for the deaf disabled, the voice information of the other person needs to be recognized as the text information and displayed to the disabled so as to communicate with the other person. Therefore, as an optional embodiment, after calling the corresponding language conversion module according to the language to which the language information belongs, and converting the voice information into the corresponding text information, the method further includes: acquiring a preset character display type; and displaying the text information according to the text display type.
Specifically, the character display types may include characters of characters, font sizes, colors, and the like, taking chinese as an example, the character display types corresponding to chinese may include simplified chinese and traditional chinese, in traditional chinese, port traditional chinese and desk traditional chinese, and may be classified according to sons and regular script; for English, the display type may include a flower body or a body.
In the above scheme, the user selects the type required by the user in advance, and the device displays the recognition result according to the type selected by the user.
As an optional embodiment, before acquiring the voice information, the method further includes: obtaining a language judgment model, wherein obtaining the language judgment model comprises: acquiring training data and an initial convolutional neural network model, wherein the initial convolutional neural network model has initial network parameters; and training the initial convolutional neural network model by using the training data to obtain target network parameters, wherein the target network parameters are used for forming a language judgment model.
Specifically, the training data may include multiple sets of training data, each set of training data includes at least one piece of voice information and a label corresponding to the voice information, and the label corresponding to the voice information is used to indicate a language to which the voice information belongs. And training the initial convolutional neural network model by using the training data to obtain a language judgment model.
In an optional embodiment, the speech information in the training data may be input to the initial convolutional neural network model to obtain a result output by the initial convolutional neural network model, a cross entropy loss function of the result output by the initial convolutional neural network model and a label corresponding to the speech information is obtained according to the result output by the initial convolutional neural network model and the label corresponding to the speech information, and after a large amount of training, the cross entropy loss function is converged to obtain a network parameter of the language judgment model, so as to obtain the language judgment model.
Example 2
According to an embodiment of the present invention, there is provided a speech recognition apparatus, and fig. 2 is a schematic diagram of a speech recognition apparatus according to an embodiment of the present invention, and with reference to fig. 2, the apparatus includes:
and an obtaining module 20, configured to obtain the voice information.
The determining module 22 is configured to determine, based on the speech information, a language to which the speech information belongs according to a language judgment model obtained in advance, where the language judgment model is obtained by training according to training data, and the training data includes: the voice information of a plurality of languages and a label for indicating the language to which the voice information belongs.
And the conversion module 24 is configured to call the corresponding language conversion module according to the language to which the language information belongs, and convert the voice information into corresponding text information.
As an alternative embodiment, the obtaining module includes: the acquisition submodule is used for acquiring the acquired sound information; and the processing submodule is used for carrying out denoising processing on the sound information to obtain the voice information of the target object.
As an alternative embodiment, the determining module includes: the input submodule is used for inputting the voice information into the language judgment model, wherein the language judgment model outputs the probability that the voice information belongs to each candidate language; and the determining submodule is used for determining the language to which the voice information belongs according to the probability corresponding to each candidate language.
As an alternative embodiment, the determining sub-module includes: the acquiring unit is used for acquiring preset candidate languages and weights corresponding to the candidate languages; and the determining unit is used for determining the language to which the voice information belongs according to the probability corresponding to each candidate language and the weight of each candidate language.
As an alternative embodiment, the apparatus further comprises: the type acquisition module is used for calling the corresponding language conversion module according to the language to which the language information belongs, converting the voice information into corresponding character information and then acquiring a preset character display type; and the display module is used for displaying the text information according to the text display type.
As an alternative embodiment, the apparatus further comprises: the model acquisition module is used for acquiring the language judgment model before acquiring the voice information, wherein the model acquisition module comprises: the acquisition submodule is used for acquiring training data and an initial convolutional neural network model, wherein the initial convolutional neural network model has initial network parameters; and the training module is used for training the initial convolutional neural network model by using the training data to obtain target network parameters, wherein the target network parameters are used for forming a language judgment model.
Example 3
According to an embodiment of the present invention, there is provided a storage medium including a stored program, wherein, when the program runs, a device in which the storage medium is controlled to execute the voice recognition method described in embodiment 1.
Example 4
According to an embodiment of the present invention, there is provided a processor, wherein the processor is configured to execute a program, and when the program runs, the voice recognition method described in embodiment 1 is performed.
The above-mentioned serial numbers of the embodiments of the present invention are merely for description and do not represent the merits of the embodiments.
In the above embodiments of the present invention, the descriptions of the respective embodiments have respective emphasis, and for parts that are not described in detail in a certain embodiment, reference may be made to related descriptions of other embodiments.
In the embodiments provided in the present application, it should be understood that the disclosed technology can be implemented in other ways. The above-described embodiments of the apparatus are merely illustrative, and for example, the division of the units may be a logical division, and in actual implementation, there may be another division, for example, multiple units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, units or modules, and may be in an electrical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present invention, which is substantially or partly contributed by the prior art, or all or part of the technical solution may be embodied in a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to perform all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a removable hard disk, a magnetic or optical disk, and other various media capable of storing program codes.
The foregoing is only a preferred embodiment of the present invention, and it should be noted that, for those skilled in the art, various modifications and decorations can be made without departing from the principle of the present invention, and these modifications and decorations should also be regarded as the protection scope of the present invention.
Claims (8)
1. A speech recognition method, comprising:
acquiring voice information;
inputting the voice information into a language judgment model based on the voice information, wherein the language judgment model outputs the probability that the voice information belongs to each candidate language; determining the language to which the voice information belongs according to the probability corresponding to each candidate language, wherein the language judgment model is obtained by training according to training data, and the training data comprises: determining the language to which the voice information belongs according to the probability corresponding to each candidate language, wherein the method comprises the following steps of: acquiring preset candidate languages and weights corresponding to the candidate languages; determining the language to which the voice information belongs according to the probability corresponding to each candidate language and the weight of each candidate language;
and calling a corresponding language conversion module according to the language to which the language information belongs, and converting the voice information into corresponding text information.
2. The method of claim 1, wherein obtaining voice information comprises:
acquiring collected sound information;
and denoising the sound information to obtain the voice information of the target object.
3. The method according to claim 1, wherein after calling a corresponding language conversion module according to a language to which the language information belongs to convert the voice information into corresponding text information, the method further comprises:
acquiring a preset character display type;
and displaying the text information according to the text display type.
4. The method of claim 1, wherein prior to obtaining the voice information, the method further comprises: obtaining the language judgment model, wherein obtaining the language judgment model comprises:
obtaining the training data and an initial convolutional neural network model, wherein the initial convolutional neural network model has initial network parameters;
and training the initial convolutional neural network model by using the training data to obtain target network parameters, wherein the target network parameters are used for forming the language judgment model.
5. A speech recognition apparatus, comprising:
the acquisition module is used for acquiring voice information;
the determining module is used for inputting the voice information into a language judgment model based on the voice information, wherein the language judgment model outputs the probability that the voice information belongs to each candidate language; determining the language to which the voice information belongs according to the probability corresponding to each candidate language, wherein the language judgment model is obtained by training according to training data, and the training data comprises: determining the language to which the voice information belongs according to the probability corresponding to each candidate language, wherein the method comprises the following steps of: acquiring preset candidate languages and weights corresponding to the candidate languages; determining the language to which the voice information belongs according to the probability corresponding to each candidate language and the weight of each candidate language;
and the conversion module is used for calling the corresponding language conversion module according to the language to which the language information belongs and converting the voice information into corresponding character information.
6. The apparatus of claim 5, wherein the obtaining module comprises:
the acquisition submodule is used for acquiring the acquired sound information;
and the processing submodule is used for carrying out denoising processing on the sound information to obtain the voice information of the target object.
7. A storage medium comprising a stored program, wherein the program, when executed, controls an apparatus in which the storage medium is located to perform the speech recognition method according to any one of claims 1 to 4.
8. A processor, characterized in that the processor is configured to run a program, wherein the program is configured to perform the speech recognition method according to any one of claims 1 to 4 when running.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811143178.6A CN110970018B (en) | 2018-09-28 | 2018-09-28 | Speech recognition method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811143178.6A CN110970018B (en) | 2018-09-28 | 2018-09-28 | Speech recognition method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110970018A CN110970018A (en) | 2020-04-07 |
CN110970018B true CN110970018B (en) | 2022-05-27 |
Family
ID=70027236
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811143178.6A Active CN110970018B (en) | 2018-09-28 | 2018-09-28 | Speech recognition method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110970018B (en) |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111477220B (en) * | 2020-04-15 | 2023-04-25 | 南京邮电大学 | Neural network voice recognition method and system for home spoken language environment |
CN113539233B (en) * | 2020-04-16 | 2024-07-30 | 北京搜狗科技发展有限公司 | Voice processing method and device and electronic equipment |
CN111798836B (en) * | 2020-08-03 | 2023-12-05 | 上海茂声智能科技有限公司 | An automatic language switching method, device, system, equipment and storage medium |
CN112259111A (en) * | 2020-09-18 | 2021-01-22 | 惠州高盛达智显科技有限公司 | Emergency broadcast method and system based on Raspberry Pi |
CN112885346B (en) * | 2021-01-19 | 2024-07-23 | 安徽迪科数金科技有限公司 | Intelligent language identification and translation system |
CN112908333B (en) * | 2021-05-08 | 2021-07-16 | 鹏城实验室 | Speech recognition method, device, equipment and computer readable storage medium |
CN113380224A (en) * | 2021-06-04 | 2021-09-10 | 北京字跳网络技术有限公司 | Language determination method and device, electronic equipment and storage medium |
CN113077781B (en) * | 2021-06-04 | 2021-09-07 | 北京世纪好未来教育科技有限公司 | Speech recognition method, device, electronic device and storage medium |
CN113597641A (en) * | 2021-06-22 | 2021-11-02 | 华为技术有限公司 | Voice processing method, device and system |
CN116597827A (en) * | 2023-05-23 | 2023-08-15 | 苏州科帕特信息科技有限公司 | Target language model determining method and device |
CN117558269B (en) * | 2024-01-11 | 2024-03-15 | 深圳波洛斯科技有限公司 | Voice recognition method, device, medium and electronic equipment |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102262440A (en) * | 2010-06-11 | 2011-11-30 | 微软公司 | Multi-modal gender recognition |
CN103853703A (en) * | 2014-02-19 | 2014-06-11 | 联想(北京)有限公司 | Information processing method and electronic equipment |
CN105389162A (en) * | 2014-09-09 | 2016-03-09 | 北京金山安全软件有限公司 | Method and device for changing terminal system language and terminal |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GR20050100216A (en) * | 2005-05-04 | 2006-12-18 | i-sieve :����������� ����������� ������������ ����������� �.�.�. | Method for probabilistic information fusion to filter multi-lingual, semi-structured and multimedia electronic content. |
CN101702314B (en) * | 2009-10-13 | 2011-11-09 | 清华大学 | Method for establishing identified type language recognition model based on language pair |
CN102695096A (en) * | 2012-05-17 | 2012-09-26 | 深圳市九洲电器有限公司 | Method and device for acquiring output image-text language of bit stream and set top box |
US9966064B2 (en) * | 2012-07-18 | 2018-05-08 | International Business Machines Corporation | Dialect-specific acoustic language modeling and speech recognition |
CN103400577B (en) * | 2013-08-01 | 2015-09-16 | 百度在线网络技术(北京)有限公司 | The acoustic model method for building up of multilingual speech recognition and device |
CN104036774B (en) * | 2014-06-20 | 2018-03-06 | 国家计算机网络与信息安全管理中心 | Tibetan dialect recognition methods and system |
CN105138663A (en) * | 2015-09-01 | 2015-12-09 | 百度在线网络技术(北京)有限公司 | Word bank query method and device |
CN105222275B (en) * | 2015-09-25 | 2018-04-13 | 珠海格力电器股份有限公司 | Display data switching method, device and system |
CN106598937B (en) * | 2015-10-16 | 2019-10-18 | 阿里巴巴集团控股有限公司 | Language Identification, device and electronic equipment for text |
CN105957516B (en) * | 2016-06-16 | 2019-03-08 | 百度在线网络技术(北京)有限公司 | More voice identification model switching method and device |
CN106328125B (en) * | 2016-10-28 | 2023-08-04 | 许昌学院 | A Speech Recognition System for Henan Dialect |
CN108510976B (en) * | 2017-02-24 | 2021-03-19 | 芋头科技(杭州)有限公司 | Multi-language mixed voice recognition method |
CN106997762A (en) * | 2017-03-08 | 2017-08-01 | 广东美的制冷设备有限公司 | The sound control method and device of household electrical appliance |
CN106961508A (en) * | 2017-06-01 | 2017-07-18 | 诺雪(重庆)科技有限公司 | Communication means and device based on Sex criminals |
CN107169566A (en) * | 2017-06-09 | 2017-09-15 | 山东师范大学 | Dynamic neural network model training method and device |
-
2018
- 2018-09-28 CN CN201811143178.6A patent/CN110970018B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102262440A (en) * | 2010-06-11 | 2011-11-30 | 微软公司 | Multi-modal gender recognition |
CN103853703A (en) * | 2014-02-19 | 2014-06-11 | 联想(北京)有限公司 | Information processing method and electronic equipment |
CN105389162A (en) * | 2014-09-09 | 2016-03-09 | 北京金山安全软件有限公司 | Method and device for changing terminal system language and terminal |
Non-Patent Citations (1)
Title |
---|
基于独立分量分析的语种识别方法;陈刚等;《计算机工程》;20061220(第24期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN110970018A (en) | 2020-04-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110970018B (en) | Speech recognition method and device | |
CN107945792B (en) | Voice processing method and device | |
CN110610707B (en) | Voice keyword recognition method and device, electronic equipment and storage medium | |
CN110619889B (en) | Sign data identification method, device, electronic device and storage medium | |
US10438586B2 (en) | Voice dialog device and voice dialog method | |
CN107492379B (en) | Voiceprint creating and registering method and device | |
CN108305643B (en) | Method and device for determining emotion information | |
CN109660865B (en) | Method and device for automatically labeling videos, medium and electronic equipment | |
CN108447471A (en) | Audio recognition method and speech recognition equipment | |
CN110942763B (en) | Speech recognition method and device | |
CN110148399A (en) | A kind of control method of smart machine, device, equipment and medium | |
CN104391673A (en) | Voice interaction method and voice interaction device | |
US9251808B2 (en) | Apparatus and method for clustering speakers, and a non-transitory computer readable medium thereof | |
CN114387945B (en) | Voice generation method, device, electronic equipment and storage medium | |
CN110262278B (en) | Control method and device of intelligent household electrical appliance and intelligent household electrical appliance | |
CN110931018A (en) | Intelligent voice interaction method and device and computer readable storage medium | |
KR102345625B1 (en) | Caption generation method and apparatus for performing the same | |
CN112420049A (en) | Data processing method, device and storage medium | |
CN113362815A (en) | Voice interaction method, system, electronic equipment and storage medium | |
CN111477212A (en) | Content recognition, model training and data processing method, system and equipment | |
CN112597889A (en) | Emotion processing method and device based on artificial intelligence | |
CN110895936B (en) | Voice processing method and device based on household appliance | |
KR102631143B1 (en) | Voice synthesizer using artificial intelligence, operating method of voice synthesizer and computer redable recording medium | |
CN110970019A (en) | Control method and device of intelligent home system | |
CN111128127A (en) | Voice recognition processing method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |