TWI675305B - Intelligent voice control management system and method thereof - Google Patents
Intelligent voice control management system and method thereof Download PDFInfo
- Publication number
- TWI675305B TWI675305B TW107125806A TW107125806A TWI675305B TW I675305 B TWI675305 B TW I675305B TW 107125806 A TW107125806 A TW 107125806A TW 107125806 A TW107125806 A TW 107125806A TW I675305 B TWI675305 B TW I675305B
- Authority
- TW
- Taiwan
- Prior art keywords
- controlled device
- voice
- control
- control instruction
- management system
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 15
- 238000007726 management method Methods 0.000 claims abstract description 83
- 238000004458 analytical method Methods 0.000 claims abstract description 28
- 238000012795 verification Methods 0.000 claims abstract description 15
- 238000012545 processing Methods 0.000 claims description 33
- 238000012216 screening Methods 0.000 claims description 7
- 230000005540 biological transmission Effects 0.000 claims description 2
- 230000000694 effects Effects 0.000 abstract description 5
- 230000006399 behavior Effects 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 238000004364 calculation method Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000001364 causal effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Landscapes
- Selective Calling Equipment (AREA)
Abstract
一種智能語音控制管理方法,其步驟包括:接收的語音信號通過聲紋驗證後進行語音轉文本識別與語義分析而獲得控制指令;根據多個麥克風單元取得語音信號的來源位置與來源方向;取得每一受控裝置與該智能語音控制管理系統之間的相對距離與相對位置;依據每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置、語音信號的來源位置與來源方向、控制資料庫與控制指令篩選出對應控制指令的受控裝置;以及傳輸控制指令予對應的受控裝置,使其執行對應操作。因此,可達成透過聲紋辨識安全地對區域範圍內的受控裝置進行語音控制之技術功效。An intelligent voice control management method includes the steps of: obtaining voice control text recognition and semantic analysis to obtain control instructions after receiving a voice signal through voiceprint verification; obtaining a source position and a source direction of a voice signal according to a plurality of microphone units; Relative distance and relative position between a controlled device and the intelligent voice control management system; according to the relative distance and relative position between each controlled device and the intelligent voice control management system, the source location and direction of the voice signal, The control database and the control instructions screen out the controlled devices corresponding to the control instructions; and transmit the control instructions to the corresponding controlled devices to perform the corresponding operations. Therefore, the technical effect of performing voice control on the controlled devices in the area safely through voiceprint recognition can be achieved.
Description
本發明涉及一種管理系統及其方法,特別是智能語音控制管理系統及其方法。The invention relates to a management system and a method thereof, in particular to an intelligent voice control management system and a method thereof.
隨著通訊技術的進步,和語音辨識相關的許多應用技術也逐漸成為不可或缺的項目之一,使用者可直接使用語言與電子化產品溝通,讓電子裝置之操控介面更易被使用者所接受。With the development of communication technology, many application technologies related to speech recognition have gradually become one of the indispensable items. Users can directly use language to communicate with electronic products, making the control interface of electronic devices easier to be accepted by users. .
然而,目前市面部分電子裝置不具有語音辨識功能,令使用者無法有良好的操控體驗。此外,習知語音控制技術僅針對說話者的語音內容進行辨識,導致任何人都可利用語音方式來操作對應的電子裝置。因此,可能造成幼童誤用危險性高的電子裝置,甚至陌生人也能夠隨意使用電子裝置,嚴重影響電子裝置的使用安全或侵犯電子裝置合法使用者的權利。However, currently some electronic devices on the market do not have a voice recognition function, which prevents users from having a good control experience. In addition, the conventional voice control technology only recognizes the speaker's voice content, so that anyone can use the voice to operate the corresponding electronic device. Therefore, it may cause young children to misuse high-risk electronic devices, and even strangers can use the electronic devices at will, which seriously affects the use safety of electronic devices or violates the rights of legal users of electronic devices.
本發明揭露一種智能語音控制管理系統及其方法。The invention discloses an intelligent voice control management system and method.
首先,本發明揭露一種智能語音控制管理系統,此系統包括:接收模組、聲紋辨識模組、語義解析模組、定位模組及處理模組。其中,接收模組包括多個麥克風單元,用以接收語音信號;聲紋辨識模組用以自語音信號取得聲音特徵,並依據聲紋資料庫對聲音特徵進行聲紋驗證;語義解析模組用以當聲音特徵通過聲紋驗證後對語音信號進行語音轉文本識別與語義分析而獲得控制指令;定位模組用以根據多個麥克風單元接收到語音信號的時間差與該些麥克風單元的相對配置位置取得語音信號的來源位置與來源方向,以及依據定位信號與每一受控裝置所傳輸的位置信號取得每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置;以及處理模組用以依據每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置、語音信號的來源位置與來源方向、控制資料庫與控制指令篩選出對應控制指令的受控裝置,並傳輸控制指令予對應的受控裝置,使其執行對應操作。First, the present invention discloses an intelligent voice control management system. The system includes a receiving module, a voiceprint recognition module, a semantic analysis module, a positioning module, and a processing module. Among them, the receiving module includes multiple microphone units for receiving voice signals; the voiceprint recognition module is used to obtain the voice characteristics from the voice signals, and the voice characteristics are verified according to the voiceprint database; the semantic analysis module is used for The control instruction is obtained by performing speech-to-text recognition and semantic analysis on the voice signal after the voice feature is verified by the voiceprint; the positioning module is used to determine the time difference between the voice signals received by the multiple microphone units and the relative positions of the microphone units. Obtaining the source position and direction of the voice signal, and obtaining the relative distance and relative position between each controlled device and the intelligent voice control management system according to the positioning signal and the position signal transmitted by each controlled device; and a processing module It is used to filter out the controlled devices corresponding to the control instruction according to the relative distance and relative position between each controlled device and the intelligent voice control management system, the source location and direction of the voice signal, the control database and the control instruction, and transmit The control instruction is given to the corresponding controlled device so that it performs the corresponding operation.
另外,本發明揭露一種智能語音控制管理方法,其步驟包括:多個麥克風單元接收語音信號;自語音信號取得聲音特徵,並依據聲紋資料庫對聲音特徵進行聲紋驗證;當聲音特徵通過聲紋驗證後對語音信號進行語音轉文本識別與語義分析而獲得控制指令;根據多個麥克風單元接收到語音信號的時間差與該些麥克風單元的相對配置位置取得語音信號的來源位置與來源方向;依據定位信號與每一受控裝置所傳輸的位置信號取得每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置;依據每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置、語音信號的來源位置與來源方向、控制資料庫與控制指令篩選出對應控制指令的受控裝置;以及傳輸控制指令予對應的受控裝置,使其執行對應操作。In addition, the present invention discloses an intelligent voice control management method, whose steps include: multiple microphone units receive voice signals; obtaining voice characteristics from the voice signals, and performing voiceprint verification on the voice characteristics according to the voiceprint database; when the voice characteristics pass the voice After voice verification, the speech signal is subjected to speech-to-text recognition and semantic analysis to obtain control instructions; the source position and direction of the speech signal are obtained based on the time difference between the speech signals received by multiple microphone units and the relative positions of the microphone units; The positioning signal and the position signal transmitted by each controlled device obtain the relative distance and relative position between each controlled device and the intelligent voice control management system; according to the relative distance between each controlled device and the intelligent voice control management system The distance and relative position, the source position and direction of the voice signal, the control database and the control instruction screen out the controlled device corresponding to the control instruction; and transmit the control instruction to the corresponding controlled device to perform the corresponding operation.
本發明所揭露之系統與方法如上,與先前技術的差異在於本發明是透過接收的語音信號通過聲紋驗證後進行語音轉文本識別與語義分析而獲得控制指令;根據多個麥克風單元取得語音信號的來源位置與來源方向;取得每一受控裝置與該智能語音控制管理系統之間的相對距離與相對位置;依據每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置、語音信號的來源位置與來源方向、控制資料庫與控制指令篩選出對應控制指令的受控裝置;以及傳輸控制指令予對應的受控裝置,使其執行對應操作。The system and method disclosed in the present invention are as described above. The difference from the prior art is that the present invention obtains control instructions through voice-to-text recognition and semantic analysis after the received voice signal is verified by voiceprint verification; the voice signals are obtained according to multiple microphone units. Source location and source direction; obtain the relative distance and relative position between each controlled device and the intelligent voice control management system; based on the relative distance and relative position between each controlled device and the intelligent voice control management system, The source location and direction of the voice signal, the control database and the control instruction screen out the controlled devices corresponding to the control instruction; and transmit the control instructions to the corresponding controlled device to perform the corresponding operation.
透過上述的技術手段,本發明可以達成透過聲紋辨識安全地對區域範圍內的受控裝置進行語音控制之技術功效。Through the above-mentioned technical means, the present invention can achieve the technical effect of safely controlling the voice of controlled devices in the area through voiceprint recognition.
在說明本發明所揭露之智能語音控制管理系統及其方法之前,先對本發明所自行定義的名詞作說明,本發明所述的智能語音控制管理系統所包括的聲紋辨識模組、語義解析模組、定位模組、處理模組、統計模組及學習模組可以利用各種方式來實現,包括軟體、硬體、韌體或其任意組合。在實施中提出的技術使用軟體或韌體可以被儲存在機器可讀儲存媒體上,例如:唯讀記憶體(ROM)、隨機存取記憶體(RAM)、磁盤儲存媒體、光儲存媒體、快閃記憶體裝置等等,並且可以由一個或多個通用或專用的可程式化微處理器執行。本發明所述的智能語音控制管理系統與受控裝置之間可通過網路,例如:行動通訊網路、網際網路、局域網路、廣域網路和/或無線網路相互連通。Before describing the intelligent voice control management system and method disclosed in the present invention, the terms defined by the present invention are explained first. The voiceprint recognition module and semantic analysis module included in the intelligent voice control management system according to the present invention. Groups, positioning modules, processing modules, statistics modules, and learning modules can be implemented in various ways, including software, hardware, firmware, or any combination thereof. The software or firmware used in the proposed technology can be stored on machine-readable storage media, such as: read-only memory (ROM), random access memory (RAM), disk storage media, optical storage media, flash memory Flash memory devices, etc., and can be executed by one or more general-purpose or special-purpose programmable microprocessors. The intelligent voice control management system and the controlled device according to the present invention can communicate with each other through a network, such as: a mobile communication network, the Internet, a local area network, a wide area network, and / or a wireless network.
以下將配合圖式及實施例來詳細說明本發明之實施方式,藉此對本發明如何應用技術手段來解決技術問題並達成技術功效的實現過程能充分理解並據以實施。In the following, the embodiments of the present invention will be described in detail with reference to the drawings and examples, so as to fully understand and implement the implementation process of how the present invention applies technical means to solve technical problems and achieve technical effects.
請先參閱「第1圖」,「第1圖」為本發明智能語音控制管理系統之一實施例系統方塊圖,智能語音控制管理系統100包括:接收模組101、聲紋辨識模組102、語義解析模組103、定位模組104及處理模組105。其中,接收模組101連接聲紋辨識模組102與語義解析模組103,聲紋辨識模組102連接語義解析模組103,語義解析模組103連接定位模組104,定位模組104連接處理模組105。Please refer to "Figure 1" first. "Figure 1" is a system block diagram of an embodiment of the intelligent voice control management system of the present invention. The intelligent voice control management system 100 includes a receiving module 101, a voiceprint recognition module 102, The semantic analysis module 103, the positioning module 104, and the processing module 105. The receiving module 101 is connected to the voiceprint identification module 102 and the semantic analysis module 103, the voiceprint identification module 102 is connected to the semantic analysis module 103, the semantic analysis module 103 is connected to the positioning module 104, and the positioning module 104 is connected to process Module 105.
在本實施例中,接收模組101可包括多個麥克風單元50,用以接收語音信號。其中,麥克風單元50的數量可為但不限於五個且該些麥克風單元50的配置位置可為但不限於呈圓周狀設置於適用智能語音控制管理系統100的管理裝置的殼體上,但本實施例並非用以限定本發明,麥克風單元50的數量及其配置位置可依據實際需求進行調整。此外,相鄰兩個麥克風單元50之間的距離可以是相等的,以提高後續取得語音信號的來源位置與來源方向之精度,但本實施例並非用以限定本發明。再者,呈圓周狀設置的該些麥克風單元50所在圓周的半徑可大於八十毫米,以能夠便取得語音信號進行來源方向的識別,但本實施例並非用以限定本發明。In this embodiment, the receiving module 101 may include a plurality of microphone units 50 for receiving voice signals. The number of the microphone units 50 may be, but is not limited to, five, and the configuration positions of the microphone units 50 may be, but are not limited to, arranged on the casing of the management device to which the intelligent voice control management system 100 is applied. The embodiment is not intended to limit the present invention, and the number of microphone units 50 and their configuration positions can be adjusted according to actual needs. In addition, the distance between two adjacent microphone units 50 may be equal to improve the accuracy of the source location and source direction of subsequent voice signals, but this embodiment is not intended to limit the present invention. Furthermore, the radius of the circumference of the microphone units 50 arranged in a circle may be greater than eighty millimeters, so as to obtain the voice signal for identification of the source direction, but this embodiment is not intended to limit the present invention.
在本實施例中,聲紋辨識模組102用以自語音信號取得聲音特徵,並依據聲紋資料庫60對聲音特徵進行聲紋驗證。其中,聲紋資料庫60儲存多個預設聲紋資料,每一預設聲紋資料代表一位可合法使用智能語音控制管理系統100的使用者的聲紋資料(即聲音特徵)。由於人在講話時使用的發聲器官都不盡相同,使得任何兩個人的聲紋圖譜都有差異,可從聲紋圖譜上觀察說話人發音的聲音特徵,利用這些聲音特徵進行說話人的辨識,因此在一般情況下可透過聲音特徵區別不同的人的聲音。在本實施例中,聲紋辨識模組102接收語音信號後將其轉換成聲紋圖譜,再從該聲紋圖譜取得聲音特徵,接著,將取得的聲音特徵逐一比對聲紋資料庫60中的每一預設聲紋資料(即進行聲紋驗證),當判斷取得的聲音特徵與該些預設聲紋資料其中之一相符時,代表取得的聲音特徵通過聲紋驗證;當判斷取得的聲音特徵與該些預設聲紋資料其中之一不相符時,代表取得的聲音特徵無法通過聲紋驗證。In this embodiment, the voiceprint recognition module 102 is configured to obtain voice characteristics from a voice signal, and perform voiceprint verification on the voice characteristics according to the voiceprint database 60. The voiceprint database 60 stores a plurality of preset voiceprint data, and each preset voiceprint data represents voiceprint data (ie, voice characteristics) of a user who can legally use the intelligent voice control management system 100. Because the vocal organs used by people are different, any two people have different voiceprint patterns. You can observe the voice characteristics of the speaker's pronunciation from the voiceprint pattern and use these voice characteristics to identify the speaker. , So in general, the sound of different people can be distinguished through sound characteristics. In this embodiment, the voiceprint recognition module 102 converts the voice signal into a voiceprint map after receiving the voice signal, and then obtains the voice characteristics from the voiceprint map, and then compares the acquired voice characteristics one by one to the voiceprint database 60. For each preset voiceprint data (that is, to perform voiceprint verification), when it is determined that the acquired voice characteristics match one of the preset voiceprint data, it means that the acquired voice characteristics pass voiceprint verification; When the sound feature does not match one of the preset voiceprint data, the obtained voice feature cannot be verified by the voiceprint.
此外,聲紋資料庫60也可儲存多個身分資料,該些身分資料以一對一方式對應該些預設聲紋資料。其中,每一身分資料代表一位可合法使用智能語音控制管理系統100的使用者的識別資料,可為但不限於身分證字號,可依據實際需求進行調整。當聲紋辨識模組102判斷取得的聲音特徵與該些預設聲紋資料其中之一相符時,即可知道說話人的身分。In addition, the voiceprint database 60 can also store multiple identity data, which correspond to the preset voiceprint data in a one-to-one manner. Wherein, each identity data represents identification data of a user who can legally use the intelligent voice control management system 100, and may be, but not limited to, an identity card number, and may be adjusted according to actual needs. When the voiceprint recognition module 102 determines that the acquired voice characteristics match one of the preset voiceprint data, the identity of the speaker can be known.
在本實施例中,語義解析模組103用以當聲音特徵通過聲紋驗證後對語音信號進行語音轉文本識別與語義分析而獲得控制指令。更詳細地說,語義解析模組103可包含聲學模型單元(未繪製)與語義分析單元(未繪製),當聲音特徵通過聲紋驗證後,聲學模型單元可透過隱藏式馬可夫模型(Hidden Markov Model,HMM)、神經網路(Neural Networks)、動態時間校準(Dynamic Time Warping,DTW)或是語音模版比對(Template Matching)等各式聲學模型來進行語音辨識。當聲學模型單元以聲學模型比對方式來進行語音辨識時,可先從語音信號中擷取多個聲學特徵(即得到語音信號對應的線性預測倒頻譜係數(Linear Predictive Cepstral Coefficient,LPCC)、梅爾倒頻譜係數(Mel Frequency Cepstral Coefficient,MFCC)、MPEG7及對數能量(Log Energy)等參數),接著,將該些聲學特徵與上述其中一聲學模型進行比對,找出與該些聲學特徵最為匹配的對應字彙。舉例而言,當語音信號為「開啟電視」時,經聲學模型單元解析所獲得的對應字彙可為「『開啟』、『電視』」。In this embodiment, the semantic analysis module 103 is configured to obtain a control instruction by performing speech-to-text recognition and semantic analysis on a speech signal after a voice feature is verified by a voiceprint. In more detail, the semantic analysis module 103 may include an acoustic model unit (not drawn) and a semantic analysis unit (not drawn). After the sound features are verified by the voiceprint, the acoustic model unit may be passed through a hidden Markov model (Hidden Markov Model). , HMM), Neural Networks, Dynamic Time Warping (DTW), or Template Matching to perform speech recognition. When the acoustic model unit uses the acoustic model comparison method for speech recognition, multiple acoustic features can be extracted from the speech signal (that is, the Linear Predictive Cepstral Coefficient (LPCC) corresponding to the speech signal, Parameters such as Mel Frequency Cepstral Coefficient (MFCC), MPEG7, and Log Energy, and then compare these acoustic characteristics with one of the above acoustic models to find the most acoustic characteristics Matching corresponding vocabulary. For example, when the voice signal is "turn on the TV", the corresponding vocabulary obtained through the analysis of the acoustic model unit can be "" on "," television ".
接著,語義分析單元可於語義資料庫70查詢該些字彙所對應的類別,以取得控制指令,其中,控制指令可包含操作命令與物件名稱。具體而言,類別可為行為(對應操作命令)或物件(對應物件名稱)。更具體而言,語義資料庫70可記錄有字彙與對應的類別之關係(如「第2圖」所示,「第2圖」為「第1圖」的智能語音控制管理系統的語義資料庫所記錄之字彙與類別的一實施例對照表),每一字彙對應一個類別。由於一個完整的控制指令可包含一行為類別(即說話者欲執行的操作)與一物件類別(即說話者欲操作的目標),因此,語義分析單元需自語義資料庫70查詢聲學模型單元解析語音信號所獲得的該些字彙是否至少對應一行為類別與一物件類別,進而確認是否已收集完整的控制指令,其中,控制指令可具有一特定格式,其可包含一物件名稱以及一操作命令(如「第3A圖」所示,「第3A圖」為「第1圖」的智能語音控制管理系統中控制指令的一實施例格式),於實際運用上, 物件名稱與操作命令的排序可依據實際需求進行調整,物件名稱用於表示欲被操作的物件,操作命令用於表示說話者欲對該物件執行的操作。舉例而言,當聲學模型單元解析語音信號所獲得的對應字彙為「『開啟』、『電視』」時,語義分析單元可根據語義資料庫70得到查詢到「開啟」對應的類別為“行為”, 「電視」對應的類別為“物件”,因此,控制指令的內容可能如「第3B圖」所示,「第3B圖」為「第1圖」的智能語音控制管理系統中控制指令的一實施例內容,「第3B圖」表示說話者想要「開啟」物件,該物件的名稱為「電視」。Then, the semantic analysis unit may query the category corresponding to the vocabulary in the semantic database 70 to obtain a control instruction. The control instruction may include an operation command and an object name. Specifically, the category can be a behavior (corresponding to an operation command) or an object (corresponding to an object name). More specifically, the semantic database 70 may record the relationship between the vocabulary and the corresponding category (as shown in "Figure 2", "Figure 2" is the semantic database of the intelligent voice control management system of "Figure 1" An example comparison table of the recorded vocabulary and category), each vocabulary corresponds to a category. Since a complete control instruction can include a behavior category (that is, the operation that the speaker wants to perform) and an object category (that is, the target that the speaker wants to operate), the semantic analysis unit needs to query the acoustic model unit analysis from the semantic database 70 Whether the vocabulary obtained by the voice signal corresponds to at least a behavior category and an object category, and then confirm whether a complete control instruction has been collected. The control instruction may have a specific format, which may include an object name and an operation command ( As shown in "Figure 3A", "Figure 3A" is an example format of control instructions in the intelligent voice control management system of "Figure 1." In practice, the ordering of object names and operation commands can be based on The actual needs are adjusted. The object name is used to indicate the object to be operated, and the operation command is used to indicate the operation that the speaker wants to perform on the object. For example, when the corresponding vocabulary obtained by the acoustic model unit analyzing the speech signal is "" on "," television "", the semantic analysis unit can obtain the query "on" and the corresponding category is "behavior" according to the semantic database 70 The category corresponding to "TV" is "Object", so the content of the control instruction may be as shown in "Figure 3B", and "Figure 3B" is one of the control instructions in the intelligent voice control management system of "Figure 1." In the content of the embodiment, "Figure 3B" indicates that the speaker wants to "open" the object, and the name of the object is "television".
需注意的是,聲學模型單元解析所獲得的字彙可能無法自語義資料庫70查詢到對應的類別(即代表該字彙無法辨識其類別),但語義資料庫70所記錄之字彙與類別的對照表可隨時依據需求進行更新,其中,對照表中的字彙需包含可透過智能語音控制管理系統100進行語音控制的每一受控裝置的名稱且該些受控裝置的名稱所對應的類別為“物件",操作命令可依據使用者需求進行刪減且該些操作命令所對應的類別為“行為"。此外,當語義解析模組102確認該些字彙對應的類別缺少行為類別或物件類別,或者確認該些字彙對應多個行為類別或多個物件類別時,則判斷無法收集完整的控制指令,而無法取得控制指令。It should be noted that the vocabulary obtained from the analysis of the acoustic model unit may not be able to query the corresponding category from the semantic database 70 (that is, it indicates that the vocabulary cannot identify its category), but the correspondence table between the vocabulary and category recorded in the semantic database 70 It can be updated at any time according to requirements. The vocabulary in the comparison table needs to include the name of each controlled device that can be controlled by voice through the intelligent voice control management system 100 and the category corresponding to the names of the controlled devices is "object ", Operation commands can be deleted according to user needs and the category corresponding to these operation commands is" behavior ". In addition, when the semantic analysis module 102 confirms that the categories corresponding to the vocabularies lack behavior categories or object categories, or confirms that the vocabularies correspond to multiple behavior categories or multiple object categories, it is judged that a complete control instruction cannot be collected, and cannot Get control instructions.
在本實施例中,定位模組104可用以根據該些麥克風單元接收到語音信號的時間差與該些麥克風單元的相對配置位置取得語音信號的來源位置與來源方向。需注意的是,定位模組104取得語音信號的來源位置與來源方向的精準度與麥克風單元50的排列密度及數量有正相關性,換句話說,當麥克風單元50的排列密度越密集及數量越多時,定位模組104取得語音信號的來源位置與來源方向的精準度會越高。In this embodiment, the positioning module 104 may be configured to obtain a source position and a source direction of a voice signal according to a time difference between the microphone units receiving the voice signals and a relative configuration position of the microphone units. It should be noted that the accuracy of the location and direction of the voice signal obtained by the positioning module 104 has a positive correlation with the arrangement density and quantity of the microphone units 50. In other words, when the arrangement density and quantity of the microphone units 50 are more dense The more the positioning module 104 obtains the accuracy of the source position and source direction of the voice signal, the higher the accuracy.
此外,定位模組104可依據定位信號與每一受控裝置所傳輸的位置信號取得每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置。其中,定位信號可為但不限於透過全球定位系統(Global Positioning System, GPS)定位智能語音控制管理系統100所提供的定位信號,定位信號可包含智能語音控制管理系統100的辨識碼或網際網路協定位址(Internet Protocol Address);每一受控裝置所傳輸的位置信號可為但不限於透過GPS定位每一受控裝置所提供的位置信號,位置信號可包含該受控裝置的辨識碼或IP位址。因此,定位模組104可依據智能語音控制管理系統100所在的位置(即對應定位信號)與每一受控裝置所在的位置(即對應每一受控裝置所傳輸的位置信號)取得每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置。In addition, the positioning module 104 can obtain the relative distance and relative position between each controlled device and the intelligent voice control management system according to the positioning signal and the position signal transmitted by each controlled device. The positioning signal may be, but is not limited to, a positioning signal provided by the Global Positioning System (GPS) positioning intelligent voice control management system 100, and the positioning signal may include an identification code of the intelligent voice control management system 100 or the Internet. Internet Protocol Address; the position signal transmitted by each controlled device can be, but is not limited to, the position signal provided by each controlled device through GPS positioning, and the position signal can include the identification code of the controlled device or IP address. Therefore, the positioning module 104 can obtain each target based on the location of the intelligent voice control management system 100 (that is, the corresponding positioning signal) and the location of each controlled device (that is, the corresponding position signal transmitted by each controlled device). The relative distance and relative position between the control device and the intelligent voice control management system.
在本實施例中,處理模組105可用以依據每一受控裝置與智能語音控制管理系統100之間的相對距離與相對位置、語音信號的來源位置與來源方向、控制資料庫90與控制指令篩選出對應控制指令的受控裝置,並傳輸控制指令予對應的受控裝置,使其執行對應操作。其中,控制資料庫90可包括每一受控裝置的名稱、位址與其對應的操作(如「第4圖」所示,「第4圖」為「第1圖」的智能語音控制管理系統的控制資料庫所記錄之內容的一實施例示意圖),每一受控裝置的名稱(即物件名稱)可對應多個操作(即對應多個操作命令),每一受控裝置的位址可為但不限於IP位址,雖然受控裝置的名稱可能相同,但可透過IP位址區分不同的受控裝置。In this embodiment, the processing module 105 can be used to determine the relative distance and relative position between each controlled device and the intelligent voice control management system 100, the source position and direction of the voice signal, the control database 90 and the control instruction. The controlled devices corresponding to the control instructions are screened out, and the control instructions are transmitted to the corresponding controlled devices to perform the corresponding operations. Among them, the control database 90 may include the name and address of each controlled device and its corresponding operation (as shown in "Figure 4", "Figure 4" is the intelligent voice control management system of "Figure 1" A schematic diagram of an embodiment of the content recorded in the control database), the name of each controlled device (ie, the object name) can correspond to multiple operations (that is, corresponding to multiple operation commands), and the address of each controlled device can be But it is not limited to the IP address. Although the names of the controlled devices may be the same, different controlled devices can be distinguished by the IP address.
更詳細地說,處理模組105可依據控制指令(其包含操作命令與物件名稱)從控制資料庫90找到對應該控制指令的受控裝置,當處理模組105篩選出來對應該控制指令的受控裝置數量為一個時,處理模組105透過控制資料庫90中該受控裝置的位址傳輸該控制指令予該受控裝置,使其執行對應操作;當處理模組105篩選出來對應該控制指令的受控裝置數量為多個時,處理模組105可進一步依據每一受控裝置與智能語音控制管理系統100之間的相對距離與相對位置以及語音信號的來源位置與來源方向判斷使用者最有可能欲進行控制的受控裝置。由於受控裝置執行對應操作的影響範圍應與使用者所在位置(即語音信號的來源位置與來源方向)有關,因此,處理模組105可進一步依據每一受控裝置與智能語音控制管理系統100之間的相對距離與相對位置以及語音信號的來源位置與來源方向(即說話者與智能語音控制管理系統100之間的相對距離與相對方位)篩選出對應該控制指令的最佳受控裝置,其中,篩選的依據可為但不限於說話者與受控裝置之間的相對距離與/或相對方位為最小或最接近(可由該受控裝置與智能語音控制管理系統100之間的相對距離與相對位置以及語音信號的來源位置與來源方向計算而取得),實際的篩選機制可依據實際需求進行調整。因此,本發明可以達成透過聲紋辨識安全地對區域範圍內的受控裝置進行語音控制之技術功效。In more detail, the processing module 105 can find the controlled device corresponding to the control instruction from the control database 90 according to the control instruction (which includes the operation command and the object name). When the number of control devices is one, the processing module 105 transmits the control instruction to the controlled device through the address of the controlled device in the control database 90, so that it performs corresponding operations; when the processing module 105 screens out the corresponding control When the number of instructed controlled devices is multiple, the processing module 105 can further judge the user based on the relative distance and relative position between each controlled device and the intelligent voice control management system 100, and the source position and direction of the voice signal. The controlled device most likely to be controlled. Since the affected range of the corresponding operation performed by the controlled device should be related to the location of the user (that is, the source location and direction of the voice signal), the processing module 105 can further control each of the controlled devices and the intelligent voice control management system 100 The relative distance and relative position and the source position and direction of the voice signal (ie, the relative distance and relative orientation between the speaker and the intelligent voice control management system 100) screen out the best controlled device corresponding to the control instruction, The basis for screening may be, but is not limited to, the relative distance and / or the relative orientation between the speaker and the controlled device is the smallest or the closest (can be determined by the relative distance between the controlled device and the intelligent voice control management system 100 and The relative position and the source position and direction of the voice signal are calculated). The actual screening mechanism can be adjusted according to actual needs. Therefore, the present invention can achieve the technical effect of safely controlling the voice of controlled devices in the area through voiceprint recognition.
由於處理模組105依據上述方法(利用說話者與受控裝置之間的相對距離與/或相對方位為最小或最接近)所篩選出的受控裝置可能不是說話者真正項進行語音控制的受控裝置,因此,使用者可透過再次輸入相同的語音信號的方法使智能語音控制管理系統100進行修正與調整。更詳細地說,當某一受控裝置依據處理模組105所傳輸的控制指令執行對應操作一預定時間後(即錯誤的受控裝置執行控制命令一預定時間後),處理模組105再次接收到相同的控制指令(即使用者可透過再次輸入相同的語音信號)且判斷每一受控裝置與智能語音控制管理系統100之間的相對距離與相對位置以及語音信號的來源位置與來源方向不變時(即說話者與受控裝置之間的相對距離與/或相對方位不變),處理模組105可依據控制資料庫90篩選出另一受控裝置,並傳輸控制指令予該另一受控裝置,使其執行對應操作以及傳輸停止指令予原先執行控制指令的受控裝置(即錯誤的受控裝置),使其停止對應操作。其中,處理模組105可依據控制資料庫90篩選出另一受控裝置的方法可為但不限於說話者與該受控裝置之間的相對距離與/或相對方位為第二小或第二接近,實際的篩選機制可依據實際需求進行調整。Because the controlled device selected by the processing module 105 according to the above method (using the minimum or closest relative distance and / or relative orientation between the speaker and the controlled device) may not be the subject of the speaker's real item for voice control Therefore, the user can cause the intelligent voice control management system 100 to make corrections and adjustments by re-entering the same voice signal. In more detail, when a controlled device performs a corresponding operation according to a control instruction transmitted by the processing module 105 for a predetermined time (that is, after a wrong controlled device executes a control command for a predetermined time), the processing module 105 receives again To the same control instruction (that is, the user can input the same voice signal again) and judge the relative distance and relative position between each controlled device and the intelligent voice control management system 100, and the source position and direction of the voice signal. Time-varying (that is, the relative distance and / or relative orientation between the speaker and the controlled device remains the same), the processing module 105 can select another controlled device according to the control database 90, and transmit a control instruction to the other controlled device The controlled device causes it to perform the corresponding operation and transmits a stop instruction to the controlled device that originally executed the control instruction (ie, the wrong controlled device) to stop the corresponding operation. The processing module 105 may select another controlled device according to the control database 90. The method may be, but is not limited to, the relative distance and / or the relative orientation between the speaker and the controlled device is the second smallest or the second. Approach, the actual screening mechanism can be adjusted according to actual needs.
此外,為了讓處理模組105可以更為精準地篩選出說話者欲執行語音控制的受控裝置,智能語音控制管理系統100還可包括統計模組106,用以統計每一受控裝置依據控制指令執行對應操作的次數,為處理模組105篩選出對應該控制指令的受控裝置提供進一步的依據。換句話說,可透過統計模組106了解說話者其語音控制的習慣,使得處理模組105除了根據說話者與該受控裝置之間的相對距離與/或相對方位進行受控裝置的篩選外,也可參考說話者其語音控制的習慣。其中,統計模組106需針對不同說話者的語音控制習慣進行統計(可由聲紋辨識的方式得知不同的說話者身分,以進行後續語音控制的統計),處理模組105可將兩個參考依據(即說話者與該受控裝置之間的相對距離與/或相對方位、該說話者的語音控制習慣)以加權計算方式進行篩選受控裝置的評估,進而篩選出對應該控制指令的最佳受控裝置。舉例而言,可先針對說話者與該受控裝置之間的相對距離與/或相對方位、該說話者的語音控制習慣分別進行對應受控裝置的排名,再依據兩個參考依據的排名以加權方式進行計算,以篩選出對應該控制指令的最佳受控裝置。In addition, in order to allow the processing module 105 to more accurately screen out the controlled devices that the speaker wants to perform voice control, the intelligent voice control management system 100 may further include a statistics module 106 for counting the control of each controlled device. The number of times the instruction performs the corresponding operation provides a further basis for the processing module 105 to screen out the controlled device corresponding to the control instruction. In other words, the speaker's habits of voice control can be understood through the statistics module 106, so that the processing module 105 performs screening of controlled devices in addition to the relative distance and / or relative position between the speaker and the controlled device. , You can also refer to the speaker's habits of voice control. Among them, the statistics module 106 needs to perform statistics on the voice control habits of different speakers (the identity of different speakers can be obtained by means of voiceprint recognition for subsequent voice control statistics), and the processing module 105 can refer to two references Based on (the relative distance and / or relative position between the speaker and the controlled device, the speaker's voice control habits), a weighted calculation method is used to evaluate the screening of the controlled device, and then the most appropriate control instruction is selected. Best controlled device. For example, the relative distance and / or relative position between the speaker and the controlled device, the speaker's voice control habits, and the corresponding controlled device may be ranked separately. The calculation is performed in a weighted manner to screen out the best controlled device corresponding to the control instruction.
再者,本實施例的智能語音控制管理系統100更可包括學習模組107,用以統計每一受控裝置依據其對應的控制指令執行對應操作的發生時間,並在相同控制指令於相同的時間點或時間段重複發生超過預設次數時,自動在該時間點或時間段時使對應的受控裝置重複執行該控制指令。因此,可透過學習模組107了解說話者其語音控制的習慣(相同控制指令於相同的時間點或時間段重複發生超過預設次數),自動地在該時間點或時間段時生成並傳輸該控制指令予對應的該受控裝置,使對應的該受控裝置重複執行該控制指令。舉例而言,當說話者常透過智能語音控制管理系統100控制某一電視於下午五點開啟,且次數超過五次(即超過預設次數),則智能語音控制管理系統100可透過學習模組107自動在下午五點生成並傳輸該控制指令予該電視,使其開啟。Furthermore, the intelligent voice control management system 100 of this embodiment may further include a learning module 107 for counting the time when each controlled device performs a corresponding operation according to its corresponding control instruction, and in the same control instruction in the same When the time point or time period repeatedly occurs over a preset number of times, the corresponding controlled device is automatically caused to repeatedly execute the control instruction at the time point or time period. Therefore, the learning module 107 can be used to understand the speaker's habits of voice control (the same control instruction repeatedly occurs at the same time point or time period over a preset number of times), and automatically generates and transmits the same at that time point or time period. The control instruction is given to the corresponding controlled device, so that the corresponding controlled device repeatedly executes the control instruction. For example, when a speaker often controls a TV to be turned on at 5 pm through the intelligent voice control management system 100 and the number of times exceeds five times (that is, more than a preset number of times), the intelligent voice control management system 100 can use a learning module 107 automatically generates and transmits the control instruction to the TV at 5 pm to turn it on.
接著,請參閱「第5圖」,「第5圖」為「第1圖」的智能語音控制管理系統執行智能語音控制管理方法之一實施例方法流程圖,其步驟包括:多個麥克風單元接收一語音信號(步驟210);自語音信號取得聲音特徵,並依據聲紋資料庫對聲音特徵進行聲紋驗證(步驟220);當聲音特徵通過聲紋驗證後對語音信號進行語音轉文本識別與語義分析而獲得控制指令(步驟230);根據該些麥克風單元接收到語音信號的時間差與該些麥克風單元的相對配置位置取得語音信號的來源位置與來源方向(步驟240);依據定位信號與每一受控裝置所傳輸的位置信號取得每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置(步驟250);依據每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置、語音信號的來源位置與來源方向、控制資料庫與控制指令篩選出對應控制指令的受控裝置(步驟260);傳輸控制指令予對應的受控裝置,使其執行對應操作(步驟270)。其中,控制資料庫可包括每一受控裝置的名稱、位址與其對應的操作。Next, please refer to "Figure 5". "Figure 5" is a method flowchart of one embodiment of the intelligent voice control management method performed by the intelligent voice control management system of "Figure 1." The steps include: multiple microphone units receive A voice signal (step 210); obtaining a voice feature from the voice signal, and performing voiceprint verification on the voice feature according to the voiceprint database (step 220); performing voice-to-text recognition on the voice signal after the voice feature passes the voiceprint verification and Obtain a control instruction by semantic analysis (step 230); obtain the source position and direction of the voice signal according to the time difference between the time when the microphone units receive the voice signal and the relative position of the microphone units (step 240); according to the positioning signal and each The position signal transmitted by a controlled device obtains the relative distance and relative position between each controlled device and the intelligent voice control management system (step 250); according to the relative distance between each controlled device and the intelligent voice control management system Distance and relative position, source position and direction of voice signal, control database and control finger Selected controlled device (step 260) corresponding to a control command; transmission control instruction corresponding to the controlled device to execute a corresponding operation (step 270). The control database may include the name and address of each controlled device and the corresponding operations.
透過上述步驟,即可透過聲紋辨識安全地對區域範圍內的受控裝置(根據說話者與受控裝置之間的相對距離與/或相對方位進行受控裝置的篩選)進行語音控制。詳細描述已於上述段落加以說明,於此不再贅述。Through the above steps, voice control can be used to safely control the controlled devices within the area (selection of controlled devices based on the relative distance and / or relative orientation between the speaker and the controlled device) through voice recognition. The detailed description has been described in the above paragraphs, and will not be repeated here.
在本實施例中,智能語音控制管理方法還可包括:統計每一受控裝置依據控制指令執行對應操作的次數,為處理模組篩選出對應該控制指令的受控裝置提供進一步的依據(未繪製)。詳細描述已於上述段落加以說明,於此不再贅述。In this embodiment, the intelligent voice control management method may further include: counting the number of times each controlled device performs a corresponding operation according to the control instruction, and providing a further basis for the processing module to screen out the controlled device corresponding to the control instruction (not draw). The detailed description has been described in the above paragraphs, and will not be repeated here.
此外,在本實施例中,智能語音控制管理方法更可包括:當某一受控裝置依據處理模組所傳輸的控制指令執行對應操作一預定時間後,處理模組再次接收到相同的控制指令且判斷每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置與語音信號的來源位置與來源方向不變時,處理模組依據控制資料庫篩選出另一該受控裝置(未繪製);傳輸該控制指令予該另一受控裝置,使其執行對應操作(未繪製);以及傳輸停止指令予原先執行該控制指令的受控裝置,使其停止對應操作(未繪製)。詳細描述已於上述段落加以說明,於此不再贅述。In addition, in this embodiment, the intelligent voice control management method may further include: when a controlled device performs a corresponding operation according to a control instruction transmitted by the processing module for a predetermined time, the processing module receives the same control instruction again When determining the relative distance and relative position between each controlled device and the intelligent voice control management system and the source location and direction of the voice signal, the processing module selects another controlled device based on the control database ( (Not drawn); transmitting the control instruction to the other controlled device to perform the corresponding operation (not drawn); and transmitting a stop instruction to the controlled device that originally executed the control instruction to stop the corresponding operation (not drawn) . The detailed description has been described in the above paragraphs, and will not be repeated here.
再者,在本實施例中,智能語音控制管理方法也可包括:統計每一受控裝置依據其對應的控制指令執行對應操作的發生時間(未繪製);以及在相同控制指令於相同的時間點或時間段重複發生超過預設次數時,自動在該時間點或時間段時使生成並傳輸該控制指令予對應的該受控裝置,使對應的該受控裝置重複執行該控制指令(未繪製)。詳細描述已於上述段落加以說明,於此不再贅述。Furthermore, in this embodiment, the intelligent voice control management method may also include: counting the time when each controlled device performs a corresponding operation according to its corresponding control instruction (not drawn); and the same control instruction at the same time When the point or time period repeatedly occurs more than a preset number of times, the control instruction is automatically generated and transmitted to the corresponding controlled device at the time point or time period, so that the corresponding controlled device repeatedly executes the control instruction (not draw). The detailed description has been described in the above paragraphs, and will not be repeated here.
需要特別注意的是,上述智能語音控制管理方法中除了有說明其因果關係之外,可以依照任何順序執行上述步驟。It should be particularly noted that, in addition to explaining the causal relationship in the above intelligent voice control management method, the above steps may be performed in any order.
綜上所述,可知本發明與先前技術之間的差異在於透過接收的語音信號通過聲紋驗證後進行語音轉文本識別與語義分析而獲得控制指令;根據多個麥克風單元取得語音信號的來源位置與來源方向;取得每一受控裝置與該智能語音控制管理系統之間的相對距離與相對位置;依據每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置、語音信號的來源位置與來源方向、控制資料庫與控制指令篩選出對應控制指令的受控裝置;以及傳輸控制指令予對應的受控裝置,使其執行對應操作,藉由此一技術手段可以解決先前技術所存在的問題,進而達成透過聲紋辨識安全地對區域範圍內的受控裝置進行語音控制之技術功效。In summary, it can be seen that the difference between the present invention and the prior art is that the control instruction is obtained through speech-to-text recognition and semantic analysis after the received voice signal is verified by voiceprint verification; the source position of the voice signal is obtained according to multiple microphone units And source direction; obtain the relative distance and relative position between each controlled device and the intelligent voice control management system; according to the relative distance and relative position between each controlled device and the intelligent voice control management system, the The source location and source direction, the control database and the control instruction screen out the controlled devices corresponding to the control instruction; and transmit the control instruction to the corresponding controlled device to perform the corresponding operation, by which a technical means can solve the problem of the previous technology The existing problems further achieve the technical effect of voice control of controlled devices in the area safely through voiceprint recognition.
雖然本發明以前述之實施例揭露如上,然其並非用以限定本發明,任何熟習相像技藝者,在不脫離本發明之精神和範圍內,當可作些許之更動與潤飾,因此本發明之專利保護範圍須視本說明書所附之申請專利範圍所界定者為準。Although the present invention is disclosed in the foregoing embodiments as above, it is not intended to limit the present invention. Any person skilled in similar arts can make some modifications and retouches without departing from the spirit and scope of the present invention. The scope of patent protection shall be determined by the scope of the patent application attached to this specification.
50‧‧‧麥克風單元50‧‧‧ microphone unit
60‧‧‧聲紋資料庫60‧‧‧ Voiceprint Database
70‧‧‧語義資料庫70‧‧‧ Semantic Database
90‧‧‧控制資料庫90‧‧‧Control database
100‧‧‧智能語音控制管理系統100‧‧‧ Intelligent Voice Control Management System
101‧‧‧接收模組101‧‧‧Receiving module
102‧‧‧聲紋辨識模組102‧‧‧Voiceprint recognition module
103‧‧‧語義解析模組103‧‧‧Semantic parsing module
104‧‧‧定位模組104‧‧‧Positioning module
105‧‧‧處理模組105‧‧‧Processing Module
106‧‧‧統計模組106‧‧‧ Statistics Module
107‧‧‧學習模組107‧‧‧Learning Module
步驟210‧‧‧多個麥克風單元接收一語音信號Step 210‧‧‧ multiple microphone units receive a voice signal
步驟220‧‧‧自語音信號取得聲音特徵,並依據聲紋資料庫對聲音特徵進行聲紋驗證Step 220‧‧‧ Obtain a voice feature from the voice signal, and perform voiceprint verification on the voice feature according to the voiceprint database
步驟230‧‧‧當聲音特徵通過聲紋驗證後對語音信號進行語音轉文本識別與語義分析而獲得控制指令Step 230‧‧‧ After voice features are verified by voiceprint, voice-to-text recognition and semantic analysis are performed on the voice signal to obtain control instructions
步驟240‧‧‧根據該些麥克風單元接收到語音信號的時間差與該些麥克風單元的相對配置位置取得語音信號的來源位置與來源方向Step 240‧‧‧ Obtain the source position and source direction of the voice signal according to the time difference between the microphone units receiving the voice signals and the relative position of the microphone units
步驟250‧‧‧依據定位信號與每一受控裝置所傳輸的位置信號取得每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置Step 250‧‧‧ Obtain the relative distance and relative position between each controlled device and the intelligent voice control management system according to the positioning signal and the position signal transmitted by each controlled device
步驟260‧‧‧依據每一受控裝置與智能語音控制管理系統之間的相對距離與相對位置、語音信號的來源位置與來源方向、控制資料庫與控制指令篩選出對應控制指令的受控裝置Step 260‧‧‧ Screen out the controlled device corresponding to the control instruction according to the relative distance and relative position between each controlled device and the intelligent voice control management system, the source position and direction of the voice signal, the control database and the control instruction
步驟270‧‧‧傳輸控制指令予對應的受控裝置,使其執行對應操作Step 270‧‧‧ transmits a control instruction to the corresponding controlled device, so that it performs the corresponding operation
第1圖為本發明智能語音控制管理系統之一實施例系統方塊圖。 第2圖為第1圖的智能語音控制管理系統的語義資料庫所記錄之字彙與類別的一實施例對照表。 第3A圖為第1圖的智能語音控制管理系統中控制指令的一實施例格式。 第3B圖為第1圖的智能語音控制管理系統中控制指令的一實施例內容。 第4圖為第1圖的智能語音控制管理系統的控制資料庫所記錄之內容的一實施例示意圖。 第5圖為第1圖的智能語音控制管理系統執行智能語音控制管理方法之一實施例方法流程圖。FIG. 1 is a system block diagram of an embodiment of an intelligent voice control management system according to the present invention. FIG. 2 is a comparison table of the embodiment of the vocabulary and category recorded in the semantic database of the intelligent voice control management system of FIG. 1. FIG. 3A is a format of an embodiment of a control instruction in the intelligent voice control management system of FIG. 1. FIG. 3B is an embodiment of the control instructions in the intelligent voice control management system of FIG. 1. FIG. 4 is a schematic diagram of an embodiment of the content recorded in the control database of the intelligent voice control management system of FIG. 1. FIG. 5 is a method flowchart of an embodiment of the intelligent voice control management system performed by the intelligent voice control management system of FIG. 1.
Claims (8)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| TW107125806A TWI675305B (en) | 2018-07-26 | 2018-07-26 | Intelligent voice control management system and method thereof |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| TW107125806A TWI675305B (en) | 2018-07-26 | 2018-07-26 | Intelligent voice control management system and method thereof |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| TWI675305B true TWI675305B (en) | 2019-10-21 |
| TW202008192A TW202008192A (en) | 2020-02-16 |
Family
ID=69024004
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| TW107125806A TWI675305B (en) | 2018-07-26 | 2018-07-26 | Intelligent voice control management system and method thereof |
Country Status (1)
| Country | Link |
|---|---|
| TW (1) | TWI675305B (en) |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| TW201349004A (en) * | 2012-05-23 | 2013-12-01 | Transcend Information Inc | Voice control method and computer-implemented system for data management and protection |
| TW201506679A (en) * | 2013-08-02 | 2015-02-16 | Mstar Semiconductor Inc | Controller for voice-controlled device and associated method |
| TW201717655A (en) * | 2015-11-05 | 2017-05-16 | 宏碁股份有限公司 | Voice control method and voice control system |
-
2018
- 2018-07-26 TW TW107125806A patent/TWI675305B/en active
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| TW201349004A (en) * | 2012-05-23 | 2013-12-01 | Transcend Information Inc | Voice control method and computer-implemented system for data management and protection |
| TW201506679A (en) * | 2013-08-02 | 2015-02-16 | Mstar Semiconductor Inc | Controller for voice-controlled device and associated method |
| TW201717655A (en) * | 2015-11-05 | 2017-05-16 | 宏碁股份有限公司 | Voice control method and voice control system |
Also Published As
| Publication number | Publication date |
|---|---|
| TW202008192A (en) | 2020-02-16 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12149781B2 (en) | Methods and systems for detecting audio output of associated device | |
| US11138977B1 (en) | Determining device groups | |
| US11790912B2 (en) | Phoneme recognizer customizable keyword spotting system with keyword adaptation | |
| US9424837B2 (en) | Voice authentication and speech recognition system and method | |
| US11848029B2 (en) | Method and device for detecting audio signal, and storage medium | |
| CN109564759B (en) | speaker recognition | |
| US10950245B2 (en) | Generating prompts for user vocalisation for biometric speaker recognition | |
| US9443527B1 (en) | Speech recognition capability generation and control | |
| US20160372116A1 (en) | Voice authentication and speech recognition system and method | |
| CN109378006A (en) | A method and system for cross-device voiceprint recognition | |
| US20240005918A1 (en) | System For Recognizing and Responding to Environmental Noises | |
| AU2013203139A1 (en) | Voice authentication and speech recognition system and method | |
| KR20160115944A (en) | Systems and methods for evaluating strength of an audio password | |
| US20240079007A1 (en) | System and method for detecting a wakeup command for a voice assistant | |
| JP2024510798A (en) | Hybrid multilingual text-dependent and text-independent speaker verification | |
| JP2005283647A (en) | Emotion recognition device | |
| CN115373280B (en) | Remote voice control method, device and system | |
| US12190877B1 (en) | Device arbitration for speech processing | |
| CN108831468A (en) | Intelligent sound Control management system and its method | |
| TWI675305B (en) | Intelligent voice control management system and method thereof | |
| CN115376521A (en) | Voiceprint recognition method, device and computer-readable storage medium | |
| Baimirov et al. | Overview of the latest research related to smart speakers | |
| CN119724217A (en) | Speaker separation method based on voiceprint recognition in intelligent conference system | |
| CN119892544A (en) | Method, device, equipment, medium and product for equipment configuration | |
| KR20240073991A (en) | Voice synthesis service provision method and system |