WO2021080362A1 - Système de traitement de langue utilisant un écouteur - Google Patents
Système de traitement de langue utilisant un écouteur Download PDFInfo
- Publication number
- WO2021080362A1 WO2021080362A1 PCT/KR2020/014544 KR2020014544W WO2021080362A1 WO 2021080362 A1 WO2021080362 A1 WO 2021080362A1 KR 2020014544 W KR2020014544 W KR 2020014544W WO 2021080362 A1 WO2021080362 A1 WO 2021080362A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- language
- function
- interpretation
- voice
- earset
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/58—Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
- G06Q50/20—Education
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/04—Electrically-operated educational appliances with audible presentation of the material to be studied
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/02—Casings; Cabinets ; Supports therefor; Mountings therein
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
Definitions
- the present invention relates to a language processing system using an earset, and in particular, an earset that blocks the inflow of external noise to obtain a clearer voice of the user, and performs an interpreter for the acquired voice or a translation and learning function for the acquired voice. It relates to a language processing system using an earset, and in particular, an earset that blocks the inflow of external noise to obtain a clearer voice of the user, and performs an interpreter for the acquired voice or a translation and learning function for the acquired voice. It relates to a language processing system using an earset, and in particular, an earset that blocks the inflow of external noise to obtain a clearer voice of the user, and performs an interpreter for the acquired voice or a translation and learning function for the acquired voice. It relates to a language processing system using an earset, and in particular, an earset that blocks the inflow of external noise to obtain a clearer voice of the user, and performs an interpreter for the acquired voice or a translation and learning function for
- the best practice of foreign language is repetitive and accurate writing correction, hearing and speaking practice by experts, but it can also be supplemented by using a learning device.
- the present invention relates to a language processing system using an earset, and in particular, an earset that blocks the inflow of external noise to obtain a clearer voice of the user, and performs an interpreter for the acquired voice or a translation and learning function for the acquired voice. It is an object of the present invention to provide a language processing system using an earset, and in particular, an earset that blocks the inflow of external noise to obtain a clearer voice of the user, and performs an interpreter for the acquired voice or a translation and learning function for the acquired voice. It is an object of the present invention to provide a language processing system using an earset, and in particular, an earset that blocks the inflow of external noise to obtain a clearer voice of the user, and performs an interpreter for the acquired voice or a translation and learning function for the acquired voice. It is an object of the present invention to provide a language processing system using an earset, and in particular, an earset that blocks the inflow of external noise to obtain a clearer voice of the user,
- the language processing system using the earset of the present invention includes a first earset or a second earset, a wearable sound device performing wireless communication with the first earset or wired communication with the second earset, and a first earset or a wearable sound device wirelessly. It is composed of an electronic communication device including a communication unit that performs communication and communicates with a translation server, and a display unit, and the electronic communication device is a processing including a speech signal that is a processing target language received from a first earphone or a wearable sound device.
- Target information is generated and transmitted to the translation server, and processing information including translation text, which is a processing target language corresponding to the processing target information transmitted from the translation server, or audio signal converted from the translated text is received and visually displayed through the display unit. Or perform a speech processing function that is expressed aurally.
- the language processing function includes an interpretation function
- the processing information corresponds to the interpretation information
- the wearable sound device and the electronic communication device are in a communication enabled state
- the wearable sound device responds to the voice signal from the first earset or the second earset.
- an interpretation function control command corresponding to the reference voice signal is generated and transmitted to the electronic communication device, and the electronic communication device receives the interpretation function control command, and the interpretation function control command It is desirable to perform the start or end of the interpreting function in response.
- the electronic communication device wakes up the application for the interpretation function when receiving an interpretation function control command including starting the interpretation function while operating the application for the interpretation function in the foreground service state in the standby state.
- the electronic communication device receives an interpretation function control command including termination of the interpretation function while the electronic communication device is performing the interpretation function, it is preferable to terminate the interpretation function by operating the application for the interpretation function in the foreground service state. Do.
- the electronic communication device performs an interpretation function by opening a voice communication channel with the wearable sound device.
- the language processing function includes a language learning function
- the processing information corresponds to the translation information
- the electronic communication device includes an input unit, and after the display of the translation information, the selection of success in learning the first language through the input unit or It is desirable to acquire and store the learning failure selection.
- the electronic communication device receives the translation target text as a second language through the input unit, transmits the translation target text information to the translation server, and transmits the first text corresponding to the translation target text information transmitted from the translation server. It is preferable to receive the translated text, which is a language, and display it through the display unit.
- the first and second earphones define an installation space in which parts are installed, and a housing that forms an exterior and has a soundproofing hole, a sound reproduction unit installed in the installation space and emitting sound, a voice microphone installed in the installation space, It is installed in the housing, it is preferable to include a; a conduit for transmitting the voice transmitted through the soundproof hole to the voice microphone.
- the housing includes an insertion tube inserted into the user's external ear canal, and it is preferable that the insertion tube serves as a soundproofing device.
- it further includes a chamber forming a closed space surrounding the voice microphone, and the conduit is preferably formed in the chamber to transmit the voice transmitted through the insertion tube to the voice microphone.
- the chamber has an upper bracket that fixes the voice microphone in the installation space, and a lower bracket that meshes with the upper bracket to form a space, and the sound reproduction unit is installed between the upper bracket and the lower bracket, and surrounds the sound reproduction unit. It is preferable that the closed space and the installation space of the voice microphone are separated from each other.
- the housing has a back hole communicating with the rear surface of the sound reproduction unit.
- At least one bracket installed between the sound reproduction unit and the housing and capable of tuning acoustic characteristics.
- the conduit extends into the soundproofing hole.
- a clearer voice of the user can be obtained, and interpretation for the obtained voice can be performed.
- the interpretation function is performed by executing an application for an interpretation function by the user's voice.
- the present invention has an effect of blocking the inflow of external noise to obtain a clearer voice of the wearer, and helping the wearer learn language by performing translation on the acquired voice.
- the voice transmitted from the user's ear can be input more clearly.
- the present invention has an advantage in that it is possible to prevent external noise from flowing into the inner microphone by forming a chamber that closes the rear of the voice microphone.
- FIG. 1 is a control configuration diagram of a language processing system using an ear set according to the present invention.
- FIG. 2 is a cross-sectional view of an earset according to a first embodiment of the present invention.
- FIG 3 is an exploded view of an earset according to a second embodiment of the present invention.
- FIG. 4 is a perspective view of an earset according to a second embodiment of the present invention.
- FIG. 5 is a cross-sectional view of an earset according to a second embodiment of the present invention.
- FIG. 6 is a cross-sectional view of an earset according to a third embodiment of the present invention.
- expressions such as “A or B”, “at least one of A or/and B”, or “one or more of A or/and B” may include all possible combinations of the items listed together.
- “A or B”, “at least one of A and B”, or “at least one of A or B” includes (1) at least one A, (2) at least one B, Or (3) it may refer to all cases including both at least one A and at least one B.
- first, second, first, or “second” used in this document can modify various elements, regardless of their order and/or importance, and It is used to distinguish it from the component, but does not limit the component.
- a first user device and a second user device may represent different user devices regardless of order or importance.
- a first component may be referred to as a second component, and similarly, a second component may be renamed to a first component.
- Some component eg, the first component
- another component eg, the second component
- the certain component may be directly connected to the other component or may be connected through another component (eg, a third component).
- a component eg, a first component
- the component and the It may be understood that no other component (eg, a third component) exists between the different components.
- a device configured to may mean that the device “can” along with other devices or parts.
- a processor configured (or configured) to perform A, B, and C means a dedicated processor (eg, an embedded processor) for performing the corresponding operation, or executing one or more software programs stored in a memory device. By doing so, it may mean a generic-purpose processor (eg, a CPU or an application processor) capable of performing corresponding operations.
- the language processing system in the present invention performs a language processing function, but it should be recognized that the language processing function includes an interpretation function and a translation function, and additionally includes a language learning function.
- the language to be interpreted corresponds to the language to be translated or interpreted into the language spoken by the user
- the language to be interpreted is the language spoken by the user to be translated or interpreted. This corresponds to the language to be finally conveyed visually or aurally to the conversation partner.
- the target language for interpretation is English
- the target language for interpretation is Korean.
- the first language corresponds to a language to be translated as a language to be learned (a processing target language), and the second language is translated as a language used by the user.
- the selected language the target language for processing. For example, when a user who speaks Korean wants to learn English, the first language becomes English and the second language becomes Korean.
- FIG. 1 is a control configuration diagram of a language processing system using an ear set according to the present invention.
- the language processing system includes a first earset 10 for performing wireless communication with the wearable sound device 30 or the electronic communication device 40, and a second earset 20 electrically connected to the wearable sound device 30 through wired communication. And, a wearable sound device 30 that performs wireless communication with the electronic communication device 40 and performs wireless communication with the first ear set 10 and performs wired communication with the second ear set 20, and the first ear set.
- an electronic communication device 40 that performs wireless communication with the wearable sound device 30 and communicates with the translation server 50 through the network 60, and the electronic communication device ( Interpretation information (or processing information) including voice signals converted from text by performing communication with 40) and translating the voice included in the voice signal included by receiving the interpretation target information (or processing target information) It is configured to include a translation (or interpreter) server 50 to provide a.
- the first earphone 10 includes a microphone 11 that acquires a user's voice, a speaker 13 (or a receiver) that emits sound by receiving an electrical signal, and a communication for performing wireless communication with the wearable sound device 30 And a module 15 (for example, a wireless communication module such as a Bluetooth module).
- a module 15 for example, a wireless communication module such as a Bluetooth module.
- the configuration or function of the power supply unit (not shown) for supplying power, the microphone 11 and the speaker 13 corresponds to a technology that is naturally recognized by a person skilled in the art to which the present invention belongs, and the description thereof is omitted.
- FIGS. 2 to 6 the mechanical structure of the first ear set 10 will be described in detail.
- the communication module 15 performs a phone call function and a sound reproduction function, and performs an interpretation function according to the present invention, as already known to those skilled in the art to which the present invention pertains.
- the communication module 15 transmits the user's voice signal obtained from the microphone 11 to the wearable sound device 30 or the electronic communication device 40 in a communication connection state with the wearable sound device 30, and the wearable sound device 30 or an electrical signal including an audio signal is received from the electronic communication device 40 and sound is emitted through the speaker 13.
- the second earphone 20 is a connection for performing wired communication with a microphone 21 that acquires a user's voice, a speaker 23 (or a receiver) that emits sound by receiving an electrical signal, and a wearable sound device 30 It comprises a cable 24 (for example, a wired cable, etc.).
- a cable 24 for example, a wired cable, etc.
- the configuration or function of the microphone 21, the speaker 23, and the connection cable 24 corresponds to a technology that is naturally recognized by a person skilled in the art to which the present invention belongs, and the description thereof will be omitted.
- FIGS. 2 to 6 the mechanical structure of the second ear set 20 will be described in detail.
- the wearable sound device 30 is a device that includes a wireless communication function, such as a neckband type sound conversion device, and performs a phone call function, a sound reproduction function, and the like.
- the wearable sound device 30 includes a microphone 31 that acquires external sound, a speaker 33 that emits sound by receiving an electric signal, and wireless communication with the first earset 10 and the electronic communication device 40 ( For example, a communication unit 35 for performing Bluetooth communication), an input unit 37 for acquiring an input from a user, a microphone 31, a speaker 33, a communication unit 35, and an input unit 37 It is configured to include a data processor 39 for selectively performing a phone call function, a sound reproduction function, and an interpretation function by controlling.
- the configuration or function of the power supply unit (not shown) that supplies power, the microphone 31 and the speaker 33, the communication unit 35, and the input unit 37 are naturally recognized by those skilled in the art to which the present invention belongs. The description is omitted because it corresponds to the technology to be used.
- the data processor 39 is a processor that performs a phone call function and a sound reproduction function, and performs an interpretation function according to the present invention, as already known to those of ordinary skill in the art to which the present invention pertains (for example, CPU, MCU, MICROPROCESSOR, etc.), and a storage space (eg, memory, etc.) for storing voice signals, interpretation performance information (processing performance information), and the like.
- a storage space eg, memory, etc.
- Interpretation performance information is at least an audio signal for starting the interpretation function (eg, an audio signal for'interpretation start') and an audio signal for ending the interpretation function (eg, an audio signal for'interpretation end', etc.) It is configured to include a reference speech signal for identifying.
- the wearable sound device 30 may maintain a state capable of communicating with at least one or more of the first and second earphones 10 (eg, a wireless communication connection state or a wired communication connection state).
- a state capable of communicating with at least one or more of the first and second earphones 10 eg, a wireless communication connection state or a wired communication connection state.
- the data processor 39 causes the communication module 15 to transmit the voice signal acquired through the microphone 11 to the communication unit 35 in a communication connection state with the first earset 10.
- the data processor 39 acquires a voice signal from the microphone 21 inserted into the user's ear and the voice signal from the microphone 31 facing the outside of the user.
- the voice signal (first voice signal) obtained from the microphone 11 and the microphone 21 is the user's voice
- the voice signal (second voice signal) obtained from the microphone 31 is the user's voice. It is treated as the voice of the other party.
- the data processor 39 may transmit conversational identification information capable of identifying each of the first and second voice signals together when transmitting the interpretation target information.
- the data processor 39 is performing a mode or function other than a phone call function and a sound reproduction function (for example, a standby mode, an interpretation function, etc.), while performing a voice signal from the first ear set 10 or a second ear set ( 20) by comparing the voice signal from the voice signal and the reference voice signal, it is determined whether the voice signal includes or corresponds to the reference voice signal for starting the interpretation function or ending the interpretation function. If the voice signal includes or corresponds to a reference voice signal for starting an interpretation function or ending the interpretation function, the data processor 39 generates an interpretation function control command corresponding to the start or end of the interpretation function, which is a reference voice signal, and communicates electronically. Apply to the device 40.
- a mode or function other than a phone call function and a sound reproduction function for example, a standby mode, an interpretation function, etc.
- the data processor 39 performs a mode or function (e.g., a standby mode) other than a phone call function and a sound reproduction function. , Interpreting functions, etc.) are performed in the same way.
- a mode or function e.g., a standby mode
- the data processor 39 is a voice communication channel (for example, SCO: Synchronous Connection-Oriented) between the communication unit 35 and the communication unit 45 by the interpretation function started by the electronic communication device 40. Transmit information to be interpreted and receive interpretation information. More detailed interpretation functions are described below.
- SCO Synchronous Connection-Oriented
- the electronic communication device 40 corresponds to, for example, an information communication device such as a smartphone or tablet having a communication function, and input from a user (eg, selection of the start or end of an application for an interpreter function, interpretation)
- An input unit 41 that acquires a function start or end selection, selection input of an interpretation target language and/or an interpretation target language, etc.) and applies it to the data processor 49, and a user interface for an interpretation function visually or aurally.
- a communication unit 45 for performing communication a microphone 46 for acquiring voice or sound
- a data processor 49 for performing a phone call function and a sound reproduction function for performing an interpretation function according to the present invention. It is composed.
- the configuration and function of the power supply unit (not shown), the input unit 41, the display unit 43, the microphone 46, and the communication unit 45 supplying power are naturally recognized by those skilled in the art to which the present invention belongs. The description is omitted because it corresponds to the technology to be used.
- the data processor 49 is a processor (e.g., CPU, MCU, MICROPROCESSOR, etc.) that performs a phone call function, a sound reproduction function, and an interpretation function, and an application for an interpretation function, a user interface, and a storage for storing interpretation information, etc. It is configured to include space (eg, memory, etc.).
- processor e.g., CPU, MCU, MICROPROCESSOR, etc.
- space e.g., memory, etc.
- the data processor 49 executes an application for an interpretation function.
- the application for the interpretation function includes a process of selecting and setting an interpretation target language and/or an interpretation target language, and a process of generating and transmitting interpretation target information including voice information of a user, which is an interpretation target language, to the interpretation server 50, It includes a process of receiving interpretation information including voice information that is a target language for interpretation from the interpretation server 50 and transmitting it to the wearable sound device 30.
- the data processor 49 activates or executes an application for an interpretation function in a foreground service state according to an execution input from the input unit 41, and a standby state in which the phone call function and sound reproduction function are not performed.
- an interpretation function control command processing function control command
- processing function control command for example, an interpretation function start command
- the application for the interpretation function wakes up without an additional user's button or touch input.
- the data processor 49 is an additional user's button or touch according to an interpretation function control command (processing function control command) (e.g., an interpretation function termination command) from the wearable sound device 30 while the interpretation function is being performed.
- processing function control command processing function control command
- the application for the interpretation function is terminated or the interpreter function is terminated by operating in the foreground service state, and it operates as a standby state, a phone call function, or a sound reproduction function.
- the data processor 19 controls the communication unit 45 to control the interpretation target information including the voice information received from the first earphone 10 or the wearable sound device 30 while performing the interpretation function through the network 50.
- Interpretation functions such as transmitting to the translation server 50 and controlling the communication unit 45 to receive interpretation information including an interpretation target language from the translation server 50 through the network 50 will be described in detail below. do.
- the translation server 50 has a STT (Speech to Text) function (a function of extracting voice information included in the interpretation target information and recognizing it and converting it to text) and/or a function of translating a text to generate a translated text, and/or As a server including a TTS (Text to Speech) function (a function for synthesizing text into speech), such a translation server 50 corresponds to a technology naturally recognized by a person skilled in the art to which the present invention belongs, and its detailed Description is omitted.
- STT Seech to Text
- TTS Text to Speech
- the network 60 corresponds to a system for performing wired communication and/or wireless communication, and corresponds to a technology that is naturally recognized by a person skilled in the art to which the present invention pertains, and a detailed description thereof is omitted.
- the language processing system performs the interpreter function by performing the following process.
- the wearable sound device 30 is in a state capable of communicating with the first ear set 10 and/or the second ear set 20, and the electronic communication device 40 communicates with the first ear set 10 or the wearable sound device 40 It is possible.
- the data processor 49 operates an application for an interpretation function in a foreground service state.
- the data processor 49 obtains an interpretation function start command from the wearable sound device 30 through the communication unit 45 while maintaining the standby state, the application for the interpretation function wakes up to perform the interpretation function. Start.
- the data processor 49 enables the user to set the interpretation target language and/or the interpretation target language by voice or input by the input unit 41.
- the data processor 49 stores information on the language to be interpreted (eg, type of language) and information on the target language for interpretation (eg, type of language).
- the data processor 49 controls the communication unit 45 to open a voice communication channel with the communication unit 35 to enable transmission and reception of interpretation target information and interpretation information.
- the data processor 49 receives the voice signal (first or second voice signal) and/or the speaker identification information received from the first earphone 10 or the wearable sound device 30 while performing the interpretation function, and the voice signal And/or the interpretation target information including the speaker identification information is transmitted to the translation server 50 through the network 50 by controlling the communication unit 45.
- the translation server 50 translates the voice signal included in the interpretation target information into text, converts the text into a voice signal, and transmits the converted voice signal and/or interpretation information including the speaker identification information to the electronic communication device 40. send.
- the data processor 49 receives interpretation information through the communication unit 45 and transmits it to the wearable sound device 30.
- the data processor 39 receives interpretation information through the communication unit 35, and applies the converted voice signal included according to the speaker identification information included in the interpretation information to the speaker 33 or the first or second earphone 10 , 20). That is, if the conversational party identification information represents the user's voice signal, the conversation partner must listen to the converted audio signal, and therefore, when the conversation party identification information indicates the conversation partner’s voice signal, the user converts it. Since it is necessary to listen to the obtained voice signal, the converted voice is transmitted or applied to the first or second earphones 10 and 20.
- the data processor 39 does not transmit the interpretation target information to the electronic communication device 40 during sound emission by receiving interpretation information corresponding to the interpretation target information transmitted immediately before.
- the voice signal that has already been interpreted is again included in the interpretation target information to prevent transmission.
- the data processor 49 When the data processor 49 receives an interpretation function termination command from the wearable sound device 30 while performing the interpretation function, the data processor 49 terminates the application for the interpretation function or operates in a foreground service state to terminate the interpretation function and wait. It operates as a state or a phone call function or a sound reproduction function.
- the language processing system includes a first earset 10 that performs wired or wireless communication with the wearable sound device 30 or the electronic communication device 40, and a second earset that is electrically connected to the wearable sound device 30 through wired communication ( 20), a wearable sound device 30 that performs wireless communication with the electronic communication device 40 and performs wired communication with the second ear set 20, and the first ear set 10 and/or the wearable sound device 30 )
- the electronic communication device 40 that performs wireless communication with each and communicates with the translation server 50 through the network 60, and the electronic communication device 40 through the network 60, It is configured to include a translation server 50 that receives the voice information, translates the voice included in the voice information into text, and provides translation information (or processing information) corresponding to the voice information.
- the first earset 10 includes a microphone 11 for acquiring a voice, a speaker 13 (or a receiver) for emitting sound by receiving an electric signal, and a communication module 15 for performing communication with the electronic communication device 40.
- a wireless communication module such as a Bluetooth module or a wired cable.
- the configuration or function of the power supply unit (not shown) for supplying power, the microphone 11 and the speaker 13 corresponds to a technology that is naturally recognized by a person skilled in the art to which the present invention belongs, and the description thereof is omitted.
- FIGS. 2 to 6 the mechanical structure of the first ear set 10 will be described in detail.
- the communication module 15 performs a phone call function and a sound reproduction function, and performs a language learning function according to the present invention, as already known to those of ordinary skill in the art to which the present invention pertains.
- the communication module 15 transmits the voice information including the user's voice signal acquired from the microphone 11 to the electronic communication device 40 when performing the language learning function.
- the second earphone 20 includes a microphone 21 for acquiring a voice, a speaker 23 (or a receiver) that emits sound by receiving an electric signal, and a connection cable for performing wired communication with the wearable sound device 30 ( 24) (for example, a wired cable, etc.).
- a microphone 21 for acquiring a voice
- a speaker 23 or a receiver
- a connection cable for performing wired communication with the wearable sound device 30 ( 24) (for example, a wired cable, etc.).
- the configuration or function of the microphone 21, the speaker 23, and the connection cable 24 corresponds to a technology that is naturally recognized by a person skilled in the art to which the present invention belongs, and the description thereof will be omitted.
- FIGS. 2 to 6 the mechanical structure of the second ear set 20 will be described in detail.
- the wearable sound device 30 is a device that includes a wireless communication function, such as a neckband type sound conversion device, and performs a phone call function, a sound reproduction function, and the like.
- the wearable sound device 30 includes a microphone 31 that acquires external sound, a speaker 33 that emits sound by receiving an electric signal, and wireless communication (for example, Bluetooth communication, etc.) ), the input unit 37 for acquiring input from the user, the microphone 31, the speaker 33, the communication unit 35, and the input unit 37 to control the phone call function, sound And a data processor 39 that selectively performs a reproduction function and a language learning function.
- the configuration or function of the power supply unit (not shown) that supplies power, the microphone 31 and the speaker 33, the communication unit 35, and the input unit 37 are naturally recognized by those skilled in the art to which the present invention belongs. The description is omitted because it corresponds to the technology to be used.
- the data processor 39 is a processor that performs a phone call function and a sound reproduction function, and performs a language learning function according to the present invention, as already known to a person skilled in the art. , CPU, MCU, MICROPROCESSOR, etc.).
- the data processor 39 transmits the voice information including the user's voice signal obtained from the microphone 21 to the electronic communication device 40 when performing the language learning function.
- the language learning functions performed by the data processor 39 are described in detail below.
- the electronic communication device 40 corresponds to, for example, an information communication device such as a smartphone or tablet having a communication function, and input from a user (for example, selection of the start or end of a language learning function, a language to be learned).
- An input unit 41 that acquires selection of, evaluation selection for translation information (learning success, learning failure), input of words or sentences composed of a language to be studied, etc., and applies it to the data processor 49, and a language learning function.
- Wireless communication for example, Bluetooth communication, etc.
- a communication unit 45 that communicates with the translation server 50 through the network 60, a microphone 46 that acquires voice or sound, a phone call function and a sound reproduction function, and a language according to the present invention. It is configured to include a data processor 49 performing a learning function.
- the configuration and function of the power supply unit (not shown), the input unit 41, the display unit 43, the microphone 46, and the communication unit 45 supplying power are naturally recognized by those skilled in the art to which the present invention belongs. The description is omitted because it corresponds to the technology to be used.
- the data processor 49 includes a processor (eg, CPU, MCU, MICROPROCESSOR, etc.) that performs a phone call function, a sound reproduction function, and a language learning function, and an application and user interface for a language learning function, translation information, and translation text. It is configured to include a storage space (eg, memory, etc.) for storing the and the like.
- the data processor 19 controls the communication unit 45 to transmit information to be translated, including voice information received from the first earset 10 or the wearable sound device 30, and the translation server 50 through the network 50. Language learning functions such as transmission to and controlling the communication unit 45 to receive translation information from the translation server 50 through the network 50 will be described in detail below.
- the translation server 50 generates translation information including the translated text by translating the text and/or the STT (Speech to Text) function (a function to extract voice information included in the translation target information and recognize it and convert it into text).
- STT Seech to Text
- a server including a function to perform and/or a TTS (Text to Speech) function (a function to synthesize text into speech)
- TTS Text to Speech
- a translation server 50 is a technology that is naturally recognized by a person skilled in the art to which the present invention belongs. Corresponds to the detailed description thereof will be omitted.
- the network 60 corresponds to a system for performing wired communication and/or wireless communication, and corresponds to a technology that is naturally recognized by a person skilled in the art to which the present invention pertains, and a detailed description thereof is omitted.
- the language processing system performs the language learning function by performing the following process.
- the data processor 49 controls the communication unit 45 to perform a pairing operation with the first ear set 10 or the wearable sound device 30 to enter a communication enabled state. That is, the communication unit 45 performs wireless communication with the communication module 15 or the communication unit 35.
- the data processor 49 executes an application for the language learning function according to the start selection input of the language learning function from the input unit 41 and displays a user interface on the display unit 43.
- the data processor 39 obtains a start selection input of the language learning function from the input unit 37, and transmits a start selection input of the language learning function through the communication unit 45 to the electronic communication device 40 through the communication unit 35.
- the data processor 49 executes an application for the language learning function according to the received start selection input of the language learning function, and displays the user interface on the display unit 43.
- the data processor 49 may start an application for a language learning function in various other ways.
- the data processor 49 displays a language to be learned that can be selected in the user interface displayed on the display unit 43 and stores a language to be learned (ie, a first language) by a language selection input from the input unit 41.
- the data processor 49 enables a user's voice command, that is, checks the language included in the voice input through the microphone 46, and selects the identified language (eg, English, Chinese, etc.). Save as the language to be learned (first language).
- the data processor 49 may visually and/or audibly display the selection of the language to be learned to the user through the display unit 43.
- the data processor 49 displays the language to be learned on the display unit 43, and prompts the user to speak a word or sentence to be learned through the display unit 43 visually and/or aurally.
- the second language may be preset or the data processor 49 may allow the user to set it in the same manner as above.
- the data processor 49 communicates with the communication module 15 through the communication unit 45 to perform a language learning function, or communicates with the communication unit 35 to allow the data processor 39 to perform a language learning function.
- the communication module 15 of the first earset 10 electronically communicates first voice information including a voice signal (first voice signal) that is the user's first language in which external noise obtained by the microphone 11 is reduced. Transmitted to the device 40, or the data processor 39 includes a voice signal (a second voice signal) that is the first language of the user in which the external noise obtained by the microphone 21 of the second ear set 20 has been reduced.
- the second voice information to be transmitted is transmitted to the electronic communication device 40 through the communication unit 35.
- the data processor 49 receives and stores the first or second voice information through the communication unit 45, and generates translation target information (processing target information) including the first or second voice information, and the communication unit 45 And transmits it to the translation server 50 through the network 60.
- the translation target information includes a first or second voice signal, a first language type code, and a second language type code.
- the translation server 50 receives and stores the translation target information, extracts the first or second voice signal included in the translation target information, recognizes it, and converts it into text by referring to the language type code included in the translation target information. Save it.
- the translation server 50 translates the text into a second language, generates and stores the translated text, which is a second language, and generates translation information (processing information) including the translated text, and communicates electronically through the network 60. It is transmitted to the device 40.
- the data processor 49 receives and stores the translation information through the communication unit 45, and displays the translated text included in the translation information through the display unit 43 to display the contents spoken by the user in the first language and the displayed second language. Make it possible to compare and learn the translated texts.
- the data processor 49 allows the user to select whether or not the content intended by the user is included in the translated text. If the user determines that the content of the translated text includes or is the same as the content spoken in the first language, it is determined as success and inputs a selection of success in learning for the first language through the input unit 41, and the data processor 49 Save your learning success choices. Otherwise, if the content spoken in the first language is not included in the translated text, since there is an error in the content spoken in the first language by the user, the user inputs a selection of learning failure for the first language through the input unit 41, The data processor 49 stores the learning failure selection.
- the data processor 49 when the learning failure selection is obtained from the input unit 41, the user interface of the display unit 43, an input window for inputting the contents spoken in the first language as text in the second language. Is displayed.
- the data processor 49 receives and stores the text to be translated, inputted in the input window through the input unit 41, and stores the text to be translated, the type of the language to be translated (i.e., the second language), and the type of the language to be translated (i.e. , The first language) to be transmitted to the translation server 50 through the network 60 by controlling the communication unit 45.
- the translation server 50 receives the translation target text information, refers to the type of the translation target language and the type of the translation language, and translates the included translation target text into a translation language to generate the translated text.
- the translation server 50 transmits the generated translated text to the electronic communication device 40 through the network 60.
- the data processor 49 receives and stores the translated text through the communication unit 45, and displays the translated text through the display unit 43, so that the user can check and learn the translated text.
- the translation server 50 generates translated speech information including a speech signal of the translated text by synthesizing the translated text into speech, in addition to the translation text, and transmitting it to the electronic communication device 40, and the data processor 49
- the voice signal included in the translated voice information through the communication unit 45 may be audibly expressed through the display unit 43.
- the language processing system enables the user to learn the first language.
- FIGS. 2 to 6 are cross-sectional views of an earset equipped with a voice microphone according to a first embodiment of the present invention.
- the ear sets in FIGS. 2 to 6 may be applied to the first ear set 10 and the second ear set 20, the voice microphone corresponds to the microphone 11 and the microphone 21 described above, and the sound reproduction unit is a speaker ( 13) and the speaker 23.
- the housing 100 has an installation space 110 in which parts can be installed and an insertion tube 120 structure that can be inserted into the ear canal, and the insertion tube 120 has an ear of an elastic member so that it can be gently in close contact with the user's ear.
- the bud can be detached.
- Components such as a sound reproduction unit 200 such as a microspeaker, a voice microphone 300 to which a user's voice signal is input, and a communication module 15 for controlling them may be installed in the installation space 110.
- the sound reproducing unit 200 emits sound toward the insertion tube 120 to direct sound to the user's ear, and the voice microphone 300 receives the user's voice from the insertion tube.
- the sound reproduction unit 200 and the voice microphone 300 are assembled on the upper bracket 410 to facilitate installation, and the lower bracket 430 is coupled to the upper bracket 410 to facilitate the installation. To form a closed space 400 for the volume of the bag. Meanwhile, between the upper bracket 410 and the lower bracket 430, a microphone bracket 420 is additionally installed, and a voice microphone 300 is installed in the microphone bracket 420.
- the upper bracket 410 communicates with the insertion tube 120, a first conduit 412 leading to the sound reproducing unit 200 and a second conduit 414 leading to the microphone bracket 420 are formed.
- a conduit 422 connecting the second conduit 414 and the voice microphone 300 to the microphone bracket 420 is formed.
- a terminal capable of transmitting an electrical signal to the sound reproduction unit 200 and the voice microphone 300 may be additionally provided.
- the terminal (not shown) may be connected to a communication module 15 such as a PCB or a connection cable 24.
- the assembly After the assembly of the upper bracket 410, the sound reproduction unit 200, the microphone bracket 420, the voice microphone 300, the terminal (not shown), and the lower bracket 430 is completed, the assembly is installed in the housing 100. It is inserted and fixed in the space 110.
- the upper bracket 410 has a shape corresponding to the installation space 110 of the housing 100.
- the size of the backhaul of the sound reproducing unit 200 can be increased to approximately 1.0 mm, and accordingly, the sound pressure can be increased by 6 dB or more in a low frequency band.
- FIG. 3 is an exploded view of an earset according to a second embodiment of the present invention
- FIG. 4 is a perspective view of an earset according to a second embodiment of the present invention
- FIG. 5 is a cross-sectional view of an earset according to a second embodiment of the present invention.
- the earset according to the second embodiment of the present invention applies a conduit structure for transmitting the voice transmitted through the soundproofing device, which is a technical feature of the present invention, to an open earset in which a bag 122a is formed on the rear housing 120a. .
- the earset according to the second embodiment of the present invention has a front housing 110a facing the user's ear and a rear housing 120a facing the user's ear, and the front housing 110a and the rear housing 120a are Components are installed in the installation space formed by combining.
- Components such as a sound reproducing unit 200 such as a microspeaker, a voice microphone 300 to which a user's voice signal is input, and a communication module 15 for controlling them may be installed in the housings 110a and 120a.
- the front housing 110a includes one or more soundproofing holes 112a and 114a, and the earset according to the second embodiment of the present invention has two soundproofing holes 112a and 114a formed at a predetermined angle to each other.
- the soundproofing holes 112a and 114a may be divided into a first soundproofing hole 112a having a relatively large size and a second soundproofing hole 114a having a relatively small size.
- the first soundproofing hole 112a outputs sound from the acoustic conversion device 300 to the ear canal
- the second soundproofing hole 114a is a structure for the overall balance of the SPL and plays a role of flatly tuning the sound pressure in the mid-range, and the high-frequency sound pressure Raises. It is preferable that the acoustic radiation angle of the first soundproofing hole 112a and that of the second soundproofing hole 114a be 90 degrees or more.
- the voice microphone 300 receives the user's voice from the first soundproofing hole 112a.
- a conduit 420a communicating with the first soundproofing hole 112a and transmitting the user's voice to the voice microphone 300 is provided.
- the conduit 420a is coupled to the front housing 110a. Accordingly, when the user speaks, the voice coming into the first soundproofing hole 112a through the Eustachian tube can be transmitted to the voice microphone 300.
- the rear housing 120a includes a back hole 122a through which the inside of the housing communicates with the outside so as to maintain a constant sound pressure inside the ear.
- a bracket 430a may be installed between the rear housing 120a and the acoustic conversion device 200.
- the bracket 430a covers the back hole 122 to form a pipeline.
- the bracket 430a is formed at a position spaced apart from the back hole 122a of the rear housing 120a and includes a communication hole 432a for communicating the inside of the housing and the pipeline. That is, the conduit connects the communication hole 432a and the back hole 122a.
- the pipe structure formed by the bracket 430a serves to enhance low-frequency sound by generating internal resonance within the housings 110a and 120a.
- the backhaul 122a serves to cancel Dip occurring in the 2 kHz band.
- FIG. 6 is a cross-sectional view of an earset according to a third embodiment of the present invention.
- Earsets according to the third embodiment of the present invention are all the same as those of the second embodiment, but the pipe 420b connecting the voice microphone 300 and the first soundproofing hole 112a is bent so that the first soundproofing hole 112a It is characterized by extending into ). As the pipe 420b is bent into the first soundproof hole, there is an advantage that howling can be suppressed during a voice call.
- Earset provided by the present invention can suppress external noise by forming a voice microphone channel in the soundproofing hole to input a voice from the Eustachian tube during a call with a microphone.
- howling may occur during a voice call, howling can be suppressed by separately manufacturing and installing a pipe 420b that extends into the soundproofing hole to guide the voice.
- the structure of forming the voice microphone channel in the soundproofing hole may be applied to a kernel-type earset or an open-type earset. It can also be applied to wireless earsets and TWS earsets.
- At least a part of a device eg, a processor or its functions
- a method eg, operations
- a computer-readable storage media in the form of, for example, a program module. It can be implemented as a stored command.
- the one or more processors may perform a function corresponding to the command.
- the computer-readable storage medium may be, for example, a memory.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Tourism & Hospitality (AREA)
- General Health & Medical Sciences (AREA)
- Educational Administration (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Acoustics & Sound (AREA)
- Educational Technology (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Artificial Intelligence (AREA)
- Economics (AREA)
- Human Resources & Organizations (AREA)
- Marketing (AREA)
- Primary Health Care (AREA)
- Strategic Management (AREA)
- General Business, Economics & Management (AREA)
- Machine Translation (AREA)
Abstract
La présente invention concerne un système de traitement de langue utilisant un écouteur, et en particulier un système de traitement de langue utilisant un écouteur dans lequel le système acquiert plus clairement la parole d'un utilisateur en bloquant l'entrée de bruit externe et interprète la parole acquise ou effectue une fonction de traduction et d'apprentissage pour la parole acquise. Un système de traitement de langue utilisant un écouteur selon la présente invention est conçu à partir d'un dispositif de communication électronique comprenant : un premier écouteur ou un deuxième écouteur ; un dispositif acoustique pouvant être porté sur soi qui réalise une communication sans fil avec le premier écouteur ou une communication filaire avec le deuxième écouteur ; une unité de communication qui effectue une communication sans fil avec le premier écouteur ou le dispositif acoustique pouvant être porté sur soi, et communique avec un serveur de traduction ; et une unité d'affichage. Le dispositif de communication électronique effectue une fonction de traitement de langue consistant à : générer des informations à traiter comprenant un signal vocal, qui est dans une langue source et reçu depuis le premier écouteur ou le dispositif acoustique pouvant être porté, et transmettre les informations à traiter au serveur de traduction ; recevoir des informations traitées transmises depuis le serveur de traduction et comprenant un texte de traduction qui est dans une langue cible et qui correspond aux informations à traiter, ou un signal vocal converti depuis le texte de traduction ; et afficher visuellement les informations traitées par l'intermédiaire de l'unité d'affichage, ou exprimer de manière acoustique les informations traitées.
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR10-2019-0133599 | 2019-10-25 | ||
| KR10-2019-0133598 | 2019-10-25 | ||
| KR1020190133598A KR102219494B1 (ko) | 2019-10-25 | 2019-10-25 | 이어셋 및 이를 이용한 언어 학습 시스템 |
| KR1020190133599A KR102285877B1 (ko) | 2019-10-25 | 2019-10-25 | 이어셋을 이용한 통역 시스템 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2021080362A1 true WO2021080362A1 (fr) | 2021-04-29 |
Family
ID=75619982
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/KR2020/014544 Ceased WO2021080362A1 (fr) | 2019-10-25 | 2020-10-23 | Système de traitement de langue utilisant un écouteur |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2021080362A1 (fr) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN113411995A (zh) * | 2021-05-27 | 2021-09-17 | 王强 | 识别多语种的语言翻译机 |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| KR20150081157A (ko) * | 2014-01-03 | 2015-07-13 | 엘지전자 주식회사 | 넥밴드형 단말기 |
| KR101693268B1 (ko) * | 2015-04-10 | 2017-01-05 | 해보라 주식회사 | 이어셋 |
| KR101767467B1 (ko) * | 2016-04-19 | 2017-08-11 | 주식회사 오르페오사운드웍스 | 소음 차폐 이어셋 및 이의 제조방법 |
| KR101834546B1 (ko) * | 2013-08-28 | 2018-04-13 | 한국전자통신연구원 | 핸즈프리 자동 통역 서비스를 위한 단말 장치 및 핸즈프리 장치와, 핸즈프리 자동 통역 서비스 방법 |
| JP2019175426A (ja) * | 2018-12-20 | 2019-10-10 | 株式会社フォルテ | 翻訳システム、翻訳方法、翻訳装置、及び音声入出力装置 |
-
2020
- 2020-10-23 WO PCT/KR2020/014544 patent/WO2021080362A1/fr not_active Ceased
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| KR101834546B1 (ko) * | 2013-08-28 | 2018-04-13 | 한국전자통신연구원 | 핸즈프리 자동 통역 서비스를 위한 단말 장치 및 핸즈프리 장치와, 핸즈프리 자동 통역 서비스 방법 |
| KR20150081157A (ko) * | 2014-01-03 | 2015-07-13 | 엘지전자 주식회사 | 넥밴드형 단말기 |
| KR101693268B1 (ko) * | 2015-04-10 | 2017-01-05 | 해보라 주식회사 | 이어셋 |
| KR101767467B1 (ko) * | 2016-04-19 | 2017-08-11 | 주식회사 오르페오사운드웍스 | 소음 차폐 이어셋 및 이의 제조방법 |
| JP2019175426A (ja) * | 2018-12-20 | 2019-10-10 | 株式会社フォルテ | 翻訳システム、翻訳方法、翻訳装置、及び音声入出力装置 |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN113411995A (zh) * | 2021-05-27 | 2021-09-17 | 王强 | 识别多语种的语言翻译机 |
| CN113411995B (zh) * | 2021-05-27 | 2023-05-23 | 德州学院 | 识别多语种的语言翻译机 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9484017B2 (en) | Speech translation apparatus, speech translation method, and non-transitory computer readable medium thereof | |
| TWI695281B (zh) | 翻譯系統、翻譯方法、以及翻譯裝置 | |
| JP2019175426A (ja) | 翻訳システム、翻訳方法、翻訳装置、及び音声入出力装置 | |
| JPWO2013077110A1 (ja) | 翻訳装置、翻訳システム、翻訳方法およびプログラム | |
| WO2021071271A1 (fr) | Appareil électronique et procédé de commande associé | |
| WO2019112181A1 (fr) | Dispositif électronique pour exécuter une application au moyen d'informations de phonème comprises dans des données audio, et son procédé de fonctionnement | |
| KR101619133B1 (ko) | 통역용 이어셋 | |
| WO2020080635A1 (fr) | Dispositif électronique permettant d'effectuer une reconnaissance vocale à l'aide de microphones sélectionnés d'après un état de fonctionnement, et procédé de fonctionnement associé | |
| KR101517975B1 (ko) | 동시 통/번역 기능을 가지는 이어폰 장치 | |
| US20210090548A1 (en) | Translation system | |
| WO2021118180A1 (fr) | Terminal d'utilisateur, appareil de diffusion, système de diffusion le comprenant et procédé de commande associé | |
| WO2019004762A1 (fr) | Procédé et dispositif permettant de fournir une fonction d'interprétation à l'aide d'un écouteur | |
| WO2021080362A1 (fr) | Système de traitement de langue utilisant un écouteur | |
| WO2020009261A1 (fr) | Dispositif numérique assurant une reconnaissance vocale et son procédé de commande | |
| WO2020101174A1 (fr) | Procédé et appareil pour produire un modèle de lecture sur les lèvres personnalisé | |
| JP6813176B2 (ja) | 音声抑制システム及び音声抑制装置 | |
| US20240370669A1 (en) | Speech translation processing apparatus | |
| WO2022177103A1 (fr) | Dispositif électronique de prise en charge de service pour agent à intelligence artificielle (ia) parlant avec un utilisateur | |
| JP2010128766A (ja) | 情報処理装置、情報処理方法、プログラム及び記憶媒体 | |
| CN106031135A (zh) | 可穿戴设备和通信控制方法 | |
| WO2020091482A1 (fr) | Procédé et dispositif permettant de réduire la diaphonie dans un système d'interprétation automatique | |
| KR102285877B1 (ko) | 이어셋을 이용한 통역 시스템 | |
| WO2019103340A1 (fr) | Dispositif électronique et son procédé de commande | |
| WO2022177063A1 (fr) | Dispositif électronique et son procédé de commande | |
| KR102219494B1 (ko) | 이어셋 및 이를 이용한 언어 학습 시스템 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20878291 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 20878291 Country of ref document: EP Kind code of ref document: A1 |