CN1292400C - Expression figure explanation treatment method for text and voice transfer system - Google Patents
Expression figure explanation treatment method for text and voice transfer system Download PDFInfo
- Publication number
- CN1292400C CN1292400C CNB2004100781977A CN200410078197A CN1292400C CN 1292400 C CN1292400 C CN 1292400C CN B2004100781977 A CNB2004100781977 A CN B2004100781977A CN 200410078197 A CN200410078197 A CN 200410078197A CN 1292400 C CN1292400 C CN 1292400C
- Authority
- CN
- China
- Prior art keywords
- expression
- explanation
- text
- voice
- character string
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000012546 transfer Methods 0.000 title claims abstract description 10
- 238000000034 method Methods 0.000 title claims description 11
- 230000033764 rhythmic process Effects 0.000 claims abstract description 12
- 238000012545 processing Methods 0.000 claims abstract description 5
- 238000006243 chemical reaction Methods 0.000 claims description 7
- 230000008451 emotion Effects 0.000 claims description 6
- 230000001105 regulatory effect Effects 0.000 claims description 4
- 230000008676 import Effects 0.000 claims description 2
- 238000005516 engineering process Methods 0.000 abstract description 9
- 230000035807 sensation Effects 0.000 abstract 1
- 206010053238 Amimia Diseases 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000001771 impaired effect Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A23—FOODS OR FOODSTUFFS; TREATMENT THEREOF, NOT COVERED BY OTHER CLASSES
- A23L—FOODS, FOODSTUFFS OR NON-ALCOHOLIC BEVERAGES, NOT OTHERWISE PROVIDED FOR; PREPARATION OR TREATMENT THEREOF
- A23L7/00—Cereal-derived products; Malt products; Preparation or treatment thereof
- A23L7/10—Cereal-derived products
-
- A—HUMAN NECESSITIES
- A23—FOODS OR FOODSTUFFS; TREATMENT THEREOF, NOT COVERED BY OTHER CLASSES
- A23P—SHAPING OR WORKING OF FOODSTUFFS, NOT FULLY COVERED BY A SINGLE OTHER SUBCLASS
- A23P30/00—Shaping or working of foodstuffs characterised by the process or apparatus
- A23P30/20—Extruding
Landscapes
- Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Chemical & Material Sciences (AREA)
- Food Science & Technology (AREA)
- Polymers & Plastics (AREA)
- Manufacturing & Machinery (AREA)
- Health & Medical Sciences (AREA)
- Nutrition Science (AREA)
- Machine Translation (AREA)
- Document Processing Apparatus (AREA)
Abstract
The invention relates to technology which outputs corresponding pronunciation if expression graph has been found when a text-voice transfer system (TTS) transfers character string into voice signal. The invention is comprised by following three steps: The first step of using corresponding pronunciation to output the expression graph contained in the character string while implementing glossology processing in order to transfer the input character string in the TTS engine into voice signal; The second step of confirming voice related information such as tone, sound length and so on in order to confirm rhythm of the sentence of the voice signal, and properly adjusting rhythm according to the sensation represented by the expression graph; the third step of outputting the adjusted voice signal of the rhythm to external.
Description
Technical field
The present invention relates to text voice conversion (TTS:Text To Speech) system handles expression (emotion) illustrated technology, particularly relate to a kind of when tts engine is converted to voice signal with character string, if the discovery expression figure explanation is then exported the expression figure explanation treatment method of tts system of its corresponding pronunciation.
Background of invention
Tts system mainly is the system that character string is converted to human speech, and basic goal is the people not to be spent see the text that is made of character string, but goes to listen with ear.This TTS technology is than the speech recognition technology technology of closeness to life more, can be used for various text messages are converted to the service of voice.Recently, along with popularizing of Email, can utilize phone to read the mail of new reception from the outside, this also be to benefit from the TTS technology.In addition, the TTS technology can also utilize voice to listen to by the sentence of word processor input, the html document that web browser shows on screen.For visually impaired person, can be to listen to behind the voice the information translation on the internet, thereby can equally with the ordinary people obtain various Useful Informations.Recently, develop and surmounted the synthesized voice of mechanical sound level in the past, can send the technology of the synthesized voice of similar human speech, utilize the service of TTS technology just presenting the trend that enlarges gradually towards masses.
Yet the language that people use has vitality, and is constantly changing, and under as it is by the situation of diverse network with written communication thought, the speed of its variation is accelerated just day by day.
Recently, in fields such as compunication, the frequency of utilization of expression figure explanation increases just gradually.Above-mentioned " expression figure explanation " is used to show user's emotion or wish, is the synthetic language of emotion and icon (icon), makes by various symbols and literal on the integral keyboard.For example, the smiling face can be expressed as :) or:-), turn to the left side to see to be exactly a smiling face.Use the student's Scott Fil graceful (Scott Fahlman) that allegedly is the 1980's Ka Naiji-Mei Long university at first.It can be artificially guided into the stiff inflexible compunication that is easy to become soft and is full of the boundary of enjoyment, makes the hommization more that becomes of communicating by letter between machine and the machine.
Yet, in the tts system under the prior art, can only be voice with common text conversion, expression figure explanation is processed into simple sentence symbol or totally uninteresting symbol, have difficulties aspect the document content fully transmitting to the user.
Summary of the invention
Therefore, the present invention seeks to overcome deficiency of the prior art, a kind of expression figure explanation treatment method of tts system is provided.
For achieving the above object, the expression figure explanation treatment method of tts system of the present invention is made of following three steps: the 1st step, for the character string that will import in the tts engine is converted to voice signal, carrying out the sentence of text handles, non-Chinese text is handled, operations such as analysis of part of speech such as verb, adjective and syntactic analysis and the conversion of pronunciation mark are exported the expression figure explanation that comprises in this character string with corresponding pronunciation simultaneously; The 2nd step is determined acoustic informations such as tone, the duration of a sound for the rhythm of the sentence of determining to be converted to described voice signal, and the emotion of representing according to described expression figure explanation is suitably regulated the rhythm then; The 3rd step after speech database generation actual speech signal, is carried out the D/A conversion and is carried out processing and amplifying this voice signal.
As the above-mentioned detailed description of doing, the present invention has following effect, promptly, when tts engine is converted to voice signal with character string, if the discovery expression figure explanation then utilizes the output of expression figure explanation pronunciation dictionary and this corresponding sound of expressing one's feelings, thereby, when the content of the literal that will comprise expression figure explanation is converted to corresponding sound and exports, can directly transmit its content.
Description of drawings
Fig. 1 is system's pie graph of expression figure explanation treatment method of the present invention;
Fig. 2 is the pronunciation sample table of the expression figure explanation pronunciation dictionary among Fig. 1.
The symbol description of accompanying drawing major part
1: text input part 2: the linguistics handling part
3: rhythm handling part 4: the voice signal handling part
5: voice signal efferent 6: dictionary portion
7: expression figure explanation pronunciation dictionary 8: speech database
Embodiment
Below with reference to Fig. 1 and Fig. 2, describe expression figure explanation treatment step of the present invention in detail.
Character string is behind the text input part 1 of external device (ED) or internal storage input tts engine, linguistics handling part 2 is in order to be converted into voice signal, with reference to the various data in the numeral/abbreviation symbol dictionary of dictionary portion 6, part of speech dictionary, the pronunciation dictionary, operations such as the sentence processing of execution contexts, syntactic analysis, non-Chinese text processing, morpheme analysis and syntactic analysis, the conversion of pronunciation mark.
At this moment, above-mentioned linguistics handling part 2 is not that it is processed into simple symbol, but exports with the pronunciation of income in the expression figure explanation pronunciation dictionary 7 after utilizing the expression figure explanation that comprises in the above-mentioned character string of expression figure explanation pronunciation dictionary 7 identifications.
As a reference, Fig. 2 has shown the pronunciation example of each expression figure explanation of record in the above-mentioned expression figure explanation pronunciation dictionary 7.For instance, ^^, ^-^ :), the pronunciation of expression figure explanations such as ^o^, ^_^ is " having laughed at ".Again for example, for-.-,--,-expression figure explanation of .-, pronounce to be " amimia ".
Then, rhythm handling part 3 is determined the length of tone, sound etc. in order to determine the rhythm with the voice output sentence time.At this moment, the emotion of representing according to above-mentioned expression figure explanation is suitably regulated the rhythm.
Then, voice signal handling part 4 generates actual speech signal with reference to the actual speech data in voice data storehouse 8, the actual speech signal of 5 pairs of above-mentioned generations of voice signal efferent is carried out D/A conversion, exports after being enlarged into suitable level simultaneously, so that the people can hear.
Claims (5)
1. the expression figure explanation treatment method of a text and voice transfer system is characterized in that being made of following steps:
The 1st step when carrying out the linguistics processing, is exported the expression figure explanation that comprises in this character string with corresponding pronunciation, is converted to voice signal in order to the character string that will import in the text voice transform engine;
The 2nd step is determined the information relevant with sound for the rhythm of the sentence of determining to be converted to described voice signal, and the emotion of representing according to described expression figure explanation is suitably regulated the rhythm then;
The 3rd step outputs to the outside with the voice signal of the described rhythm regulated.
2. the expression figure explanation treatment method of text and voice transfer system according to claim 1 is characterized in that the character string of input text speech conversion engine is provided by external device (ED) or internal storage.
3. the expression figure explanation treatment method of text and voice transfer system according to claim 1 is characterized in that described the 1st step also comprises:
After utilizing the expression figure explanation pronunciation dictionary to discern the expression figure explanation that comprises in the described character string, the step of exporting with the pronunciation of taking in the expression figure explanation pronunciation dictionary.
4. the expression figure explanation treatment method of text and voice transfer system according to claim 1 is characterized in that the described information relevant with sound comprises the tone and the duration of a sound.
5. the expression figure explanation treatment method of text and voice transfer system according to claim 3 is characterized in that, described expression figure explanation pronunciation dictionary is being stored the pronunciation corresponding with each expression figure explanation.
Applications Claiming Priority (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR1020040008771 | 2004-02-10 | ||
| KR1020040008771A KR20050080671A (en) | 2004-02-10 | 2004-02-10 | Emoticon processing method for text to speech system |
| KR10-2004-0008771 | 2004-02-10 |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN1655231A CN1655231A (en) | 2005-08-17 |
| CN1292400C true CN1292400C (en) | 2006-12-27 |
Family
ID=34909935
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CNB2004100781977A Expired - Fee Related CN1292400C (en) | 2004-02-10 | 2004-09-17 | Expression figure explanation treatment method for text and voice transfer system |
Country Status (2)
| Country | Link |
|---|---|
| KR (1) | KR20050080671A (en) |
| CN (1) | CN1292400C (en) |
Families Citing this family (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101217685A (en) * | 2007-01-04 | 2008-07-09 | 中兴通讯股份有限公司 | Method and device for sending MMS via text message |
| CN101072258B (en) * | 2007-03-29 | 2012-06-06 | 腾讯科技(深圳)有限公司 | Voice on demand method and device |
| CN104053131A (en) * | 2013-03-12 | 2014-09-17 | 华为技术有限公司 | Text communication information processing method and related equipment |
| CN103761963A (en) * | 2014-02-18 | 2014-04-30 | 大陆汽车投资(上海)有限公司 | Method for processing text containing emotion information |
| CN104699675B (en) * | 2015-03-18 | 2018-01-30 | 北京交通大学 | The method and apparatus of translation information |
| CN105139848B (en) * | 2015-07-23 | 2019-01-04 | 小米科技有限责任公司 | Data transfer device and device |
| CN106951105A (en) * | 2017-03-03 | 2017-07-14 | 深圳市联谛信息无障碍有限责任公司 | A kind of method that use Barrier-free Service reads emoticon |
| CN110189742B (en) * | 2019-05-30 | 2021-10-08 | 芋头科技(杭州)有限公司 | Method and related device for determining emotion audio frequency, emotion display and text-to-speech |
| CN118541750A (en) | 2022-01-11 | 2024-08-23 | 三星电子株式会社 | Method and electronic device for intelligently reading aloud displayed content |
| WO2023136605A1 (en) * | 2022-01-11 | 2023-07-20 | Samsung Electronics Co., Ltd. | Method and electronic device for intelligently reading displayed contents |
-
2004
- 2004-02-10 KR KR1020040008771A patent/KR20050080671A/en not_active Ceased
- 2004-09-17 CN CNB2004100781977A patent/CN1292400C/en not_active Expired - Fee Related
Also Published As
| Publication number | Publication date |
|---|---|
| KR20050080671A (en) | 2005-08-17 |
| CN1655231A (en) | 2005-08-17 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US7062437B2 (en) | Audio renderings for expressing non-audio nuances | |
| CN108962219B (en) | method and device for processing text | |
| CN111223498A (en) | Intelligent emotion recognition method and device and computer readable storage medium | |
| Wilpon et al. | Voice communication between humans and machines | |
| US9240180B2 (en) | System and method for low-latency web-based text-to-speech without plugins | |
| US20060069567A1 (en) | Methods, systems, and products for translating text to speech | |
| US20100217591A1 (en) | Vowel recognition system and method in speech to text applictions | |
| WO2016159961A1 (en) | Voice driven operating system for interfacing with electronic devices | |
| JP2004355629A (en) | Semantic object synchronous understanding for highly interactive interface | |
| GB2307619A (en) | Internet information access system | |
| CN1292400C (en) | Expression figure explanation treatment method for text and voice transfer system | |
| EP1371057A1 (en) | Method for enabling the voice interaction with a web page | |
| EP1685556A1 (en) | Audio dialogue system and voice browsing method | |
| CN118136047A (en) | A speech sentiment analysis method based on semantic intonation | |
| CN104050962B (en) | Multifunctional Reader Based on Speech Synthesis Technology | |
| CN115249480A (en) | Method and related device for converting speech and text based on Beidou short message | |
| Trivedi | Fundamentals of Natural Language Processing | |
| JP6289950B2 (en) | Reading apparatus, reading method and program | |
| Kehoe et al. | Designing help topics for use with text-to-speech | |
| CN118312040A (en) | Virtual digital person interaction method, device, equipment and storage medium | |
| JP2003044072A (en) | Voice reading setting device, voice reading device, voice reading setting method, voice reading setting program, and recording medium | |
| Spiliotopoulos et al. | Acoustic rendering of data tables using earcons and prosody for document accessibility | |
| Sasmal et al. | Acoustic and spectral analysis of adi triphthongs | |
| Shanavas | Malayalam Text-to-Speech Conversion: An Assistive Tool for Visually Impaired People. | |
| Jeevitha et al. | A study on innovative trends in multimedia library using speech enabled softwares |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| C06 | Publication | ||
| PB01 | Publication | ||
| C10 | Entry into substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| C14 | Grant of patent or utility model | ||
| GR01 | Patent grant | ||
| C17 | Cessation of patent right | ||
| CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20061227 Termination date: 20091019 |