[go: up one dir, main page]

CN1292400C - Expression figure explanation treatment method for text and voice transfer system - Google Patents

Expression figure explanation treatment method for text and voice transfer system Download PDF

Info

Publication number
CN1292400C
CN1292400C CNB2004100781977A CN200410078197A CN1292400C CN 1292400 C CN1292400 C CN 1292400C CN B2004100781977 A CNB2004100781977 A CN B2004100781977A CN 200410078197 A CN200410078197 A CN 200410078197A CN 1292400 C CN1292400 C CN 1292400C
Authority
CN
China
Prior art keywords
expression
explanation
text
voice
character string
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CNB2004100781977A
Other languages
Chinese (zh)
Other versions
CN1655231A (en
Inventor
姜容成
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
LG Electronics China Research and Development Center Co Ltd
Original Assignee
LG Electronics China Research and Development Center Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by LG Electronics China Research and Development Center Co Ltd filed Critical LG Electronics China Research and Development Center Co Ltd
Publication of CN1655231A publication Critical patent/CN1655231A/en
Application granted granted Critical
Publication of CN1292400C publication Critical patent/CN1292400C/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A23FOODS OR FOODSTUFFS; TREATMENT THEREOF, NOT COVERED BY OTHER CLASSES
    • A23LFOODS, FOODSTUFFS OR NON-ALCOHOLIC BEVERAGES, NOT OTHERWISE PROVIDED FOR; PREPARATION OR TREATMENT THEREOF
    • A23L7/00Cereal-derived products; Malt products; Preparation or treatment thereof
    • A23L7/10Cereal-derived products
    • AHUMAN NECESSITIES
    • A23FOODS OR FOODSTUFFS; TREATMENT THEREOF, NOT COVERED BY OTHER CLASSES
    • A23PSHAPING OR WORKING OF FOODSTUFFS, NOT FULLY COVERED BY A SINGLE OTHER SUBCLASS
    • A23P30/00Shaping or working of foodstuffs characterised by the process or apparatus
    • A23P30/20Extruding

Landscapes

  • Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Chemical & Material Sciences (AREA)
  • Food Science & Technology (AREA)
  • Polymers & Plastics (AREA)
  • Manufacturing & Machinery (AREA)
  • Health & Medical Sciences (AREA)
  • Nutrition Science (AREA)
  • Machine Translation (AREA)
  • Document Processing Apparatus (AREA)

Abstract

The invention relates to technology which outputs corresponding pronunciation if expression graph has been found when a text-voice transfer system (TTS) transfers character string into voice signal. The invention is comprised by following three steps: The first step of using corresponding pronunciation to output the expression graph contained in the character string while implementing glossology processing in order to transfer the input character string in the TTS engine into voice signal; The second step of confirming voice related information such as tone, sound length and so on in order to confirm rhythm of the sentence of the voice signal, and properly adjusting rhythm according to the sensation represented by the expression graph; the third step of outputting the adjusted voice signal of the rhythm to external.

Description

The expression figure explanation treatment method of text and voice transfer system
Technical field
The present invention relates to text voice conversion (TTS:Text To Speech) system handles expression (emotion) illustrated technology, particularly relate to a kind of when tts engine is converted to voice signal with character string, if the discovery expression figure explanation is then exported the expression figure explanation treatment method of tts system of its corresponding pronunciation.
Background of invention
Tts system mainly is the system that character string is converted to human speech, and basic goal is the people not to be spent see the text that is made of character string, but goes to listen with ear.This TTS technology is than the speech recognition technology technology of closeness to life more, can be used for various text messages are converted to the service of voice.Recently, along with popularizing of Email, can utilize phone to read the mail of new reception from the outside, this also be to benefit from the TTS technology.In addition, the TTS technology can also utilize voice to listen to by the sentence of word processor input, the html document that web browser shows on screen.For visually impaired person, can be to listen to behind the voice the information translation on the internet, thereby can equally with the ordinary people obtain various Useful Informations.Recently, develop and surmounted the synthesized voice of mechanical sound level in the past, can send the technology of the synthesized voice of similar human speech, utilize the service of TTS technology just presenting the trend that enlarges gradually towards masses.
Yet the language that people use has vitality, and is constantly changing, and under as it is by the situation of diverse network with written communication thought, the speed of its variation is accelerated just day by day.
Recently, in fields such as compunication, the frequency of utilization of expression figure explanation increases just gradually.Above-mentioned " expression figure explanation " is used to show user's emotion or wish, is the synthetic language of emotion and icon (icon), makes by various symbols and literal on the integral keyboard.For example, the smiling face can be expressed as :) or:-), turn to the left side to see to be exactly a smiling face.Use the student's Scott Fil graceful (Scott Fahlman) that allegedly is the 1980's Ka Naiji-Mei Long university at first.It can be artificially guided into the stiff inflexible compunication that is easy to become soft and is full of the boundary of enjoyment, makes the hommization more that becomes of communicating by letter between machine and the machine.
Yet, in the tts system under the prior art, can only be voice with common text conversion, expression figure explanation is processed into simple sentence symbol or totally uninteresting symbol, have difficulties aspect the document content fully transmitting to the user.
Summary of the invention
Therefore, the present invention seeks to overcome deficiency of the prior art, a kind of expression figure explanation treatment method of tts system is provided.
For achieving the above object, the expression figure explanation treatment method of tts system of the present invention is made of following three steps: the 1st step, for the character string that will import in the tts engine is converted to voice signal, carrying out the sentence of text handles, non-Chinese text is handled, operations such as analysis of part of speech such as verb, adjective and syntactic analysis and the conversion of pronunciation mark are exported the expression figure explanation that comprises in this character string with corresponding pronunciation simultaneously; The 2nd step is determined acoustic informations such as tone, the duration of a sound for the rhythm of the sentence of determining to be converted to described voice signal, and the emotion of representing according to described expression figure explanation is suitably regulated the rhythm then; The 3rd step after speech database generation actual speech signal, is carried out the D/A conversion and is carried out processing and amplifying this voice signal.
As the above-mentioned detailed description of doing, the present invention has following effect, promptly, when tts engine is converted to voice signal with character string, if the discovery expression figure explanation then utilizes the output of expression figure explanation pronunciation dictionary and this corresponding sound of expressing one's feelings, thereby, when the content of the literal that will comprise expression figure explanation is converted to corresponding sound and exports, can directly transmit its content.
Description of drawings
Fig. 1 is system's pie graph of expression figure explanation treatment method of the present invention;
Fig. 2 is the pronunciation sample table of the expression figure explanation pronunciation dictionary among Fig. 1.
The symbol description of accompanying drawing major part
1: text input part 2: the linguistics handling part
3: rhythm handling part 4: the voice signal handling part
5: voice signal efferent 6: dictionary portion
7: expression figure explanation pronunciation dictionary 8: speech database
Embodiment
Below with reference to Fig. 1 and Fig. 2, describe expression figure explanation treatment step of the present invention in detail.
Character string is behind the text input part 1 of external device (ED) or internal storage input tts engine, linguistics handling part 2 is in order to be converted into voice signal, with reference to the various data in the numeral/abbreviation symbol dictionary of dictionary portion 6, part of speech dictionary, the pronunciation dictionary, operations such as the sentence processing of execution contexts, syntactic analysis, non-Chinese text processing, morpheme analysis and syntactic analysis, the conversion of pronunciation mark.
At this moment, above-mentioned linguistics handling part 2 is not that it is processed into simple symbol, but exports with the pronunciation of income in the expression figure explanation pronunciation dictionary 7 after utilizing the expression figure explanation that comprises in the above-mentioned character string of expression figure explanation pronunciation dictionary 7 identifications.
As a reference, Fig. 2 has shown the pronunciation example of each expression figure explanation of record in the above-mentioned expression figure explanation pronunciation dictionary 7.For instance, ^^, ^-^ :), the pronunciation of expression figure explanations such as ^o^, ^_^ is " having laughed at ".Again for example, for-.-,--,-expression figure explanation of .-, pronounce to be " amimia ".
Then, rhythm handling part 3 is determined the length of tone, sound etc. in order to determine the rhythm with the voice output sentence time.At this moment, the emotion of representing according to above-mentioned expression figure explanation is suitably regulated the rhythm.
Then, voice signal handling part 4 generates actual speech signal with reference to the actual speech data in voice data storehouse 8, the actual speech signal of 5 pairs of above-mentioned generations of voice signal efferent is carried out D/A conversion, exports after being enlarged into suitable level simultaneously, so that the people can hear.

Claims (5)

1. the expression figure explanation treatment method of a text and voice transfer system is characterized in that being made of following steps:
The 1st step when carrying out the linguistics processing, is exported the expression figure explanation that comprises in this character string with corresponding pronunciation, is converted to voice signal in order to the character string that will import in the text voice transform engine;
The 2nd step is determined the information relevant with sound for the rhythm of the sentence of determining to be converted to described voice signal, and the emotion of representing according to described expression figure explanation is suitably regulated the rhythm then;
The 3rd step outputs to the outside with the voice signal of the described rhythm regulated.
2. the expression figure explanation treatment method of text and voice transfer system according to claim 1 is characterized in that the character string of input text speech conversion engine is provided by external device (ED) or internal storage.
3. the expression figure explanation treatment method of text and voice transfer system according to claim 1 is characterized in that described the 1st step also comprises:
After utilizing the expression figure explanation pronunciation dictionary to discern the expression figure explanation that comprises in the described character string, the step of exporting with the pronunciation of taking in the expression figure explanation pronunciation dictionary.
4. the expression figure explanation treatment method of text and voice transfer system according to claim 1 is characterized in that the described information relevant with sound comprises the tone and the duration of a sound.
5. the expression figure explanation treatment method of text and voice transfer system according to claim 3 is characterized in that, described expression figure explanation pronunciation dictionary is being stored the pronunciation corresponding with each expression figure explanation.
CNB2004100781977A 2004-02-10 2004-09-17 Expression figure explanation treatment method for text and voice transfer system Expired - Fee Related CN1292400C (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
KR1020040008771 2004-02-10
KR1020040008771A KR20050080671A (en) 2004-02-10 2004-02-10 Emoticon processing method for text to speech system
KR10-2004-0008771 2004-02-10

Publications (2)

Publication Number Publication Date
CN1655231A CN1655231A (en) 2005-08-17
CN1292400C true CN1292400C (en) 2006-12-27

Family

ID=34909935

Family Applications (1)

Application Number Title Priority Date Filing Date
CNB2004100781977A Expired - Fee Related CN1292400C (en) 2004-02-10 2004-09-17 Expression figure explanation treatment method for text and voice transfer system

Country Status (2)

Country Link
KR (1) KR20050080671A (en)
CN (1) CN1292400C (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101217685A (en) * 2007-01-04 2008-07-09 中兴通讯股份有限公司 Method and device for sending MMS via text message
CN101072258B (en) * 2007-03-29 2012-06-06 腾讯科技(深圳)有限公司 Voice on demand method and device
CN104053131A (en) * 2013-03-12 2014-09-17 华为技术有限公司 Text communication information processing method and related equipment
CN103761963A (en) * 2014-02-18 2014-04-30 大陆汽车投资(上海)有限公司 Method for processing text containing emotion information
CN104699675B (en) * 2015-03-18 2018-01-30 北京交通大学 The method and apparatus of translation information
CN105139848B (en) * 2015-07-23 2019-01-04 小米科技有限责任公司 Data transfer device and device
CN106951105A (en) * 2017-03-03 2017-07-14 深圳市联谛信息无障碍有限责任公司 A kind of method that use Barrier-free Service reads emoticon
CN110189742B (en) * 2019-05-30 2021-10-08 芋头科技(杭州)有限公司 Method and related device for determining emotion audio frequency, emotion display and text-to-speech
CN118541750A (en) 2022-01-11 2024-08-23 三星电子株式会社 Method and electronic device for intelligently reading aloud displayed content
WO2023136605A1 (en) * 2022-01-11 2023-07-20 Samsung Electronics Co., Ltd. Method and electronic device for intelligently reading displayed contents

Also Published As

Publication number Publication date
KR20050080671A (en) 2005-08-17
CN1655231A (en) 2005-08-17

Similar Documents

Publication Publication Date Title
US7062437B2 (en) Audio renderings for expressing non-audio nuances
CN108962219B (en) method and device for processing text
CN111223498A (en) Intelligent emotion recognition method and device and computer readable storage medium
Wilpon et al. Voice communication between humans and machines
US9240180B2 (en) System and method for low-latency web-based text-to-speech without plugins
US20060069567A1 (en) Methods, systems, and products for translating text to speech
US20100217591A1 (en) Vowel recognition system and method in speech to text applictions
WO2016159961A1 (en) Voice driven operating system for interfacing with electronic devices
JP2004355629A (en) Semantic object synchronous understanding for highly interactive interface
GB2307619A (en) Internet information access system
CN1292400C (en) Expression figure explanation treatment method for text and voice transfer system
EP1371057A1 (en) Method for enabling the voice interaction with a web page
EP1685556A1 (en) Audio dialogue system and voice browsing method
CN118136047A (en) A speech sentiment analysis method based on semantic intonation
CN104050962B (en) Multifunctional Reader Based on Speech Synthesis Technology
CN115249480A (en) Method and related device for converting speech and text based on Beidou short message
Trivedi Fundamentals of Natural Language Processing
JP6289950B2 (en) Reading apparatus, reading method and program
Kehoe et al. Designing help topics for use with text-to-speech
CN118312040A (en) Virtual digital person interaction method, device, equipment and storage medium
JP2003044072A (en) Voice reading setting device, voice reading device, voice reading setting method, voice reading setting program, and recording medium
Spiliotopoulos et al. Acoustic rendering of data tables using earcons and prosody for document accessibility
Sasmal et al. Acoustic and spectral analysis of adi triphthongs
Shanavas Malayalam Text-to-Speech Conversion: An Assistive Tool for Visually Impaired People.
Jeevitha et al. A study on innovative trends in multimedia library using speech enabled softwares

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
C17 Cessation of patent right
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20061227

Termination date: 20091019