[go: up one dir, main page]

WO2018094952A1 - Procédé et appareil de recommandation de contenu - Google Patents

Procédé et appareil de recommandation de contenu Download PDF

Info

Publication number
WO2018094952A1
WO2018094952A1 PCT/CN2017/079624 CN2017079624W WO2018094952A1 WO 2018094952 A1 WO2018094952 A1 WO 2018094952A1 CN 2017079624 W CN2017079624 W CN 2017079624W WO 2018094952 A1 WO2018094952 A1 WO 2018094952A1
Authority
WO
WIPO (PCT)
Prior art keywords
voice
user
information
feature information
recommendation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2017/079624
Other languages
English (en)
Chinese (zh)
Inventor
崔宝宏
王方舟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Baidu Online Network Technology Beijing Co Ltd
Original Assignee
Baidu Online Network Technology Beijing Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Baidu Online Network Technology Beijing Co Ltd filed Critical Baidu Online Network Technology Beijing Co Ltd
Publication of WO2018094952A1 publication Critical patent/WO2018094952A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/735Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/635Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/54Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for retrieval

Definitions

  • the present invention relates to the field of voice information search technology, and in particular, to a content recommendation technology.
  • voice search schemes are based on the recognition of the user's voice content. After identifying the user's voice content, the search engine provides the user with various content corresponding to the voice content. Therefore, these voice search schemes are still traditional content search in their essence, but the user's input mode has changed, instead of text input, voice input can be used.
  • a content recommendation method comprising the following steps:
  • a content recommendation apparatus comprising:
  • the present invention can recommend matching voice materials, such as songs, video clips, etc., according to the user's voice feature information.
  • the present invention uses a fun operation method to increase the user's willingness to use the voice automatically in the gradual outbreak of the voice interaction application scenario, which is beneficial to cultivating the habit of the user using the voice to search. For example, based on the social and entertainment needs of the current voice audience, when the user searches for music or uses voice to automatically analyze various voice feature information of the user, such as sound color, audio, etc., the present invention appropriately recommends to the user to sing through voice big data analysis. Songs or singers suitable for imitation.
  • FIG. 1 shows a flow chart of a method for performing content recommendation in accordance with one embodiment of the present invention
  • FIG. 2 shows a schematic diagram of an apparatus for performing content recommendation in accordance with one embodiment of the present invention.
  • Computer equipment also known as “computer” means that it can be transported
  • An intelligent electronic device that schedules a program or instruction to perform a predetermined process such as numerical calculations and/or logic calculations, which may include a processor and a memory, the processor executing program instructions pre-stored in the memory to perform a predetermined process, or The predetermined processing is performed by hardware such as an ASIC, an FPGA, or a DSP, or a combination of the two.
  • Computer devices include, but are not limited to, servers, personal computers (PCs), notebook computers, tablets, smart phones, and the like.
  • the computer device includes, for example, a user device and a network device.
  • the user equipment includes, but is not limited to, a personal computer (PC), a notebook computer, a mobile terminal, etc., and the mobile terminal includes, but is not limited to, a smart phone, a PDA, etc.;
  • the network device includes but is not limited to a single network server, and more A network of server servers or a cloud-based cloud consisting of a large number of computers or network servers, where cloud computing is a type of distributed computing, a super-virtual set of a loosely coupled set of computers. computer.
  • the computer device can be operated separately to implement the present invention, and can also access the network and implement the present invention by interacting with other computer devices in the network.
  • the network in which the computer device is located includes, but is not limited to, the Internet, a wide area network, a metropolitan area network, a local area network, a VPN network, and the like.
  • the user equipment, the network equipment, the network, and the like are merely examples, and other existing or future possible computer equipment or networks, such as those applicable to the present invention, are also included in the scope of the present invention. It is included here by reference.
  • the invention can be implemented by a computer device.
  • the present invention can be implemented by a network device, but those skilled in the art will appreciate that the solution of the present invention can be implemented by a user device as long as it has the computing/processing capabilities required by the present invention.
  • the following description of the implementation of the network device is exemplified in the present specification, but those skilled in the art should understand that the examples are only for explaining the purpose of the present invention, and should not be construed as limiting the present invention.
  • FIG. 1 shows a method flow diagram in accordance with one embodiment of the present invention, in which a content recommendation process is specifically illustrated.
  • step S1 the network device receives the voice information submitted by the user; in step S2, the network device extracts the voice feature information of the user according to the voice information; in step S3, the network device is configured according to The sound feature information queries a preset first voice database to obtain a matched first voice data; in step S4, the network device recommends the first voice data to the user.
  • the present invention is typically applicable to search scenes or music interaction class scenarios.
  • the first voice material recommended to the user may be any voice material that is matched by the voice feature information, including but not limited to songs and video clips.
  • voice data is not intended to limit its expression, such as various audio materials, but may include any material having audio content, so that it can include not only songs but also songs.
  • a video clip can be included.
  • the network device recommends a first voice material according to the voice feature information of the user, for example, a video segment suitable for the user to imitate.
  • the network device also needs to determine whether the user has entertainment requirements, such as identifying the voice content submitted by the user, to analyze whether the content delivered by the voice has video, music, and other related entertainment needs, for example, by using a preset keyword table.
  • Entertainment needs, such as whether the voice content is a lyric, a line of words, and the like.
  • the network device can recommend the first voice material according to the voice feature information of the user, for example, a song suitable for the user to sing.
  • the content recommendation scheme of the present invention may be embodied as a “matching” function, which may provide the user with the first voice material that matches the voice feature information.
  • the client may present a "match" function button to the user, such as placed on the right side of the voice search box or at a specific location in the interface of the music class APP.
  • a "match" function button to the user, such as placed on the right side of the voice search box or at a specific location in the interface of the music class APP.
  • the "match" function can be directly prompted to the user to remind the user that voice information can be recommended for the previously submitted voice information.
  • match function button is only used to illustrate the purpose of the present invention, and its expression in the actual application may be different, for example, it may be presented as a "main song” function button, "large The “star” function button or the like, as long as the function buttons are intended to provide the user with a content recommendation scheme as in the present invention, it should fall within the scope of the patent protection of the present invention.
  • network device in the present invention is not limited to one computer device, and may be a plurality of computer devices, typically such as multiple servers, which cooperate with each other to implement the present invention.
  • Content recommendation program is not limited to one computer device, and may be a plurality of computer devices, typically such as multiple servers, which cooperate with each other to implement the present invention.
  • the interface server is responsible for interacting with the client and other servers, wherein the interaction with the client is such as receiving voice information sent by the client and returning the recommended first voice data to the client, and interacting with the matching server, such as submitting voice information. And matching the server and obtaining the matched first voice data from the matching server; the matching server is responsible for performing matching query on the voice information submitted by the interface server in the first voice database to obtain the matched first voice data.
  • These servers are considered as a whole relative to other external devices and are generally considered as "network devices" in their entirety.
  • step S1 the network device receives the voice information submitted by the user.
  • the present invention requires the network device to interact with the client to implement.
  • the user end is, for example, various functional entities running in the user equipment, such as a webpage in a PC, an APP in a mobile phone, and the like.
  • the user submits a piece of voice information through the user equipment, and the user equipment sends the voice information to the network device, so that the network device receives the voice information submitted by the user.
  • step S2 the network device extracts the voice feature information of the user according to the voice information submitted by the user.
  • the user's voice feature information can be defined from at least the following two dimensions:
  • the perceived dimension is intended to be defined from the perceived characteristics of the sound.
  • the sound feature information is specifically in this dimension such as pitch feature information, pitch feature information, melody feature information, rhythm feature information, audio feature information, speech rate feature information, and the like.
  • the acoustic dimension is intended to be expressed from various features defined by the acoustic angle of the sound.
  • the sound feature information is specifically such as energy feature information, zero-crossing rate, LPC (Linear Prediction Coefficient) parameter information, and the like in this dimension.
  • the zero-crossing rate means the ratio of the sign change of the sound signal, such as the sound signal changing from a positive number to a negative number or a reverse direction.
  • the network device may extract the voice feature information by using any known or future feasible technology, which is not limited by the present invention.
  • step S3 the network device queries the preset first voice data database according to the user's voice feature information to obtain the matched first voice data.
  • the network device may pre-establish a first voice database in which various sound feature information of each voice material is stored.
  • the first voice database may also be established and maintained by other devices, and the network device may have access rights thereto.
  • a song or a participating video segment of each star is stored in the first voice database, and the voice materials are labeled with various sound feature information.
  • the network device performs a matching query in the first voice database according to the sound feature information recognized by the user to obtain a matched song or video segment.
  • step S4 the network device recommends the matched first voice material to the user.
  • the network device recommends the matched songs or video clips and the like to the user, so that the present invention can provide the user with a song or video dubbing suitable for singing or imitating, so as to customize the "title song” or "famous song” for the user. And help users find their singer that is suitable for imitation. This is practical in commercial applications and significantly increases user stickiness and length of use.
  • the invention automatically matches singers, poets, celebrities, etc. by guiding the user through various voice inputs such as speaking, shouting, singing two sentences, reading two poems and the like. Further, the present invention can also Continue to receive the user's singing of the recommended songs, and then score the user's own singing relative to the original singing, and by sharing the friends, inviting friends to PK, etc. to enhance the fun to help users find the songs that are most suitable for singing, create users I imitate the famous songs, and gradually guide the user to form a voice search habit and a second interaction with the search engine to provide user stickiness and duration of use. Accordingly, the present invention can also be derived from a search engine to a social field by providing a sharing function.
  • the content recommendation process shown in FIG. 1 further includes a step S5.
  • step S5 the network device queries the preset second voice data database according to the content of the user voice information to obtain the matched second voice data; and then the network device recommends the matched second voice data to the user.
  • step of recommending the second voice material by the network device may be merged with the recommendation of step S4, so that step S5 occurs before step S4, and the network device may recommend the first voice data and the second to the user in step S4. Voice data.
  • the recommendation of the network device for the second voice material may also be independent of the recommendation of the first voice material in step S4.
  • the acquisition of the second voice material in step S5 and the acquisition of the second voice material in step S6 may be performed before, after or simultaneously with the two steps. recommend.
  • the network device After identifying the content in the voice information of the user, the network device performs a matching query in the second voice database to obtain second voice data with the same content, such as voice data with the same content but different sounds. If you use the same content without dialect.
  • the voice information submitted by the user is "Hello” in Mandarin
  • the network device recognizes that the content information is “Hello” and performs a matching query in the second voice database to obtain “Hello” in various dialects. For example, “Hello” in Shanghai dialect and Cantonese.
  • the first voice database can be integrated with the second voice library. That is, in the present invention, only one voice database is used to perform matching query of the first voice data and the second voice data, so that at least the voice feature information and the content information of each voice material are stored in the voice database.
  • the content recommendation device 20 is installed in the network device 200, and specifically includes a receiving device 21, an extracting device 22, a matching device 23, and a recommending device 24.
  • the receiving device 21 receives the voice information submitted by the user; the extracting device 22 extracts the sound feature information of the user according to the voice information; the matching device 23 queries the preset first voice database according to the sound feature information. A matching first voice material is obtained; the recommendation device 24 recommends the first voice material to the user.
  • the present invention is typically applicable to search scenes or music interaction class scenarios.
  • the first voice material recommended to the user may be any voice material that is matched by the voice feature information, including but not limited to songs and video clips.
  • voice data is not intended to limit its expression, such as various audio materials, but may include any material having audio content, so that it can include not only songs but also songs.
  • a video clip can be included.
  • the content recommendation device 20 recommends a first voice material for the user according to the voice feature information of the user, for example, a video segment suitable for the user to imitate.
  • the content recommendation device 20 or other devices in the network device also need to determine whether the user has entertainment requirements, such as identifying the voice content submitted by the user, to analyze whether the content delivered by the voice has video, music, and other related entertainment needs, for example, Identify the entertainment needs through a preset keyword list, such as whether the voice content is a lyric, a line, etc.
  • the user currently uses the music class APP.
  • the content recommendation device 20 may recommend the first voice material according to the voice feature information of the user, for example, a song suitable for the user to sing. .
  • the content recommendation scheme of the present invention may be embodied as a “matching” function, which may provide the user with the first voice material that matches the voice feature information.
  • the client may present a "match" function button to the user, such as placed on the right side of the voice search box or at a specific location in the interface of the music class APP.
  • the user sends the voice information submitted by the user to the network device, so that the content recommendation device 20 recommends the corresponding first according to the user's voice feature information.
  • Voice data such as placed on the right side of the voice search box or at a specific location in the interface of the music class APP.
  • the "match" function can be directly prompted to the user to remind the user that voice information can be recommended for the previously submitted voice information.
  • match function button is only used to illustrate the purpose of the present invention, and its expression in the actual application may be different, for example, it may be presented as a "main song” function button, "large The “star” function button or the like, as long as the function buttons are intended to provide the user with a content recommendation scheme as in the present invention, it should fall within the scope of the patent protection of the present invention.
  • network device in the present invention is not limited to one computer device, and may be a plurality of computer devices, typically such as multiple servers, which cooperate with each other to implement the present invention.
  • Content recommendation program is not limited to one computer device, and may be a plurality of computer devices, typically such as multiple servers, which cooperate with each other to implement the present invention.
  • the interface server is responsible for interacting with the client and other servers, wherein the interaction with the client is such as receiving voice information sent by the client and returning the recommended first voice data to the client, and interacting with the matching server, such as submitting voice information. And matching the server and obtaining the matched first voice data from the matching server; the matching server is responsible for performing matching query on the voice information submitted by the interface server in the first voice database to obtain the matched first voice data.
  • These servers are considered as a whole relative to other external devices and are generally considered as "network devices" in their entirety.
  • the receiving device 21 receives the voice information submitted by the user.
  • the present invention requires the network device to interact with the client to implement.
  • the user end is, for example, various functional entities running in the user equipment, such as a webpage in a PC, an APP in a mobile phone, and the like.
  • the user submits a piece of voice information through the user equipment, and the user equipment sends the voice information to the network device, so that the receiving device 21 receives the voice information submitted by the user.
  • the extracting means 22 extracts the sound feature information of the user based on the voice information submitted by the user.
  • the user's voice feature information can be defined from at least the following two dimensions:
  • the perceived dimension is intended to be defined from the perceived characteristics of the sound.
  • the sound feature information is specifically in this dimension such as pitch feature information, pitch feature information, melody feature information, rhythm feature information, audio feature information, speech rate feature information, and the like.
  • the acoustic dimension is intended to be expressed from various features defined by the acoustic angle of the sound.
  • the sound feature information is specifically such as energy feature information, zero-crossing rate, LPC (Linear Prediction Coefficient) parameter information, and the like in this dimension.
  • the zero-crossing rate means the ratio of the sign change of the sound signal, such as the sound signal changing from a positive number to a negative number or a reverse direction.
  • the extracting device 22 may extract the sound feature information by using any known or future feasible technology, which is not limited by the present invention.
  • the matching device 23 queries the preset first voice data database according to the user's voice feature information to obtain a matched first voice data.
  • the network device may pre-establish a first voice database in which various sound feature information of each voice material is stored.
  • the first voice database may also be established and maintained by other devices, and the matching device 23 may have access rights thereto.
  • a song or a participating video segment of each star is stored in the first voice database, and the voice materials are labeled with various sound feature information.
  • the matching device 23 performs a matching query in the first voice database according to the sound feature information recognized by the user to obtain a matched song or video segment.
  • the recommendation device 24 recommends the matched first voice material to the user.
  • the recommendation device 24 recommends the matched songs or video clips and the like to the user, so that the present invention can provide the user with a song or video dubbing suitable for singing or imitating, in order to customize the user's "title song” or “famous". "Song" and help users find their singer suitable for imitation. . This is practical in commercial applications and significantly increases user stickiness and length of use.
  • the invention automatically matches singers, poets, celebrities, etc. by guiding the user through various voice inputs such as speaking, shouting, singing two sentences, reading two poems and the like. Further, the present invention can continue to receive the user's singing of the recommended song, and then the user's own singing is relative to the original singing. Score and share the fun with friends, invite friends to PK, etc. to help users find the songs that they are most suitable to sing, create their own imitations, and gradually guide users to form voice search habits and search engines. A second interaction that provides user stickiness and length of use. Accordingly, the present invention can also be derived from a search engine to a social field by providing a sharing function.
  • the content recommendation device 20 shown in FIG. 2 further includes another matching device (hereinafter referred to as a second matching device).
  • the second matching device queries the preset second voice database according to the content of the user voice information to obtain the matched second voice data; subsequently, the content recommendation device 20 recommends the matched second voice data to the user.
  • the recommendation of the second voice material by the content recommendation device 20 may be combined with the above recommendation for the first voice material, so that the operation performed by the second matching device occurs before the operation performed by the recommendation device 24, and the recommendation device 24 may The first voice material and the second voice data are recommended to the user.
  • the recommendation of the second voice material by the content recommendation device 20 may also be independent of the recommendation of the first voice material by the recommendation device 24.
  • the second matching device may perform the second voice data before, after, or while performing operations of the two devices.
  • the content recommendation device 20 also needs to include another recommendation device (hereinafter referred to as a second recommendation device). To perform a recommendation for the second voice material.
  • the second matching device can still be integrated with the matching device 23, and the second recommending device is integrated with the recommendation device 24. At this time, each integrated device performs two matching operations or two recommended operations, respectively.
  • the second matching device performs a matching query in the second voice database to obtain the same content.
  • Two voice data such as voice data with the same content but different sounds, such as using the same content without dialect.
  • the voice message submitted by the user is “hello” in Mandarin
  • the second matching device is in the first
  • match queries are made to obtain “hello” in various dialects, such as “Hello” in Shanghai dialect and Cantonese.
  • the first voice database can be integrated with the second voice library. That is, in the present invention, only one voice database is used to perform matching query of the first voice data and the second voice data, so that at least the voice feature information and the content information of each voice material are stored in the voice database.
  • the present invention can be implemented in software and/or a combination of software and hardware.
  • the various devices of the present invention can be implemented using an application specific integrated circuit (ASIC) or any other similar hardware device.
  • the software program of the present invention may be executed by a processor to implement the steps or functions described above.
  • the software program (including related data structures) of the present invention can be stored in a computer readable recording medium such as a RAM memory, a magnetic or optical drive or a floppy disk and the like.
  • some of the steps or functions of the present invention may be implemented in hardware, for example, as a circuit that cooperates with a processor to perform various steps or functions.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Health & Medical Sciences (AREA)
  • Signal Processing (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

Procédé et appareil de recommandation de contenu. Le procédé comprend les étapes suivantes : un dispositif de réseau reçoit des informations vocales soumises par un utilisateur (S1) ; extrait des informations de caractéristiques vocales de l'utilisateur selon les informations vocales (S2) ; interroge une première base de données vocale prédéfinie selon les informations de caractéristiques vocales afin d'obtenir des premières données vocales appariées (S3) ; et recommande les premières données vocales à l'utilisateur (S4). La présente solution technique permet de recommander à un utilisateur des données vocales appariées, telles que des chansons et des clips vidéo, selon les informations de caractéristiques vocales de l'utilisateur. Dans les applications vocales interactives actuelles et émergentes, il est possible d'utiliser des procédés d'exploitation intéressants pour inciter les utilisateurs à utiliser automatiquement la voix, et à acquérir ainsi l'habitude d'utiliser la voix pour effectuer une recherche. Par exemple, sur la base des besoins sociaux et de divertissement du public actuel en matière de données vocales, diverses informations de caractéristiques vocales d'un utilisateur, telles que des données vocales et audio, sont analysées automatiquement lorsque l'utilisateur recherche de la musique ou utilise sa voix ; et des chansons appropriées pour une performance de chant, ou un chanteur approprié pour une imitation sont recommandés à l'utilisateur de manière appropriée au moyen d'une analyse vocale de mégadonnées.
PCT/CN2017/079624 2016-11-22 2017-04-06 Procédé et appareil de recommandation de contenu Ceased WO2018094952A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201611036906.4A CN108090081A (zh) 2016-11-22 2016-11-22 一种内容推荐方法与装置
CN201611036906.4 2016-11-22

Publications (1)

Publication Number Publication Date
WO2018094952A1 true WO2018094952A1 (fr) 2018-05-31

Family

ID=62169766

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/079624 Ceased WO2018094952A1 (fr) 2016-11-22 2017-04-06 Procédé et appareil de recommandation de contenu

Country Status (2)

Country Link
CN (1) CN108090081A (fr)
WO (1) WO2018094952A1 (fr)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113486208A (zh) * 2021-06-09 2021-10-08 安徽沐峰数据科技有限公司 一种基于人工智能的语音搜索设备及其搜索方法
CN114697759A (zh) * 2022-04-25 2022-07-01 中国平安人寿保险股份有限公司 虚拟形象视频生成方法及其系统、电子设备、存储介质
CN116662494A (zh) * 2023-04-26 2023-08-29 华南师范大学 一种基于人工智能的辅助教学方法及装置

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109657162A (zh) * 2018-12-17 2019-04-19 掌阅科技股份有限公司 阅读地图的生成方法、电子设备及计算机存储介质
CN110083772A (zh) * 2019-04-29 2019-08-02 北京小唱科技有限公司 基于演唱技巧的歌手推荐方法及装置
CN110688586B (zh) * 2019-09-30 2023-05-09 上海掌门科技有限公司 一种为用户推荐社交活动或好友的方法与设备

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102347060A (zh) * 2010-08-04 2012-02-08 鸿富锦精密工业(深圳)有限公司 电子记录装置及方法
CN102404278A (zh) * 2010-09-08 2012-04-04 盛乐信息技术(上海)有限公司 一种基于声纹识别的点歌系统及其应用方法
CN103685520A (zh) * 2013-12-13 2014-03-26 深圳Tcl新技术有限公司 基于语音识别的歌曲推送的方法和装置

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014026603A (ja) * 2012-07-30 2014-02-06 Hitachi Ltd 音楽選択支援システム、音楽選択支援方法、および音楽選択支援プログラム
CN102833582B (zh) * 2012-08-02 2015-06-17 四川长虹电器股份有限公司 采用语音搜索音视频资源的方法
CN102880693A (zh) * 2012-09-20 2013-01-16 浙江大学 一种基于个体发声能力的音乐推荐方法
CN104657438A (zh) * 2015-02-02 2015-05-27 联想(北京)有限公司 信息处理方法及电子设备
CN105095406A (zh) * 2015-07-09 2015-11-25 百度在线网络技术(北京)有限公司 一种基于用户特征的语音搜索方法及装置
CN105070283B (zh) * 2015-08-27 2019-07-09 百度在线网络技术(北京)有限公司 为歌声语音配乐的方法和装置
CN105575393A (zh) * 2015-12-02 2016-05-11 中国传媒大学 一种基于人声音色的个性化点唱歌曲推荐方法
CN106095925B (zh) * 2016-06-12 2018-07-03 北京邮电大学 一种基于声乐特征的个性化歌曲推荐方法

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102347060A (zh) * 2010-08-04 2012-02-08 鸿富锦精密工业(深圳)有限公司 电子记录装置及方法
CN102404278A (zh) * 2010-09-08 2012-04-04 盛乐信息技术(上海)有限公司 一种基于声纹识别的点歌系统及其应用方法
CN103685520A (zh) * 2013-12-13 2014-03-26 深圳Tcl新技术有限公司 基于语音识别的歌曲推送的方法和装置

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113486208A (zh) * 2021-06-09 2021-10-08 安徽沐峰数据科技有限公司 一种基于人工智能的语音搜索设备及其搜索方法
CN114697759A (zh) * 2022-04-25 2022-07-01 中国平安人寿保险股份有限公司 虚拟形象视频生成方法及其系统、电子设备、存储介质
CN114697759B (zh) * 2022-04-25 2024-04-09 中国平安人寿保险股份有限公司 虚拟形象视频生成方法及其系统、电子设备、存储介质
CN116662494A (zh) * 2023-04-26 2023-08-29 华南师范大学 一种基于人工智能的辅助教学方法及装置

Also Published As

Publication number Publication date
CN108090081A (zh) 2018-05-29

Similar Documents

Publication Publication Date Title
CN115082602B (zh) 生成数字人的方法、模型的训练方法、装置、设备和介质
CN111667811B (zh) 语音合成方法、装置、设备和介质
JP6896690B2 (ja) マルチメディアコンテンツにおける文脈探索
JP6505903B2 (ja) 会話型相互作用システムの検索入力におけるユーザ意図を推定する方法およびそのためのシステム
US11017010B2 (en) Intelligent playing method and apparatus based on preference feedback
US10572602B2 (en) Building conversational understanding systems using a toolset
US10586541B2 (en) Communicating metadata that identifies a current speaker
CN109165302B (zh) 多媒体文件推荐方法及装置
US8972265B1 (en) Multiple voices in audio content
WO2018094952A1 (fr) Procédé et appareil de recommandation de contenu
US20150179170A1 (en) Discriminative Policy Training for Dialog Systems
US20140379323A1 (en) Active learning using different knowledge sources
CN107526809B (zh) 基于人工智能推送音乐的方法和装置
CN105224586A (zh) 从先前会话检索情境
CN107221323A (zh) 语音点歌方法、终端及存储介质
CN107247769A (zh) 语音点歌方法、装置、终端及存储介质
CN105224581A (zh) 在播放音乐时呈现图片的方法和装置
US12511338B2 (en) Conference information query method and apparatus, storage medium, terminal device, and server
CN112799630A (zh) 使用网络可寻址设备创建电影化的讲故事体验
CN110765270A (zh) 用于口语交互的文本分类模型的训练方法及系统
WO2022224584A1 (fr) Dispositif de traitement d'informations, procédé de traitement d'informations, dispositif terminal et procédé d'affichage
CN111324626B (zh) 基于语音识别的搜索方法、装置、计算机设备及存储介质
CN107608799B (zh) 一种用于执行交互指令的方法、设备及存储介质
CN118762712A (zh) 剧场音频作品的生成方法、装置、设备、介质和程序产品
CN114461749A (zh) 对话内容的数据处理方法、装置、电子设备和介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17873702

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17873702

Country of ref document: EP

Kind code of ref document: A1