[go: up one dir, main page]

CN112818841A - Method and related device for recognizing user emotion - Google Patents

Method and related device for recognizing user emotion Download PDF

Info

Publication number
CN112818841A
CN112818841A CN202110128395.3A CN202110128395A CN112818841A CN 112818841 A CN112818841 A CN 112818841A CN 202110128395 A CN202110128395 A CN 202110128395A CN 112818841 A CN112818841 A CN 112818841A
Authority
CN
China
Prior art keywords
emotion
current
input information
user
type
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110128395.3A
Other languages
Chinese (zh)
Other versions
CN112818841B (en
Inventor
叶祺
薛静静
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Sogou Technology Development Co Ltd
Original Assignee
Beijing Sogou Technology Development Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Sogou Technology Development Co Ltd filed Critical Beijing Sogou Technology Development Co Ltd
Priority to CN202110128395.3A priority Critical patent/CN112818841B/en
Publication of CN112818841A publication Critical patent/CN112818841A/en
Application granted granted Critical
Publication of CN112818841B publication Critical patent/CN112818841B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/70Multimodal biometrics, e.g. combining information from different biometric modalities

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

本申请公开了一种识别用户情绪的方法及相关装置,该方法包括:首先,获得用户的当前输入信息和情绪因素;然后,对当前输入信息进行情绪特征提取,获得当前输入信息对应的当前情绪特征;最后,利用情绪识别模型处理当前情绪特征和情绪因素,获得用户的当前情绪类型。由此可见,先对用户的当前输入信息进行情绪特征提取得到当前情绪特征,再将当前情绪特征结合用户的情绪因素进行情绪识别,得到用户的当前情绪类型;不仅分析当前情绪特征所表征的用户情绪,而且分析情绪因素对当前情绪特征所表征的用户情绪的影响,实现多角度综合情绪识别,大大提高了用户情绪识别的准确性。

Figure 202110128395

The present application discloses a method and a related device for identifying user emotions. The method includes: first, obtaining the current input information and emotional factors of the user; then, performing emotional feature extraction on the current input information to obtain the current emotion corresponding to the current input information Finally, the emotion recognition model is used to process the current emotion features and emotion factors to obtain the user's current emotion type. It can be seen that the current emotional features are first extracted from the user's current input information to obtain the current emotional features, and then the current emotional features are combined with the user's emotional factors to perform emotional recognition to obtain the user's current emotional type; not only the users represented by the current emotional features are analyzed. It also analyzes the influence of emotional factors on the user's emotions represented by the current emotional characteristics, realizes multi-angle comprehensive emotion recognition, and greatly improves the accuracy of user emotion recognition.

Figure 202110128395

Description

Method and related device for recognizing user emotion
Technical Field
The present application relates to the field of data processing technologies, and in particular, to a method and a related apparatus for recognizing a user emotion.
Background
At the present stage, the user interacts with the artificial intelligence product through the input tool, that is, the input tool is an important entrance for the user to experience the artificial intelligence. When the user experiences artificial intelligence through the input tool, in order to better understand the user, more intelligent artificial intelligence service is provided, user experience is improved, user stickiness is increased, and it is particularly important to recognize user emotion under a user input scene.
In the prior art, in a user input scene, emotion recognition is generally performed on input information of a user to obtain a user emotion. However, through research, the inventor finds that the method only identifies the user emotion represented by the input information of the user, and does not consider the influence of other influence information on the user emotion represented by the input information of the user, which easily causes that the identified user emotion is not accurate, and is likely to have a large deviation from the real user emotion.
Disclosure of Invention
In view of this, the present application provides a method and a related apparatus for recognizing a user emotion, which implement multi-angle comprehensive emotion recognition and greatly improve accuracy of user emotion recognition.
In a first aspect, an embodiment of the present application provides a method for recognizing a user emotion, where the method includes:
obtaining current input information and emotion factors of a user;
extracting emotion characteristics of the current input information to obtain current emotion characteristics corresponding to the current input information;
and obtaining the current emotion type of the user based on the current emotion characteristics, the emotion factors and an emotion recognition model.
Optionally, the current input information includes at least two of current text input information, current voice input information, and current image input information.
Optionally, the current image input information includes a current face image and/or a current body posture image.
Optionally, the emotional factors include a character type and/or a historical emotion type.
Optionally, the personality type of the user is obtained based on preset time period input information of the user and a personality identification model, wherein the personality identification model is obtained by training a first preset identification network in advance based on the personality characteristics and the personality type labels corresponding to the preset time period input information samples.
Optionally, the training step of the character recognition model includes:
inputting the corresponding character features of the input information samples in the preset time period into the first preset identification network according to the time sequence to obtain a predicted character type;
adjusting parameters of the first pre-set identification network based on the predicted personality type, the personality type label, and a first pre-set loss function;
and taking the trained first preset recognition network as the character recognition model.
Optionally, the obtaining of the personality type of the user includes:
performing character feature extraction on the input information of the user in a preset time period to obtain character features corresponding to the input information of the preset time period;
determining a time sequence of character features corresponding to the preset time period input information;
and inputting the character features corresponding to the preset time period input information into the character recognition model according to the time sequence to obtain the character type of the user.
Optionally, the historical emotion types of the user are pre-stored in a map manner and/or an embedded manner.
Optionally, the emotion recognition model is obtained by pre-training a second preset recognition network based on emotion characteristics, emotion factor samples and corresponding emotion type labels corresponding to the preset input information samples.
Optionally, the training step of the emotion recognition model includes:
inputting the emotion characteristics corresponding to the preset input information samples and the emotion factor samples into the second preset identification network to obtain predicted emotion types;
adjusting parameters of the second preset identification network based on the predicted emotion type, the emotion type tag and a second preset loss function;
and taking the trained second preset recognition network as the emotion recognition model.
In a second aspect, an embodiment of the present application provides an apparatus for recognizing a user emotion, where the apparatus includes:
a first obtaining unit for obtaining current input information and emotion factors of a user;
the second obtaining unit is used for extracting emotion characteristics of the current input information to obtain current emotion characteristics corresponding to the current input information;
and the third obtaining unit is used for obtaining the current emotion type of the user based on the current emotion characteristics, the emotion factors and the emotion recognition model.
Optionally, the current input information includes at least two of current text input information, current voice input information, and current image input information.
Optionally, the current image input information includes a current face image and/or a current body posture image.
Optionally, the emotional factors include a character type and/or a historical emotion type.
Optionally, the personality type of the user is obtained based on the preset time period input information of the user and a personality identification model; the character recognition model is obtained by pre-training a first preset recognition network based on character features and character type labels corresponding to input information samples in a preset time period.
Optionally, the apparatus further includes a first training unit, where the first training unit is configured to:
inputting the corresponding character features of the input information samples in the preset time period into the first preset identification network according to the time sequence to obtain a predicted character type;
adjusting parameters of the first pre-set identification network based on the predicted personality type, the personality type label, and a first pre-set loss function;
and taking the trained first preset recognition network as the character recognition model.
Optionally, the apparatus further includes a fourth obtaining unit, where the fourth obtaining unit is configured to:
performing character feature extraction on the input information of the user in a preset time period to obtain character features corresponding to the input information of the preset time period;
determining a time sequence of character features corresponding to the preset time period input information;
and inputting the character features corresponding to the preset time period input information into the character recognition model according to the time sequence to obtain the character type of the user.
Optionally, the historical emotion types of the user are pre-stored in a map manner and/or an embedded manner.
Optionally, the emotion recognition model is obtained by pre-training a second preset recognition network based on emotion characteristics, emotion factor samples and corresponding emotion type labels corresponding to the preset input information samples.
Optionally, the apparatus further includes a second training unit, where the second training unit is configured to:
inputting the emotion characteristics corresponding to the preset input information samples and the emotion factor samples into the second preset identification network to obtain predicted emotion types;
adjusting parameters of the second preset identification network based on the predicted emotion type, the emotion type tag and a second preset loss function;
and taking the trained second preset recognition network as the emotion recognition model.
In a third aspect, embodiments of the present application provide an apparatus for identifying a user's mood, the apparatus comprising a memory, and one or more programs, wherein the one or more programs are stored in the memory, and wherein the one or more programs configured to be executed by the one or more processors include instructions for:
obtaining current input information and emotion factors of a user;
extracting emotion characteristics of the current input information to obtain current emotion characteristics corresponding to the current input information;
and obtaining the current emotion type of the user based on the current emotion characteristics, the emotion factors and an emotion recognition model.
Optionally, the current input information includes at least two of current text input information, current voice input information, and current image input information.
Optionally, the current image input information includes a current face image and/or a current body posture image.
Optionally, the emotional factors include a character type and/or a historical emotion type.
Optionally, the personality type of the user is obtained based on preset time period input information of the user and a personality identification model, wherein the personality identification model is obtained by training a first preset identification network in advance based on the personality characteristics and the personality type labels corresponding to the preset time period input information samples.
Optionally, the device is also configured to execute the one or more programs by the one or more processors including instructions for:
inputting the corresponding character features of the input information samples in the preset time period into the first preset identification network according to the time sequence to obtain a predicted character type;
adjusting parameters of the first pre-set identification network based on the predicted personality type, the personality type label, and a first pre-set loss function;
and taking the trained first preset recognition network as the character recognition model.
Optionally, the device is also configured to execute the one or more programs by the one or more processors including instructions for:
performing character feature extraction on the input information of the user in a preset time period to obtain character features corresponding to the input information of the preset time period;
determining a time sequence of character features corresponding to the preset time period input information;
and inputting the character features corresponding to the preset time period input information into the character recognition model according to the time sequence to obtain the character type of the user.
Optionally, the historical emotion types of the user are pre-stored in a map manner and/or an embedded manner.
Optionally, the emotion recognition model is obtained by pre-training a second preset recognition network based on emotion characteristics, emotion factor samples and corresponding emotion type labels corresponding to the preset input information samples.
Optionally, the device is also configured to execute the one or more programs by the one or more processors including instructions for:
inputting the emotion characteristics corresponding to the preset input information samples and the emotion factor samples into the second preset identification network to obtain predicted emotion types;
adjusting parameters of the second preset identification network based on the predicted emotion type, the emotion type tag and a second preset loss function;
and taking the trained second preset recognition network as the emotion recognition model.
In a fourth aspect, the present application provides a machine-readable medium having stored thereon instructions, which when executed by one or more processors, cause an apparatus to perform the method for identifying a user emotion described in any of the first aspects above.
Compared with the prior art, the method has the advantages that:
by adopting the technical scheme of the embodiment of the application, firstly, the current input information and the emotion factors of a user are obtained; then, extracting emotion characteristics of the current input information to obtain current emotion characteristics corresponding to the current input information; and finally, processing the current emotion characteristics and the emotion factors by using the emotion recognition model to obtain the current emotion type of the user. Therefore, emotion characteristics of current input information of a user are extracted to obtain current emotion characteristics, and then emotion recognition is carried out on the current emotion characteristics by combining the emotion factors of the user to obtain the current emotion type of the user; the emotion recognition method and the emotion recognition device not only analyze the emotion of the user represented by the current emotion characteristics, but also analyze the influence of emotion factors on the emotion of the user represented by the current emotion characteristics, realize multi-angle comprehensive emotion recognition, and greatly improve the accuracy of emotion recognition of the user.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings needed to be used in the description of the embodiments of the present application will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments described in the present application, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 is a schematic diagram of a system framework related to an application scenario in an embodiment of the present application;
fig. 2 is a flowchart illustrating a method for recognizing a user emotion according to an embodiment of the present application;
fig. 3 is a schematic structural diagram of an apparatus for recognizing a user emotion according to an embodiment of the present application;
fig. 4 is a schematic structural diagram of an apparatus for recognizing a user emotion according to an embodiment of the present application;
fig. 5 is a schematic structural diagram of a server according to an embodiment of the present application.
Detailed Description
In order to make the technical solutions of the present application better understood, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
Currently, in a user input scenario, emotion recognition is generally performed on input information of a user to obtain a user emotion. However, the inventor finds that the method does not consider the influence of other influence information on the user emotion represented by the input information of the user, and only identifies the user emotion represented by the input information of the user, which easily causes that the identified user emotion is not accurate and is likely to have a large deviation from the true user emotion.
In order to solve this problem, in the embodiment of the present application, first, current input information and emotional factors of a user are obtained; then, extracting emotion characteristics of the current input information to obtain current emotion characteristics corresponding to the current input information; and finally, processing the current emotion characteristics and the emotion factors by using the emotion recognition model to obtain the current emotion type of the user. Therefore, emotion characteristics of current input information of a user are extracted to obtain current emotion characteristics, and then emotion recognition is carried out on the current emotion characteristics by combining the emotion factors of the user to obtain the current emotion type of the user; the emotion recognition method and the emotion recognition device not only analyze the emotion of the user represented by the current emotion characteristics, but also analyze the influence of emotion factors on the emotion of the user represented by the current emotion characteristics, realize multi-angle comprehensive emotion recognition, and greatly improve the accuracy of emotion recognition of the user.
For example, one of the scenarios in the embodiment of the present application may be applied to the scenario shown in fig. 1, where the scenario includes a client 101 and a processor 102, and the client 101 and the processor 102 are in the same artificial intelligence product. The user inputs through an input product installed on the client 101, and the processor 102 recognizes the emotion of the user by adopting the implementation mode provided by the embodiment of the application, so that the processor 102 performs artificial intelligence service based on the recognized emotion of the user, and the artificial intelligence product is more intelligent.
It is to be understood that in the above application scenarios, although the actions of the embodiments of the present application are described as being performed by the processor 102; however, the present application is not limited in terms of executing a subject as long as the actions disclosed in the embodiments of the present application are executed.
It is to be understood that the above scenario is only one example of a scenario provided in the embodiment of the present application, and the embodiment of the present application is not limited to this scenario.
The following describes in detail a specific implementation manner of the method for recognizing a user emotion and the related apparatus in the embodiments of the present application by using embodiments in conjunction with the accompanying drawings.
Exemplary method
Referring to fig. 2, a flow chart of a method for recognizing a user emotion in an embodiment of the present application is shown. In this embodiment, the method may include, for example, the steps of:
step 201: current input information and emotional factors of the user are obtained.
In the embodiment of the application, in order to avoid emotion recognition based on current input information of a single modality, the obtained emotion accuracy of the user is low, and the current input information of at least two different modalities of the user can be obtained, so that the emotion accuracy of the subsequent recognition user is improved. The current input information may refer to input information at the current time, or may refer to input information at the current time period.
Common user input modes include a text input mode, a voice input mode and an image input mode; the current input information corresponding to the text input mode is current text input information, the current input information corresponding to the voice input mode is current voice input information, the current input information corresponding to the image input mode is current image input information, and the current input information is generally image input information such as a face image and a body posture image which can represent the emotion of a user. Therefore, in an optional implementation manner of this embodiment of the present application, the current input information includes at least two of current text input information, current voice input information, and current image input information; the current image input information includes a current face image and/or a current body pose image.
The current face image may be a current still face image, for example, a current still face picture, or a current dynamic face image, for example, a current dynamic face picture or a current face video; similarly, the current body posture image may be either a current static body posture image, such as a current static body posture picture, or a current dynamic body posture image, such as a current dynamic body posture picture or a current body posture video.
Because only the user emotion represented by the input information of the user is recognized, the influence of other influence information on the user emotion represented by the input information of the user is not considered, the recognized user emotion is not accurate, and a large deviation from the real user emotion is likely to occur. Therefore, in the embodiment of the present application, on the basis of obtaining the current input information of the user, information that affects the emotion of the user represented by the current input information of the user needs to be obtained as an emotion factor of the user.
The emotion of the user represented by the current input information of the user under different character types is different, and the emotion of the user represented by the current input information of the user under different historical emotion types is also different; that is, the personality type and the historical emotion type of the user affect the emotion of the user represented by the current input information of the user, and any one or two of the personality type and the historical emotion type can be used as the emotion factor of the user. Therefore, in an alternative implementation of the embodiment of the present application, the emotional factors include a character type and/or a historical emotion type.
In the embodiment of the application, when the emotion factor includes the personality type, because the personality type of the user can be reflected by the input information of the user in the preset time period, the personality identification processing can be performed on the input information of the user in the preset time period, and the personality type of the user is obtained. For example, the personality type may be open, responsible, outward, hommized, neutral, or depolarized, and the like.
The character recognition processing is based on the premise that a character feature and a character type label corresponding to an information sample are input in a preset time period, and a first preset recognition network is trained in advance to obtain a character recognition model. Based on the character recognition model, character recognition processing is carried out on the input information of the user in the preset time period, and the character type of the user can be obtained. Therefore, in an optional implementation manner of the embodiment of the present application, the personality type of the user is obtained based on the preset time period input information of the user and a personality identification model; the character recognition model is obtained by pre-training a first preset recognition network based on character features and character type labels corresponding to input information samples in a preset time period.
The training of the character recognition model refers to: after character features corresponding to the input information samples in a preset time period ordered according to the time sequence and corresponding character type labels are used as training data; training a preset identification network, namely a first preset identification network, by the training data to fully mine the association information between the character features and the character type labels corresponding to the input information samples at the preset time period ordered according to the time sequence; after a certain amount of the training data are subjected to iterative training for multiple times, the character recognition model can be obtained.
When the method is specifically implemented, firstly, the character features corresponding to the information samples are input in a preset time period, and are input into a first preset identification network according to a time sequence, and the first preset identification network can predict character types and output the predicted character types; then, calculating the loss of the predicted character type and the character type label by using a first preset loss function so as to adjust the parameter of the first preset identification network; and after repeated iterative training until the preset iterative training times are reached or the first preset recognition network is converged, finishing the training of the first preset recognition network, and taking the trained first preset recognition network as a character recognition model. Therefore, in an optional implementation manner of the embodiment of the present application, the training step of the character recognition model includes the following steps:
step A: inputting the corresponding character features of the input information samples in the preset time period into the first preset identification network according to the time sequence to obtain a predicted character type;
and B: adjusting parameters of the first pre-set identification network based on the predicted personality type, the personality type label, and a first pre-set loss function;
and C: and taking the trained first preset recognition network as the character recognition model.
Correspondingly, when the personality type of the user is obtained, firstly, the personality characteristics in the input information of the user in a preset time period need to be extracted, then, the time sequence of the personality characteristics needs to be determined, and finally, the personality characteristics are input into the personality identification model according to the time sequence, so that the personality type of the user can be obtained. In an optional implementation manner of the embodiment of the present application, the obtaining of the personality type of the user includes:
step D: performing character feature extraction on the input information of the user in a preset time period to obtain character features corresponding to the input information of the preset time period;
step E: determining a time sequence of character features corresponding to the preset time period input information;
step F: and inputting the character features corresponding to the preset time period input information into the character recognition model according to the time sequence to obtain the character type of the user.
In the embodiment of the present application, when the emotion factor includes a historical emotion type, since the historical emotion type of the user is an emotion type obtained by historical emotion recognition for the user, the historical emotion type of the user is stored in the user emotion database in advance with respect to current input information of the user. The storage mode of the historical emotion type of the user can be, for example, a map mode and/or an embedding mode. That is, in an optional implementation manner of the embodiment of the present application, the historical emotion types of the user are pre-stored in a graph manner and/or an embedded manner.
When the historical emotion types of the user are stored in advance in a graph mode, a historical emotion type graph is obtained and is composed of a plurality of historical emotion type triples. The single historical emotion type triple is in (s, p, o) format, wherein s refers to the user, p refers to the historical emotion type, and p represents the time relationship between the user and the historical emotion type. The historical emotion type map is divided into a long-term historical emotion type map and a short-term historical emotion type map, the long-term historical emotion type of the user is memorized by the long-term historical emotion type map, the short-term historical emotion type of the user is memorized by the short-term historical emotion type map, and the short-term historical emotion type map needs to be updated in time.
When the historical emotion types of the user are stored in advance in an embedding mode, an embedded historical emotion type module is obtained, and the embedded historical emotion type module represents a plurality of historical emotion type triples by using continuous numerical values. The embedded historical emotion type module is also divided into a long-term embedded historical emotion type module and a short-term embedded historical emotion type module, the long-term embedded historical emotion type module memorizes the long-term historical emotion type of the user, the short-term embedded historical emotion type module memorizes the short-term historical emotion type of the user, and the short-term embedded historical emotion type module also needs to be updated in time.
Step 202: and extracting emotion characteristics of the current input information to obtain current emotion characteristics corresponding to the current input information.
In the embodiment of the application, for the current input information of the user, firstly, the emotional feature of the user emotion represented in the current input information needs to be extracted as the current emotional feature corresponding to the current input information. In step 202, in a specific embodiment, the current input information corresponding to the at least two different modalities includes at least two of current text input information, current voice input information, and current image input information, and the specific implementation manners of the current input information corresponding to the three different modalities of the current text input information, the current voice input information, and the current image input information are different, which are specifically as follows:
first, for the current text input information, when the current input information of at least two different modalities includes the current text input information, step 202 may include, for example: and extracting emotion-related semantic features of the first current text input information to obtain first current emotion semantic features corresponding to the first current text input information.
Second, for the current speech input information, when the current input information of at least two different modalities includes the current speech input information, step 202 may include, for example: converting the current voice input information into second current text input information; extracting emotion-related semantic features of the second current text input information to obtain second current emotion semantic features corresponding to the second current text input information; and performing emotion-related voice feature extraction on the current voice input information to obtain current emotion voice features corresponding to the current voice input information. The current voice input information is obtained through preprocessing such as pre-emphasis, framing and windowing, so that the influence of factors such as aliasing, higher harmonic distortion and high frequency caused by a user vocal organ and equipment for collecting the voice input information on the quality of the current voice input information is avoided.
Third, for the current image input information, when the current input information of at least two different modalities includes the current image input information, step 202 may include, for example: and performing emotion-related image feature extraction on the current image input information to obtain current emotion image features corresponding to the current image input information. When the current image input information comprises a current face image, the current emotion image characteristics corresponding to the current face image mainly refer to current face characteristics and the like; when the current image input information includes the current body posture image, the current emotion image characteristics corresponding to the current body posture image mainly refer to current human body characteristics and the like.
Step 203: and obtaining the current emotion type of the user based on the current emotion characteristics, the emotion factors and an emotion recognition model.
In the embodiment of the application, after the current emotion characteristics and the corresponding emotion factors are obtained, in order to avoid the phenomenon that the emotion of the user obtained by recognition is not accurate and is likely to have a large deviation with the true emotion of the user because only emotion recognition is performed on the current emotion characteristics; the current emotion characteristics and corresponding emotion factors are required to be integrated for emotion recognition, and the current emotion type of the user is obtained; the emotion recognition method and the emotion recognition device not only analyze the emotion of the user represented by the current emotion characteristics, but also analyze the influence of emotion factors on the emotion of the user represented by the current emotion characteristics, realize multi-angle comprehensive emotion recognition, and greatly improve the accuracy of emotion recognition of the user.
The emotion recognition is carried out by integrating the current emotion characteristics and the corresponding emotion factors, actually, the current emotion characteristics and the corresponding emotion factors are input into an emotion recognition model obtained through pre-training, and the emotion types obtained through recognition are used as the current emotion types of the user. In an optional implementation manner of the embodiment of the application, the emotion recognition model is obtained by training a second preset recognition network in advance based on the emotion characteristics, the emotion factor samples and the corresponding emotion type labels corresponding to the preset input information samples.
The training of the emotion recognition model refers to: taking emotion characteristics, emotion factor samples and emotion type labels corresponding to preset input information samples as training data; another preset identification network, namely a second preset identification network, needs to be trained through the training data so as to fully mine emotion characteristics corresponding to preset input information samples, and associated information between the emotion factor samples and the emotion type labels; after a certain amount of the training data are subjected to repeated iterative training, the emotion recognition model can be obtained.
When the method is specifically implemented, firstly, emotion characteristics and emotion factor samples corresponding to preset input information samples need to be input into a second preset identification network, and the second preset identification network can identify emotion types and output predicted emotion types; then, calculating the loss of the predicted emotion type and the emotion type label by using a second preset loss function so as to adjust the parameters of a second preset identification network; and after repeated iterative training, finishing the training of the second preset recognition network until the preset iterative training times are reached or the second preset recognition network is converged, and taking the trained second preset recognition network as an emotion recognition model. That is, in an optional implementation manner of the embodiment of the present application, the training step of the emotion recognition model includes:
step G: inputting the emotion characteristics corresponding to the preset input information samples and the emotion factor samples into the second preset identification network to obtain predicted emotion types;
step H: adjusting parameters of the second preset identification network based on the predicted emotion type, the emotion type tag and a second preset loss function;
step I: and taking the trained second preset recognition network as the emotion recognition model.
As an example, obtaining the current input information of the user includes current text input information and current image input information of the user, where the current text input information is that "i like to eat apple" and the current image input information is a current face image; the emotion image information includes a character type, and the character type is extroversion. Performing emotion-related emotional feature extraction on the current text input information 'i likes apple' well ', and obtaining current emotional semantic features corresponding to the current text input information' i likes apple 'well'; performing emotion-related image feature extraction on current image input information, namely a current facial image, and obtaining current image input information, namely current emotion image features corresponding to the current facial image; and the current emotion semantic features and the current emotion image features form current emotion features. Compared with the situation that the current emotion type of the user is happy, the current emotion type of the user is obtained based on the current emotion characteristics, the character type-extroversion and the emotion recognition model, and the multi-angle comprehensive emotion recognition is more accurate.
As another example, in the above example, the current input information is not changed, the emotion image information includes a character type, and the character type is a negativity. Referring to the above example details, in a case where the current emotion type of the user obtained by emotion recognition only on the current emotion feature is happy, the current emotion type of the user is obtained as very happy based on the current emotion feature, the character type — passivity, and the emotion recognition model.
Through various implementation manners provided by the embodiment, firstly, current input information and emotion factors of a user are obtained; then, extracting emotion characteristics of the current input information to obtain current emotion characteristics corresponding to the current input information; and finally, processing the current emotion characteristics and the emotion factors by using the emotion recognition model to obtain the current emotion type of the user. Therefore, emotion characteristics of current input information of a user are extracted to obtain current emotion characteristics, and then emotion recognition is carried out on the current emotion characteristics by combining the emotion factors of the user to obtain the current emotion type of the user; the emotion recognition method and the emotion recognition device not only analyze the emotion of the user represented by the current emotion characteristics, but also analyze the influence of emotion factors on the emotion of the user represented by the current emotion characteristics, realize multi-angle comprehensive emotion recognition, and greatly improve the accuracy of emotion recognition of the user.
Exemplary devices
Referring to fig. 3, a schematic structural diagram of an apparatus for recognizing a user emotion in an embodiment of the present application is shown. In this embodiment, the apparatus may specifically include:
a first obtaining unit 301 for obtaining current input information and emotional factors of a user;
a second obtaining unit 302, configured to perform emotion feature extraction on the current input information, and obtain a current emotion feature corresponding to the current input information;
a third obtaining unit 303, configured to obtain a current emotion type of the user based on the current emotion feature, the emotion factor, and an emotion recognition model.
In an optional implementation manner of the embodiment of the present application, the current input information includes at least two of current text input information, current voice input information, and current image input information.
In an optional implementation manner of the embodiment of the present application, the current image input information includes a current face image and/or a current body posture image.
In an alternative implementation of the embodiment of the present application, the emotional factors include a character type and/or a historical emotion type.
In an optional implementation manner of the embodiment of the present application, the personality type of the user is obtained based on the preset time period input information of the user and a personality identification model; the character recognition model is obtained by pre-training a first preset recognition network based on character features and character type labels corresponding to input information samples in a preset time period.
In an optional implementation manner of the embodiment of the present application, the apparatus further includes a first training unit, where the first training unit is configured to:
inputting the corresponding character features of the input information samples in the preset time period into the first preset identification network according to the time sequence to obtain a predicted character type;
adjusting parameters of the first pre-set identification network based on the predicted personality type, the personality type label, and a first pre-set loss function;
and taking the trained first preset recognition network as the character recognition model.
In an optional implementation manner of the embodiment of the present application, the apparatus further includes a fourth obtaining unit, where the fourth obtaining unit is configured to:
performing character feature extraction on the input information of the user in a preset time period to obtain character features corresponding to the input information of the preset time period;
determining a time sequence of character features corresponding to the preset time period input information;
and inputting the character features corresponding to the preset time period input information into the character recognition model according to the time sequence to obtain the character type of the user.
In an optional implementation manner of the embodiment of the present application, the historical emotion types of the user are pre-stored in a graph manner and/or an embedded manner.
In an optional implementation manner of the embodiment of the application, the emotion recognition model is obtained by training a second preset recognition network in advance based on the emotion characteristics, the emotion factor samples and the corresponding emotion type labels corresponding to the preset input information samples.
In an optional implementation manner of the embodiment of the present application, the apparatus further includes a second training unit, where the second training unit is configured to:
inputting the emotion characteristics corresponding to the preset input information samples and the emotion factor samples into the second preset identification network to obtain predicted emotion types;
adjusting parameters of the second preset identification network based on the predicted emotion type, the emotion type tag and a second preset loss function;
and taking the trained second preset recognition network as the emotion recognition model.
Through various implementation manners provided by the embodiment, firstly, current input information and emotion factors of a user are obtained; then, extracting emotion characteristics of the current input information to obtain current emotion characteristics corresponding to the current input information; and finally, processing the current emotion characteristics and the emotion factors by using the emotion recognition model to obtain the current emotion type of the user. Therefore, emotion characteristics of current input information of a user are extracted to obtain current emotion characteristics, and then emotion recognition is carried out on the current emotion characteristics by combining the emotion factors of the user to obtain the current emotion type of the user; the emotion recognition method and the emotion recognition device not only analyze the emotion of the user represented by the current emotion characteristics, but also analyze the influence of emotion factors on the emotion of the user represented by the current emotion characteristics, realize multi-angle comprehensive emotion recognition, and greatly improve the accuracy of emotion recognition of the user.
Fig. 4 is a block diagram illustrating an apparatus 400 for recognizing a user's emotion according to an exemplary embodiment. For example, the apparatus 400 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, an exercise device, a personal digital assistant, and the like.
Referring to fig. 4, the apparatus 400 may include one or more of the following components: processing components 402, memory 404, power components 406, multimedia components 408, audio components 410, input/output (I/O) interfaces 412, sensor components 414, and communication components 416.
The processing component 402 generally controls overall operation of the apparatus 400, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. The processing component 402 may include one or more processors 420 to execute instructions to perform all or a portion of the steps of the methods described above. Further, the processing component 402 can include one or more modules that facilitate interaction between the processing component 402 and other components. For example, the processing component 402 can include a multimedia module to facilitate interaction between the multimedia component 408 and the processing component 402.
The memory 404 is configured to store various types of data to support operations at the device 400. Examples of such data include instructions for any application or method operating on the device 400, contact data, phonebook data, messages, pictures, videos, and so forth. The memory 404 may be implemented by any type or combination of volatile or non-volatile memory devices such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.
Power supply components 406 provide power to the various components of device 400. The power components 406 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the apparatus 400.
The multimedia component 408 includes a screen that provides an output interface between the device 400 and the user. In some embodiments, the screen may include a Liquid Crystal Display (LCD) and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive an input signal from a user. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure correlated to the touch or slide operation. In some embodiments, the multimedia component 408 includes a front facing camera and/or a rear facing camera. The front-facing camera and/or the rear-facing camera may receive external multimedia data when the device 400 is in an operational mode, such as a shooting mode or a video mode. Each front camera and rear camera may be a fixed optical lens system or have a focal length and optical zoom capability.
The audio component 410 is configured to output and/or input audio signals. For example, audio component 410 includes a Microphone (MIC) configured to receive external audio signals when apparatus 400 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode. The received audio signals may further be stored in the memory 404 or transmitted via the communication component 416. In some embodiments, audio component 410 also includes a speaker for outputting audio signals.
The I/O interface 412 provides an interface between the processing component 402 and peripheral interface modules, which may be keyboards, click wheels, buttons, etc. These buttons may include, but are not limited to: a home button, a volume button, a start button, and a lock button.
The sensor component 414 includes one or more sensors for providing various aspects of status assessment for the apparatus 400. For example, the sensor component 414 can detect the open/closed state of the device 400, the relative positioning of components, such as a display and keypad of the apparatus 400, the sensor component 414 can also detect a change in the position of the apparatus 400 or a component of the apparatus 400, the presence or absence of user contact with the apparatus 400, orientation or acceleration/deceleration of the apparatus 400, and a change in the temperature of the apparatus 400. The sensor assembly 414 may include a proximity sensor configured to detect the presence of a nearby object without any physical contact. The sensor assembly 414 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor assembly 414 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
The communication component 416 is configured to facilitate wired or wireless communication between the apparatus 400 and other devices. The apparatus 400 may access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof. In an exemplary embodiment, the communication component 416 receives broadcast signals or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 416 further includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
In an exemplary embodiment, the apparatus 400 may be implemented by one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), controllers, micro-controllers, microprocessors or other electronic components for performing the above-described methods.
In an exemplary embodiment, a non-transitory computer-readable storage medium comprising instructions, such as the memory 404 comprising instructions, executable by the processor 420 of the apparatus 400 to perform the above-described method is also provided. For example, the non-transitory computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
A non-transitory computer readable storage medium having instructions therein, which when executed by a processor of a mobile terminal, enable the mobile terminal to perform a method of recognizing a user's mood, the method comprising:
obtaining current input information and emotion factors of a user;
extracting emotion characteristics of the current input information to obtain current emotion characteristics corresponding to the current input information;
and obtaining the current emotion type of the user based on the current emotion characteristics, the emotion factors and an emotion recognition model.
In an optional implementation manner of the embodiment of the present application, the current input information includes at least two of current text input information, current voice input information, and current image input information.
In an optional implementation manner of the embodiment of the present application, the current image input information includes a current face image and/or a current body posture image.
In an alternative implementation of the embodiment of the present application, the emotional factors include a character type and/or a historical emotion type.
In an optional implementation manner of the embodiment of the present application, the personality type of the user is obtained based on preset time period input information of the user and a personality identification model, where the personality identification model is obtained by training a first preset identification network in advance based on a personality feature and a personality type tag corresponding to a preset time period input information sample.
In an optional implementation manner of the embodiment of the present application, the training step of the character recognition model includes:
inputting the corresponding character features of the input information samples in the preset time period into the first preset identification network according to the time sequence to obtain a predicted character type;
adjusting parameters of the first pre-set identification network based on the predicted personality type, the personality type label, and a first pre-set loss function;
and taking the trained first preset recognition network as the character recognition model.
In an optional implementation manner of the embodiment of the present application, the obtaining of the personality type of the user includes:
performing character feature extraction on the input information of the user in a preset time period to obtain character features corresponding to the input information of the preset time period;
determining a time sequence of character features corresponding to the preset time period input information;
and inputting the character features corresponding to the preset time period input information into the character recognition model according to the time sequence to obtain the character type of the user.
In an optional implementation manner of the embodiment of the present application, the historical emotion types of the user are pre-stored in a graph manner and/or an embedded manner.
In an optional implementation manner of the embodiment of the application, the emotion recognition model is obtained by training a second preset recognition network in advance based on the emotion characteristics, the emotion factor samples and the corresponding emotion type labels corresponding to the preset input information samples.
In an optional implementation manner of the embodiment of the present application, the training step of the emotion recognition model includes:
inputting the emotion characteristics corresponding to the preset input information samples and the emotion factor samples into the second preset identification network to obtain predicted emotion types;
adjusting parameters of the second preset identification network based on the predicted emotion type, the emotion type tag and a second preset loss function;
and taking the trained second preset recognition network as the emotion recognition model.
Fig. 5 is a schematic structural diagram of a server in an embodiment of the present application. The server 500 may vary widely in configuration or performance and may include one or more Central Processing Units (CPUs) 522 (e.g., one or more processors) and memory 532, one or more storage media 530 (e.g., one or more mass storage devices) storing applications 542 or data 544. Memory 532 and storage media 530 may be, among other things, transient storage or persistent storage. The program stored on the storage medium 530 may include one or more modules (not shown), each of which may include a series of instruction operations for the server. Still further, the central processor 522 may be configured to communicate with the storage medium 530, and execute a series of instruction operations in the storage medium 530 on the server 500.
The server 500 may also include one or more power supplies 526, one or more wired or wireless network interfaces 550, one or more input-output interfaces 558, one or more keyboards 556, and/or one or more operating systems 541, such as Windows Server, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM, etc.
The embodiments in the present description are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other. The device disclosed by the embodiment corresponds to the method disclosed by the embodiment, so that the description is simple, and the relevant points can be referred to the method part for description.
Those of skill would further appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both, and that the various illustrative components and steps have been described above generally in terms of their functionality in order to clearly illustrate this interchangeability of hardware and software. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. The terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The foregoing is merely a preferred embodiment of the present application and is not intended to limit the present application in any way. Although the present application has been described with reference to the preferred embodiments, it is not intended to limit the present application. Those skilled in the art can now make numerous possible variations and modifications to the disclosed embodiments, or modify equivalent embodiments, using the methods and techniques disclosed above, without departing from the scope of the claimed embodiments. Therefore, any simple modification, equivalent change and modification made to the above embodiments according to the technical essence of the present application still fall within the protection scope of the technical solution of the present application without departing from the content of the technical solution of the present application.

Claims (10)

1.一种识别用户情绪的方法,其特征在于,包括:1. a method for identifying user emotion, is characterized in that, comprises: 获得用户的当前输入信息和情绪因素;Obtain the user's current input information and emotional factors; 对所述当前输入信息进行情绪特征提取,获得所述当前输入信息对应的当前情绪特征;Perform emotional feature extraction on the current input information to obtain the current emotional feature corresponding to the current input information; 基于所述当前情绪特征、所述情绪因素和情绪识别模型,获得所述用户的当前情绪类型。Based on the current emotion feature, the emotion factor and the emotion recognition model, the current emotion type of the user is obtained. 2.根据权利要求1所述的方法,其特征在于,所述当前输入信息包括当前文本输入信息、当前语音输入信息和当前图像输入信息中至少两种。2 . The method according to claim 1 , wherein the current input information comprises at least two of current text input information, current voice input information and current image input information. 3 . 3.根据权利要求2所述的方法,其特征在于,所述当前图像输入信息包括当前面部图像和/或当前身体姿态图像。3. The method according to claim 2, wherein the current image input information comprises a current facial image and/or a current body posture image. 4.根据权利要求1至3中任一项所述的方法,其特征在于,所述情绪因素包括性格类型和/或历史情绪类型。4. The method according to any one of claims 1 to 3, wherein the emotional factors include personality type and/or historical emotional type. 5.根据权利要求4所述的方法,其特征在于,所述用户的性格类型是基于所述用户的预设时间段输入信息和性格识别模型获得的,其中,所述性格识别模型基于预设时间段输入信息样本对应的性格特征和性格类型标签,预先训练第一预设识别网络获得。5. The method according to claim 4, wherein the character type of the user is obtained based on the user's preset time period input information and a character recognition model, wherein the character recognition model is based on a preset The character characteristics and character type labels corresponding to the input information samples in the time period are obtained by pre-training the first preset recognition network. 6.根据权利要求5所述的方法,其特征在于,所述性格识别模型的训练步骤包括:6. The method according to claim 5, wherein the training step of the character recognition model comprises: 将所述预设时间段输入信息样本对应的性格特征,按照时间序列输入所述第一预设识别网络,获得预测性格类型;Inputting the character characteristics corresponding to the input information samples in the preset time period into the first preset recognition network according to the time series to obtain the predicted character type; 基于所述预测性格类型、所述性格类型标签和第一预设损失函数,调整所述第一预设识别网络的参数;adjusting parameters of the first preset recognition network based on the predicted personality type, the personality type label and the first preset loss function; 将训练完成的所述第一预设识别网络作为所述性格识别模型。The trained first preset recognition network is used as the character recognition model. 7.根据权利要求6所述的方法,其特征在于,所述用户的性格类型的获得步骤,包括:7. The method according to claim 6, wherein the step of obtaining the personality type of the user comprises: 对所述用户预设时间段输入信息进行性格特征提取,获得所述预设时间段输入信息对应的性格特征;performing character feature extraction on the input information of the preset time period of the user to obtain the character feature corresponding to the input information of the preset time period; 确定所述预设时间段输入信息对应的性格特征的时间序列;determining the time series of the character characteristics corresponding to the input information in the preset time period; 按照所述时间序列将所述预设时间段输入信息对应的性格特征输入所述性格识别模型,获得所述用户的性格类型。Inputting the character characteristics corresponding to the input information of the preset time period into the character recognition model according to the time series, to obtain the character type of the user. 8.一种识别用户情绪的装置,其特征在于,包括:8. A device for identifying user emotions, comprising: 第一获得单元,用于获得用户的当前输入信息和情绪因素;a first obtaining unit, used to obtain the current input information and emotional factors of the user; 第二获得单元,用于对所述当前输入信息进行情绪特征提取,获得所述当前输入信息对应的当前情绪特征;a second obtaining unit, configured to perform emotional feature extraction on the current input information to obtain the current emotional feature corresponding to the current input information; 第三获得单元,用于基于所述当前情绪特征、所述情绪因素和情绪识别模型,获得所述用户的当前情绪类型。The third obtaining unit is configured to obtain the current emotion type of the user based on the current emotion feature, the emotion factor and the emotion recognition model. 9.一种用于识别用户情绪的装置,其特征在于,包括有存储器,以及一个或者一个以上的程序,其中一个或者一个以上程序存储于存储器中,且经配置以由一个或者一个以上处理器执行所述一个或者一个以上程序包含用于进行以下操作的指令:9. A device for recognizing user emotions, comprising a memory and one or more programs, wherein one or more programs are stored in the memory and configured to be executed by one or more processors Executing the one or more programs includes instructions for: 获得用户的当前输入信息和情绪因素;Obtain the user's current input information and emotional factors; 对所述当前输入信息进行情绪特征提取,获得所述当前输入信息对应的当前情绪特征;Perform emotional feature extraction on the current input information to obtain the current emotional feature corresponding to the current input information; 基于所述当前情绪特征、所述情绪因素和情绪识别模型,获得所述用户的当前情绪类型。Based on the current emotion feature, the emotion factor and the emotion recognition model, the current emotion type of the user is obtained. 10.一种机器可读介质,其上存储有指令,当由一个或多个处理器执行时,使得装置执行如权利要求1至7中任一项所述的识别用户情绪的方法。10. A machine-readable medium having stored thereon instructions which, when executed by one or more processors, cause an apparatus to perform the method of identifying a user's emotion as claimed in any one of claims 1 to 7.
CN202110128395.3A 2021-01-29 2021-01-29 Method and related device for identifying emotion of user Active CN112818841B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110128395.3A CN112818841B (en) 2021-01-29 2021-01-29 Method and related device for identifying emotion of user

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110128395.3A CN112818841B (en) 2021-01-29 2021-01-29 Method and related device for identifying emotion of user

Publications (2)

Publication Number Publication Date
CN112818841A true CN112818841A (en) 2021-05-18
CN112818841B CN112818841B (en) 2024-10-29

Family

ID=75858388

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110128395.3A Active CN112818841B (en) 2021-01-29 2021-01-29 Method and related device for identifying emotion of user

Country Status (1)

Country Link
CN (1) CN112818841B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113256397A (en) * 2021-07-02 2021-08-13 佛山市墨纳森智能科技有限公司 Commodity recommendation method and system based on big data and computer-readable storage medium
CN114359813A (en) * 2022-01-12 2022-04-15 中国人民大学 Depression emotion detection method and device
CN117354421A (en) * 2023-10-10 2024-01-05 中科融城科技集团有限公司 An intelligent speech analysis method and system

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103024521A (en) * 2012-12-27 2013-04-03 深圳Tcl新技术有限公司 Program screening method, program screening system and television with program screening system
US20180061407A1 (en) * 2016-08-30 2018-03-01 Beijing Baidu Netcom Science And Technology Co., Ltd. Method and apparatus for inputting information
CN108009573A (en) * 2017-11-24 2018-05-08 北京物灵智能科技有限公司 A kind of robot emotion model generating method, mood model and exchange method
CN108805089A (en) * 2018-06-14 2018-11-13 南京云思创智信息科技有限公司 Based on multi-modal Emotion identification method
CN110378428A (en) * 2019-07-23 2019-10-25 上海思依暄机器人科技股份有限公司 A kind of domestic robot and its Emotion identification method and apparatus
CN110570844A (en) * 2019-08-15 2019-12-13 平安科技(深圳)有限公司 Speech emotion recognition method and device and computer readable storage medium
CN110751943A (en) * 2019-11-07 2020-02-04 浙江同花顺智能科技有限公司 Voice emotion recognition method and device and related equipment
CN110858099A (en) * 2018-08-20 2020-03-03 北京搜狗科技发展有限公司 Candidate word generation method and device
CN111241822A (en) * 2020-01-03 2020-06-05 北京搜狗科技发展有限公司 Emotion discovery and dispersion method and device under input scene
CN111353366A (en) * 2019-08-19 2020-06-30 深圳市鸿合创新信息技术有限责任公司 Emotion detection method and device and electronic equipment
JP2020113217A (en) * 2019-01-16 2020-07-27 株式会社SR factory Emotion assessment terminal, emotion assessment program, and emotion assessment method
CN111554304A (en) * 2020-04-25 2020-08-18 中信银行股份有限公司 User tag obtaining method, device and equipment
CN111803095A (en) * 2020-07-07 2020-10-23 天津科技大学 A personal emotional state evaluation method based on ARM chip embedded device
CN112101823A (en) * 2020-11-03 2020-12-18 四川大汇大数据服务有限公司 Multidimensional emotion recognition management method, system, processor, terminal and medium

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103024521A (en) * 2012-12-27 2013-04-03 深圳Tcl新技术有限公司 Program screening method, program screening system and television with program screening system
US20180061407A1 (en) * 2016-08-30 2018-03-01 Beijing Baidu Netcom Science And Technology Co., Ltd. Method and apparatus for inputting information
CN108009573A (en) * 2017-11-24 2018-05-08 北京物灵智能科技有限公司 A kind of robot emotion model generating method, mood model and exchange method
CN108805089A (en) * 2018-06-14 2018-11-13 南京云思创智信息科技有限公司 Based on multi-modal Emotion identification method
CN110858099A (en) * 2018-08-20 2020-03-03 北京搜狗科技发展有限公司 Candidate word generation method and device
JP2020113217A (en) * 2019-01-16 2020-07-27 株式会社SR factory Emotion assessment terminal, emotion assessment program, and emotion assessment method
CN110378428A (en) * 2019-07-23 2019-10-25 上海思依暄机器人科技股份有限公司 A kind of domestic robot and its Emotion identification method and apparatus
CN110570844A (en) * 2019-08-15 2019-12-13 平安科技(深圳)有限公司 Speech emotion recognition method and device and computer readable storage medium
CN111353366A (en) * 2019-08-19 2020-06-30 深圳市鸿合创新信息技术有限责任公司 Emotion detection method and device and electronic equipment
CN110751943A (en) * 2019-11-07 2020-02-04 浙江同花顺智能科技有限公司 Voice emotion recognition method and device and related equipment
CN111241822A (en) * 2020-01-03 2020-06-05 北京搜狗科技发展有限公司 Emotion discovery and dispersion method and device under input scene
CN111554304A (en) * 2020-04-25 2020-08-18 中信银行股份有限公司 User tag obtaining method, device and equipment
CN111803095A (en) * 2020-07-07 2020-10-23 天津科技大学 A personal emotional state evaluation method based on ARM chip embedded device
CN112101823A (en) * 2020-11-03 2020-12-18 四川大汇大数据服务有限公司 Multidimensional emotion recognition management method, system, processor, terminal and medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
曹晓春;: "基于文字输入特征的情绪自动识别", 现代计算机(专业版), no. 04, 5 February 2017 (2017-02-05) *
李芮;王万森;: "融入性格的E-Learning情绪模型", 计算机工程与设计, no. 01, 16 January 2016 (2016-01-16) *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113256397A (en) * 2021-07-02 2021-08-13 佛山市墨纳森智能科技有限公司 Commodity recommendation method and system based on big data and computer-readable storage medium
CN114359813A (en) * 2022-01-12 2022-04-15 中国人民大学 Depression emotion detection method and device
CN117354421A (en) * 2023-10-10 2024-01-05 中科融城科技集团有限公司 An intelligent speech analysis method and system

Also Published As

Publication number Publication date
CN112818841B (en) 2024-10-29

Similar Documents

Publication Publication Date Title
CN108363706B (en) Method and device for human-computer dialogue interaction, device for human-computer dialogue interaction
CN107705783B (en) Voice synthesis method and device
CN109599128B (en) Speech emotion recognition method and device, electronic equipment and readable medium
CN110598504B (en) Image recognition method and device, electronic equipment and storage medium
CN107644646B (en) Voice processing method and device for voice processing
CN109961791B (en) Voice information processing method and device and electronic equipment
CN106024009A (en) Audio processing method and device
CN109360197B (en) Image processing method and device, electronic equipment and storage medium
CN108038102B (en) Recommended method, device, terminal and storage medium for facial expression images
CN107945806B (en) User identification method and device based on sound characteristics
CN109919829A (en) Image Style Transfer method, apparatus and computer readable storage medium
US20210089726A1 (en) Data processing method, device and apparatus for data processing
CN108648754A (en) Sound control method and device
CN112651235A (en) Poetry generation method and related device
CN113392898B (en) A training image classification model, image search method and related device
CN110991329A (en) Semantic analysis method and device, electronic equipment and storage medium
CN112818841A (en) Method and related device for recognizing user emotion
CN111242303A (en) Network training method and device, and image processing method and device
CN111210844B (en) Method, device and equipment for determining speech emotion recognition model and storage medium
CN109388699A (en) Input method, device, equipment and storage medium
CN110990534A (en) Data processing method and device and data processing device
CN112445906A (en) Method and device for generating reply message
CN113923517A (en) Background music generation method and device and electronic equipment
CN112784151A (en) Method and related device for determining recommendation information
CN112446265B (en) Input method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant