CN115774817A - Information processing model training method, information processing method and related equipment - Google Patents
Information processing model training method, information processing method and related equipment Download PDFInfo
- Publication number
- CN115774817A CN115774817A CN202111043012.9A CN202111043012A CN115774817A CN 115774817 A CN115774817 A CN 115774817A CN 202111043012 A CN202111043012 A CN 202111043012A CN 115774817 A CN115774817 A CN 115774817A
- Authority
- CN
- China
- Prior art keywords
- user
- information
- model
- neural network
- target
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Landscapes
- Machine Translation (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
本申请涉及人工智能领域,具体公开了一种信息处理模型的训练方法、信息处理方法及相关设备,该训练方法包括:获取第一用户的交互信息序列和第二用户的交互信息序列;根据第一用户的第一用户特征生成第一用户的局部特征和第一用户的全局特征;以及根据第二用户的第一用户特征生成第二用户的局部特征;根据第一用户的局部特征和第一用户的全局特征计算第一得分;以及根据第一用户的全局特征和第二用户的局部特征计算第二得分;根据第一得分和第二得分对第一神经网络模型、用户特征模型和判别模型进行训练;根据训练后的第一神经网络模型和训练后的用户特征模型确定信息处理模型。本方案可以有效保证在目标域中所确定目标信息与用户之间的匹配性。
This application relates to the field of artificial intelligence, and specifically discloses an information processing model training method, information processing method and related equipment. The training method includes: obtaining the interaction information sequence of the first user and the interaction information sequence of the second user; according to the A user's first user feature generates the first user's local feature and the first user's global feature; and generates the second user's local feature according to the second user's first user feature; according to the first user's local feature and the first Calculating the first score based on the global features of the user; and calculating the second score based on the global features of the first user and the local features of the second user; pairing the first neural network model, the user feature model and the discriminant model according to the first score and the second score Carry out training; determine an information processing model according to the trained first neural network model and the trained user feature model. This scheme can effectively guarantee the matching between the target information determined in the target domain and the user.
Description
技术领域technical field
本申请涉及人工智能技术领域,更具体地,涉及一种信息处理模型的训练方法、信息处理方法及相关设备。The present application relates to the technical field of artificial intelligence, and more specifically, to an information processing model training method, an information processing method and related equipment.
背景技术Background technique
随着互联网技术的发展,越来越多的用户通过发送平台进行信息分享和传播。当发送平台接收到用户的信息请求后,该发送平台从信息数据库中筛选向用户发送的信息。相关技术中,发送平台向用户所发送的信息与用户的匹配性不高,导致所发送的信息与用户实际需要的信息差异较大,进而需要用户与发送平台进行多次交互,发送平台多次进行信息筛选和信息发送,用户才能获取到所需要的信息。由于需要信息发送平台多次进行信息筛选和信息发送,导致信息发送效率不高。With the development of Internet technology, more and more users share and disseminate information through the sending platform. After the sending platform receives the user's information request, the sending platform screens the information sent to the user from the information database. In related technologies, the information sent by the sending platform to the user is not highly compatible with the user, resulting in a large difference between the sent information and the information actually needed by the user, which in turn requires the user to interact with the sending platform multiple times, and the sending platform multiple times. Only by filtering and sending information can users obtain the information they need. Since the information sending platform needs to perform information screening and information sending many times, the efficiency of information sending is not high.
发明内容Contents of the invention
鉴于上述问题,本申请实施例提出了一种信息处理模型的训练方法、信息处理方法及相关设备,以解决因信息与用户的匹配性不高所导致信息发送效率不高的问题。In view of the above problems, the embodiment of the present application proposes an information processing model training method, an information processing method, and related equipment, so as to solve the problem of low information transmission efficiency caused by the low matching between information and users.
根据本申请实施例的一个方面,提供了一种信息处理模型的训练方法,包括:从源域的第一用户行为数据中获取第一用户的交互信息序列和第二用户的交互信息序列;由第一神经网络模型根据所述第一用户的交互信息序列输出所述第一用户的第一用户特征,以及由所述第一神经网络模型根据所述第二用户的交互信息序列输出所述第二用户的第一用户特征;所述第一神经网络模型是根据源域的第二用户历史行为数据进行预训练确定的;由用户特征模型根据所述第一用户的第一用户特征生成所述第一用户的局部特征和第一用户的全局特征;以及由所述用户特征模型根据所述第二用户的第一用户特征生成所述第二用户的局部特征;由判别模型根据所述第一用户的局部特征和所述第一用户的全局特征计算第一得分;以及由所述判别模型根据所述第一用户的全局特征和所述第二用户的局部特征计算第二得分;根据所述第一得分和所述第二得分对所述第一神经网络模型、所述用户特征模型和所述判别模型进行训练,直至达到训练结束条件;根据训练后的所述第一神经网络模型和训练后的所述用户特征模型确定信息处理模型,所述信息处理模型在目标域中确定向用户待发送的目标信息。According to an aspect of an embodiment of the present application, there is provided a method for training an information processing model, including: acquiring the interaction information sequence of the first user and the interaction information sequence of the second user from the first user behavior data in the source domain; The first neural network model outputs the first user features of the first user according to the interaction information sequence of the first user, and the first neural network model outputs the first user feature according to the interaction information sequence of the second user. The first user feature of the second user; the first neural network model is determined according to the pre-training of the second user historical behavior data in the source domain; the user feature model generates the first user feature according to the first user feature The local features of the first user and the global features of the first user; and the local features of the second user are generated by the user feature model according to the first user features of the second user; calculating a first score based on the local features of the user and the global features of the first user; and calculating a second score based on the global features of the first user and the local features of the second user by the discriminant model; according to the The first score and the second score train the first neural network model, the user feature model and the discriminant model until the training end condition is reached; according to the trained first neural network model and the training The latter user characteristic model determines an information processing model, and the information processing model determines the target information to be sent to the user in the target domain.
根据本申请实施例的一个方面,提供了一种信息处理方法,包括:获取目标域中的候选信息集合;通过所述信息处理模型预测目标用户针对所述候选信息集合中各信息的行为标签;所述信息处理模型是按照如上所述的信息处理模型的训练方法训练得到的;根据所述行为标签,在所述候选信息集合中确定向所述目标用户发送的目标信息。According to an aspect of an embodiment of the present application, an information processing method is provided, including: acquiring a set of candidate information in a target domain; predicting a target user's behavior label for each information in the set of candidate information through the information processing model; The information processing model is trained according to the information processing model training method described above; according to the behavior label, the target information to be sent to the target user is determined in the candidate information set.
根据本申请实施例的一个方面,提供了一种信息处理模型的训练装置,包括:第一用户行为数据获取模块,用于从源域的第一用户行为数据中获取第一用户的交互信息序列和第二用户的交互信息序列;第一输出模块,用于由第一神经网络模型根据所述第一用户的交互信息序列输出所述第一用户的第一用户特征,以及第二输出模块,用于由所述第一神经网络模型根据所述第二用户的交互信息序列输出所述第二用户的第一用户特征;所述第一神经网络模型是根据源域的第二用户历史行为数据进行预训练确定的;第一生成模块,用于由用户特征模型根据所述第一用户的第一用户特征生成所述第一用户的局部特征和第一用户的全局特征;以及第二生成模块,用于由所述用户特征模型根据所述第二用户的第一用户特征生成所述第二用户的局部特征;第一得分计算模块,用于由判别模型根据所述第一用户的局部特征和所述第一用户的全局特征计算第一得分;以及第二得分计算模块,用于由所述判别模型根据所述第一用户的全局特征和所述第二用户的局部特征计算第二得分;第一训练模块,用于根据所述第一得分和所述第二得分对所述第一神经网络模型、所述用户特征模型和所述判别模型进行训练,直至达到训练结束条件;信息处理模型确定模块,用于根据训练后的所述第一神经网络模型和训练后的所述用户特征模型确定信息处理模型,所述信息处理模型在目标域中确定向用户待发送的目标信息。According to an aspect of an embodiment of the present application, an information processing model training device is provided, including: a first user behavior data acquisition module, configured to acquire the interaction information sequence of the first user from the first user behavior data in the source domain and the interaction information sequence of the second user; the first output module is used to output the first user characteristics of the first user according to the interaction information sequence of the first user by the first neural network model, and the second output module, The first neural network model is used to output the first user characteristics of the second user according to the interaction information sequence of the second user; the first neural network model is based on the historical behavior data of the second user in the source domain Determined by pre-training; the first generation module is used to generate the local characteristics of the first user and the global characteristics of the first user according to the first user characteristics of the first user by the user characteristic model; and the second generation module , used to generate the local features of the second user according to the first user features of the second user by the user feature model; the first score calculation module is used to use the discriminant model according to the local features of the first user calculating a first score with the global features of the first user; and a second score calculation module, configured to calculate a second score by the discriminant model according to the global features of the first user and the local features of the second user ; The first training module is used to train the first neural network model, the user feature model and the discriminant model according to the first score and the second score until the training end condition is reached; information processing A model determination module, configured to determine an information processing model according to the trained first neural network model and the trained user characteristic model, and the information processing model determines the target information to be sent to the user in the target domain.
在本申请的一些实施例中,所述用户特征模型包括至少两个卷积层和池化层,所述至少两个卷积层中每个卷积层所对应的卷积窗口的长度不同;在本实施例中,第一生成模块,包括:第一卷积处理单元,用于由所述至少两个卷积层中每个卷积层按照所对应的卷积窗口分别对所述第一用户的第一用户特征进行卷积处理,得到所述第一用户的多个中间局部特征;第一拼接单元,用于将所述第一用户的多个中间局部特征进行拼接,得到所述第一用户的局部特征;第一池化处理单元,用于由所述池化层对所述第一用户的局部特征进行池化处理,得到所述第一用户的全局特征;在本实施例中,第二生成模块,包括:第二卷积处理单元,用于由所述至少两个卷积层中每个卷积层按照所对应的卷积窗口分别对所述第二用户的第一用户特征进行卷积处理,得到所述第二用户的多个中间局部特征;第二拼接单元,用于将所述第二用户的多个中间局部特征进行拼接,得到所述第二用户的局部特征。In some embodiments of the present application, the user feature model includes at least two convolutional layers and a pooling layer, and the length of the convolution window corresponding to each convolutional layer in the at least two convolutional layers is different; In this embodiment, the first generating module includes: a first convolution processing unit, configured to perform each convolutional layer of the at least two convolutional layers on the first performing convolution processing on the first user features of the user to obtain multiple intermediate local features of the first user; the first concatenation unit is configured to concatenate the multiple intermediate local features of the first user to obtain the first user A local feature of a user; a first pooling processing unit, configured to perform pooling processing on the local feature of the first user by the pooling layer to obtain a global feature of the first user; in this embodiment , the second generating module, comprising: a second convolution processing unit, configured to respectively perform the first user of the second user by each convolution layer in the at least two convolution layers according to the corresponding convolution window performing convolution processing on features to obtain a plurality of intermediate local features of the second user; a second splicing unit configured to splice the multiple intermediate local features of the second user to obtain local features of the second user .
在本申请的一些实施例中,所述第一神经网络模型包括基础神经网络模型和插入所述基础神经网络模型的自适应网络,所述基础神经网络模型的参数是根据所述源域的第二用户历史行为数据进行预训练确定的;该信息处理模型的训练装置还包括:参数固定模块,用于固定所述基础神经网络模型的参数;在本实施例中,第一训练模块进一步被配置为:根据所述第一得分和所述第二得分,调整所述自适应网络、所述用户特征模型和所述判别模型的参数并继续训练,直至达到训练结束条件。In some embodiments of the present application, the first neural network model includes a basic neural network model and an adaptive network inserted into the basic neural network model, and the parameters of the basic neural network model are based on the first neural network model of the source domain. Two user historical behavior data are determined by pre-training; the training device of the information processing model also includes: a parameter fixing module, which is used to fix the parameters of the basic neural network model; in this embodiment, the first training module is further configured It is: according to the first score and the second score, adjust the parameters of the adaptive network, the user feature model and the discriminant model and continue training until the training end condition is reached.
在本申请的一些实施例中,所述基础神经网络模型包括嵌入层和多个级联的转换器神经网络,所述第一神经网络模型被划分为N个级联的子神经网络,所述子神经网络包括一转换器神经网络和插入所述转换器神经网络中的至少一自适应网络;N为大于1的整数;第一输出模块,包括:输入信息获取单元,用于获取k级子神经网络对应的输入信息;其中,若k=1,所述k级子神经网络对应的输入信息是所述嵌入层根据所述第一用户的交互信息序列输出的嵌入向量;若k>1,所述k级子神经网络对应的输入信息为k-1级子神经网络输出第一用户的k-1级中间用户特征;0<k≤N,k为整数;第一输入单元,用于将所述k级子神经网络对应的输入信息输入所述k级子神经网络;特征提取单元,用于由所述k级子神经网络根据所对应的输入信息进行特征提取,输出第一用户的k级中间用户特征;其中,若k=N,将第一用户的k级中间用户特征作为第一用户的第一用户特征;若k<N,将所述第一用户的k级中间用户特征作为k+1级子神经网络的输入信息。In some embodiments of the present application, the basic neural network model includes an embedding layer and multiple cascaded converter neural networks, the first neural network model is divided into N cascaded sub-neural networks, the The sub-neural network includes a converter neural network and at least one adaptive network inserted into the converter neural network; N is an integer greater than 1; the first output module includes: an input information acquisition unit for obtaining k-level sub-neural networks; The input information corresponding to the neural network; wherein, if k=1, the input information corresponding to the k-level sub-neural network is the embedding vector output by the embedding layer according to the interaction information sequence of the first user; if k>1, The input information corresponding to the k-level sub-neural network is that the k-1-level sub-neural network outputs the k-1 level intermediate user characteristics of the first user; 0<k≤N, k is an integer; the first input unit is used to The input information corresponding to the k-level sub-neural network is input into the k-level sub-neural network; the feature extraction unit is used to perform feature extraction by the k-level sub-neural network according to the corresponding input information, and output the first user's k wherein, if k=N, the k-level intermediate user features of the first user are used as the first user features of the first user; if k<N, the k-level intermediate user features of the first user are used as The input information of the k+1-level sub-neural network.
在本申请的一些实施例中,所述转换器神经网络包括级联的多头注意力层、第一前馈神经网络层、第一求和与归一化层、第二前馈神经网络层、和第二求和与归一化层;每一所述子神经网络包括插入所述第一前馈神经网络层与所述第一求和与归一化层之间的自适应网络,和插入所述第二前馈神经网络层与所述第二求和与归一化层之间的自适应网络;所述子神经网络还包括从所述多头注意力层的输入指向所述第一求和与归一化层的输入的恒等映射,和从所述第二前馈神经网络层的输入指向所述第二求和与归一化层的输入的恒等映射。In some embodiments of the present application, the converter neural network includes a cascaded multi-head attention layer, a first feedforward neural network layer, a first summation and normalization layer, a second feedforward neural network layer, and a second summation and normalization layer; each of said subneural networks includes an adaptive network inserted between said first feedforward neural network layer and said first summation and normalization layer, and inserting An adaptive network between the second feed-forward neural network layer and the second summation and normalization layer; and an identity map of inputs to the sum and normalization layer, and an identity map from the input of the second feedforward neural network layer to the input of the second sum and normalization layer.
在本申请的一些实施例中,所述自适应网络包括级联的首层全连接层、中间全连接层、激活层和末层全连接层,所述首层全连接层中神经元的数量与所述末层全连接层中神经元的数量相等,所述首层全连接层中神经元的数量大于所述中间全连接层中神经元的数量;所述自适应网络还包括由所述首层全连接层的输入指向所述末层全连接层的输出的恒等映射。In some embodiments of the present application, the adaptive network includes a cascaded first fully connected layer, an intermediate fully connected layer, an activation layer and a final fully connected layer, and the number of neurons in the first fully connected layer Equal to the number of neurons in the fully connected layer of the last layer, the number of neurons in the fully connected layer of the first layer is greater than the number of neurons in the middle fully connected layer; the adaptive network also includes the An identity mapping from the input of the first fully connected layer to the output of the last fully connected layer.
在本申请的一些实施例中,信息处理模型的训练装置还包括:目标节点选取模块,用于在参考交互信息序列中选取多个目标节点;所述参考交互信息序列是指所述第二用户历史行为数据中各用户的交互信息序列;替换模块,用于将所述参考交互信息序列中的所述多个目标节点进行遮挡,得到样本交互信息序列;样本交互信息序列输入模块,用于将所述样本交互信息序列输入所述基础神经网络模型;预测信息输出模块,用于由所述基础神经网络模型根据所述样本交互信息序列,输出对应于所述目标节点的预测信息;预测误差确定模块,用于根据所述预测信息和所述目标节点,确定预测误差;第二训练模块,用于根据所述预测误差调整所述基础神经网络模型的参数并继续训练,直至达到预训练结束条件。In some embodiments of the present application, the information processing model training device further includes: a target node selection module, configured to select multiple target nodes in the reference interaction information sequence; the reference interaction information sequence refers to the second user The interaction information sequence of each user in the historical behavior data; the replacement module is used to block the multiple target nodes in the reference interaction information sequence to obtain a sample interaction information sequence; the sample interaction information sequence input module is used to convert The sample interaction information sequence is input into the basic neural network model; the prediction information output module is used for the basic neural network model to output the prediction information corresponding to the target node according to the sample interaction information sequence; the prediction error is determined A module for determining a prediction error according to the prediction information and the target node; a second training module for adjusting the parameters of the basic neural network model according to the prediction error and continuing training until the pre-training end condition is reached .
根据本申请实施例的一个方面,提供了一种信息处理装置,包括:候选信息集合获取模块,用于获取目标域中的候选信息集合;行为标签预测模块,用于通过所述信息处理模型预测目标用户针对所述候选信息集合中各信息的行为标签;所述信息处理模型是按照如上所述的信息处理模型的训练方法训练得到的;目标信息确定模块,用于根据所述行为标签,在所述候选信息集合中确定向所述目标用户待发送的目标信息。According to an aspect of the embodiment of the present application, an information processing device is provided, including: a candidate information set acquisition module, used to acquire a candidate information set in the target domain; a behavior label prediction module, used to predict The target user's behavior label for each information in the candidate information set; the information processing model is obtained by training according to the above-mentioned information processing model training method; the target information determination module is used to, according to the behavior label, in Target information to be sent to the target user is determined in the candidate information set.
在本申请的一些实施例中,所述信息处理模型包括训练后的第一神经网络模型、训练后的所述用户特征模型、嵌入查找层和分类层;行为标签预测模块,包括:全局特征查找单元,用于在用户全局特征集合中获取所述目标用户的全局特征;所述用户全局特征集合中的全局特征是由训练后的所述第一神经网络模型和训练后的所述用户特征模型根据所述第一用户行为数据生成的;嵌入向量生成单元,用于由所述嵌入查找层生成所述候选信息集合中各信息的嵌入向量;行为标签输出单元,用于由所述分类层根据所述目标用户的全局特征和所述候选信息集合中各信息的嵌入向量,输出所述目标用户针对所述候选信息集合中各信息的行为标签。In some embodiments of the present application, the information processing model includes a trained first neural network model, a trained user feature model, an embedding search layer, and a classification layer; a behavior label prediction module includes: global feature search A unit, configured to obtain the global features of the target user in the user global feature set; the global feature in the user global feature set is composed of the trained first neural network model and the trained user feature model Generated according to the first user behavior data; the embedding vector generation unit is used to generate the embedding vector of each information in the candidate information set by the embedding search layer; the behavior label output unit is used to be used by the classification layer according to The global features of the target user and the embedding vectors of each information in the candidate information set are used to output the target user's behavior label for each information in the candidate information set.
在本申请的一些实施例中,信息处理装置还包括:接收模块,用于接收信息请求,所述信息请求指示了所述目标用户的用户标识;发送模块,用于根据所述目标用户的用户标识,向所述目标用户发送所述目标信息。In some embodiments of the present application, the information processing device further includes: a receiving module, configured to receive an information request, the information request indicating the user identifier of the target user; a sending module, configured to identify, and send the target information to the target user.
在本申请的一些实施例中,信息处理装置还包括:交互行为信息获取模块,用于获取所述目标用户针对所述目标信息的交互行为信息;更新训练模块,用于根据所述目标用户针对所述目标信息的交互行为信息,对所述信息处理模型进行更新训练。In some embodiments of the present application, the information processing device further includes: an interactive behavior information acquisition module, configured to acquire the target user's interactive behavior information for the target information; an update training module, configured to The interactive behavior information of the target information is used to update and train the information processing model.
根据本申请实施例的一个方面,提供了一种电子设备,包括:处理器;存储器,所述存储器上存储有计算机可读指令,所述计算机可读指令被所述处理器执行时,实现如上所述的信息处理模型的训练方法或者信息处理方法。According to an aspect of an embodiment of the present application, there is provided an electronic device, including: a processor; a memory, on which computer-readable instructions are stored, and when the computer-readable instructions are executed by the processor, the above The training method of the information processing model or the information processing method.
根据本申请实施例的一个方面,提供了一种计算机可读存储介质,其上存储有计算机可读指令,当所述计算机可读指令被处理器执行时,实现如上所述的信息处理模型的训练方法或者信息处理方法。According to an aspect of the embodiments of the present application, there is provided a computer-readable storage medium on which computer-readable instructions are stored. When the computer-readable instructions are executed by a processor, the information processing model described above is implemented. training methods or information processing methods.
根据本申请实施例的一个方面,提供了一种计算机程序产品,包括计算机指令,所述计算机指令被处理器执行时实现如上所述的信息处理模型的训练方法或者信息处理方法。According to an aspect of the embodiments of the present application, a computer program product is provided, including computer instructions, and when the computer instructions are executed by a processor, the above information processing model training method or information processing method is implemented.
在本申请的方案中,在通过源域的第二用户行为数据对第一神经网络模型进行预训练,以使该第一神经网络模型根据用户在源域中的交互信息序列学习用户的用户特征,在此基础上,再结合判别模型,利用源域的第一用户行为数据对第一神经网络模型、用户特征模型进行无监督对比学习,拉开不同用户的特征之间的区别,提高基于交互信息序列所提取到用户特征的准确性。从而,根据训练后的第一神经网络模型和用户特征模型所确定的信息处理模型后,由于该信息处理模型在源域中准确学习到了用户的特征,该用户的特征体现了用户对于信息的偏好,因此,由该信息处理模型根据在源域中所学习到的用户特征来在目标域中确定向用户待发送的目标信息,由于该目标信息的确定过程结合了用户在源域中的用户特征,因此,可以提高目标信息与用户之间的匹配性,从而可以有效减少与用户之间的交互次数,有效解决现有技术中因信息与用户之间匹配性不高所导致信息发送效率不高的问题。In the solution of this application, the first neural network model is pre-trained through the second user behavior data in the source domain, so that the first neural network model can learn the user's user characteristics according to the user's interaction information sequence in the source domain , on this basis, combined with the discriminant model, using the first user behavior data in the source domain to conduct unsupervised comparative learning on the first neural network model and the user feature model, to widen the difference between the features of different users, and improve the interaction-based The accuracy of user features extracted from information sequences. Therefore, after the information processing model determined according to the trained first neural network model and the user feature model, since the information processing model has accurately learned the user's features in the source domain, the user's features reflect the user's preference for information , therefore, the information processing model determines the target information to be sent to the user in the target domain according to the user characteristics learned in the source domain, because the determination process of the target information combines the user characteristics of the user in the source domain , therefore, the matching between the target information and the user can be improved, thereby effectively reducing the number of interactions with the user, and effectively solving the problem of low information sending efficiency caused by the low matching between the information and the user in the prior art The problem.
附图说明Description of drawings
此处的附图被并入说明书中并构成本说明书的一部分,示出了符合本申请的实施例,并与说明书一起用于解释本申请的原理。显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the application and together with the description serve to explain the principles of the application. Apparently, the drawings in the following description are only some embodiments of the present application, and those skilled in the art can obtain other drawings according to these drawings without creative efforts.
图1示出了本申请一个示例性实施例提供的信息发送系统的示意图。Fig. 1 shows a schematic diagram of an information sending system provided by an exemplary embodiment of the present application.
图2是根据本申请的一实施例示出的信息处理模型的训练方法的流程图。Fig. 2 is a flowchart showing a method for training an information processing model according to an embodiment of the present application.
图3是根据本申请一实施例示出在微调训练阶段进行无监督对比学习的示意图。Fig. 3 is a schematic diagram illustrating unsupervised contrastive learning in the fine-tuning training phase according to an embodiment of the present application.
图4是根据本申请的一实施例示出的基础神经网络模型的示意图。Fig. 4 is a schematic diagram of a basic neural network model according to an embodiment of the present application.
图5是根据本申请一实施例示出的子神经网络的示意图。Fig. 5 is a schematic diagram of a sub-neural network according to an embodiment of the present application.
图6是根据本申请一实施例示出的自适应网络的结构示意图。Fig. 6 is a schematic structural diagram of an adaptive network according to an embodiment of the present application.
图7是根据本申请一实施例示出的步骤220之前步骤的流程图。Fig. 7 is a flowchart showing steps before
图8是根据本申请一实施例示出的信息处理方法的流程图。Fig. 8 is a flowchart of an information processing method according to an embodiment of the present application.
图9是根据本申请一实施例示出的步骤820的流程图。Fig. 9 is a flow
图10是根据本申请一实施例示出的信息处理模型预测行为标签的示意图。Fig. 10 is a schematic diagram showing an information processing model predicting behavior labels according to an embodiment of the present application.
图11是根据本申请一实施例示出在用户界面中显示目标信息的示意图。Fig. 11 is a schematic diagram illustrating displaying target information in a user interface according to an embodiment of the present application.
图12是根据一实施例示出的信息处理模型的训练装置的框图。Fig. 12 is a block diagram of an information processing model training device according to an embodiment.
图13是根据本申请一实施例示出的信息处理装置的框图。Fig. 13 is a block diagram of an information processing device according to an embodiment of the present application.
图14示出了适于用来实现本申请实施例的电子设备的计算机系统的结构示意图。Fig. 14 shows a schematic structural diagram of a computer system suitable for implementing the electronic device of the embodiment of the present application.
具体实施方式Detailed ways
现在将参考附图更全面地描述示例实施方式。然而,示例实施方式能够以多种形式实施,且不应被理解为限于在此阐述的范例;相反,提供这些实施方式使得本申请将更加全面和完整,并将示例实施方式的构思全面地传达给本领域的技术人员。Example embodiments will now be described more fully with reference to the accompanying drawings. Example embodiments may, however, be embodied in many forms and should not be construed as limited to the examples set forth herein; rather, these embodiments are provided so that this application will be thorough and complete, and will fully convey the concepts of example embodiments to those skilled in the art.
此外,所描述的特征、结构或特性可以以任何合适的方式结合在一个或更多实施例中。在下面的描述中,提供许多具体细节从而给出对本申请的实施例的充分理解。然而,本领域技术人员将意识到,可以实践本申请的技术方案而没有特定细节中的一个或更多,或者可以采用其它的方法、组元、装置、步骤等。在其它情况下,不详细示出或描述公知方法、装置、实现或者操作以避免模糊本申请的各方面。Furthermore, the described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided in order to give a thorough understanding of the embodiments of the application. However, those skilled in the art will appreciate that the technical solutions of the present application may be practiced without one or more of the specific details, or other methods, components, devices, steps, etc. may be employed. In other instances, well-known methods, apparatus, implementations, or operations have not been shown or described in detail to avoid obscuring aspects of the application.
附图中所示的方框图仅仅是功能实体,不一定必须与物理上独立的实体相对应。即,可以采用软件形式来实现这些功能实体,或在一个或多个硬件模块或集成电路中实现这些功能实体,或在不同网络和/或处理器装置和/或微控制器装置中实现这些功能实体。The block diagrams shown in the drawings are merely functional entities and do not necessarily correspond to physically separate entities. That is, these functional entities may be implemented in software, or in one or more hardware modules or integrated circuits, or in different networks and/or processor devices and/or microcontroller devices entity.
附图中所示的流程图仅是示例性说明,不是必须包括所有的内容和操作/步骤,也不是必须按所描述的顺序执行。例如,有的操作/步骤还可以分解,而有的操作/步骤可以合并或部分合并,因此实际执行的顺序有可能根据实际情况改变。The flow charts shown in the drawings are only exemplary illustrations, and do not necessarily include all contents and operations/steps, nor must they be performed in the order described. For example, some operations/steps can be decomposed, and some operations/steps can be combined or partly combined, so the actual order of execution may be changed according to the actual situation.
需要说明的是:在本文中提及的“多个”是指两个或两个以上。“和/或”描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。字符“/”一般表示前后关联对象是一种“或”的关系。It should be noted that: the "plurality" mentioned in this article refers to two or more than two. "And/or" describes the association relationship of associated objects, indicating that there may be three types of relationships. For example, A and/or B may indicate: A exists alone, A and B exist simultaneously, and B exists independently. The character "/" generally indicates that the contextual objects are an "or" relationship.
在对本申请进行具体说明之前,针对本申请涉及的术语进行如下解释:Before the application is described in detail, the terms involved in the application are explained as follows:
源域(Source domain):在迁移学习中,源域是指已有知识所属的领域。Source domain: In transfer learning, the source domain refers to the domain to which the existing knowledge belongs.
目标域(Target domain):又称目标领域,是指待学习新知识的领域,其又可以理解为需要处理当前的目标任务所属的领域。Target domain: Also known as the target domain, it refers to the domain where new knowledge is to be learned, which can also be understood as the domain that needs to deal with the current target task.
在信息发送场景下,可以将已有数量足够的用户交互数据的应用所属的领域称为源域。而将没有用户交互数据或者用户交互数据较少的应用所属的领域称为目标域。例如在第一应用程序中,已有大量用户针对在第一应用程序中所发送信息的用户交互数据,则可以将第一应用程序所属的领域称为源域;在第二应用程序中,缺少用户的交互数据,因此,可以将第二应用程序所属的领域称为目标域。In an information sending scenario, the domain to which an application with sufficient user interaction data already exists may be referred to as a source domain. And the domain of the application with no user interaction data or less user interaction data is called the target domain. For example, in the first application program, if there are a large number of user interaction data for the information sent in the first application program, the domain to which the first application program belongs can be called the source domain; in the second application program, the lack of The user's interaction data, therefore, the domain to which the second application program belongs can be referred to as the target domain.
为解决现有技术中因信息与用户之间匹配性不高导致信息发送效率不高的问题问题,提出了本申请的方案,基于人工智能技术来构建和训练信息处理模型,并利用用户针对源域中信息的用户交互信息序列来训练信息处理模型,学习各个用户在源域的用户特征,在此基础上,利用信息处理模型在源域中所学习到的用户特征,来针对性地确定目标域中向用户发送的信息,从而保证在目标域中为用户所确定目标信息与用户之间的匹配性,进而提高信息发送效率。In order to solve the problem of low information sending efficiency due to the low matching between information and users in the prior art, the solution of this application is proposed, which builds and trains an information processing model based on artificial intelligence technology, and utilizes the user to target the source The user interaction information sequence of the information in the domain is used to train the information processing model and learn the user characteristics of each user in the source domain. On this basis, the user characteristics learned by the information processing model in the source domain are used to determine the target in a targeted manner. The information sent to the user in the target domain, so as to ensure the matching between the target information determined for the user in the target domain and the user, thereby improving the efficiency of information transmission.
在一些应用场景下,例如新应用中缺少用户交互数据的场景下(该种场景可以称为冷启动场景,相关用户称为冷用户),也存在向用户发送的信息与用户之间的匹配性不高,导致需要多次向用户发送信息的情况。因此,本申请的方案也可以应用于冷启动场景下确定向用户待发送的目标信息。In some application scenarios, such as the lack of user interaction data in new applications (this scenario can be called a cold start scenario, and the relevant users are called cold users), there is also a matching between the information sent to the user and the user Not high, resulting in the need to send information to the user multiple times. Therefore, the solution of the present application can also be applied to determine the target information to be sent to the user in a cold start scenario.
人工智能(Artificial Intelligence,AI)是利用数字计算机或者数字计算机控制的机器模拟、延伸和扩展人的智能,感知环境、获取知识并使用知识获得最佳结果的理论、方法、技术及应用系统。换句话说,人工智能是计算机科学的一个综合技术,它企图了解智能的实质,并生产出一种新的能以人类智能相似的方式做出反应的智能机器。人工智能也就是研究各种智能机器的设计原理与实现方法,使机器具有感知、推理与决策的功能。Artificial Intelligence (AI) is a theory, method, technology and application system that uses digital computers or machines controlled by digital computers to simulate, extend and expand human intelligence, perceive the environment, acquire knowledge and use knowledge to obtain the best results. In other words, artificial intelligence is a comprehensive technique of computer science that attempts to understand the nature of intelligence and produce a new kind of intelligent machine that can respond in a similar way to human intelligence. Artificial intelligence is to study the design principles and implementation methods of various intelligent machines, so that the machines have the functions of perception, reasoning and decision-making.
人工智能技术是一门综合学科,涉及领域广泛,既有硬件层面的技术也有软件层面的技术。人工智能基础技术一般包括如传感器、专用人工智能芯片、云计算、分布式存储、大数据处理技术、操作/交互系统、机电一体化等技术。人工智能软件技术主要包括计算机视觉技术、语音处理技术、自然语言处理技术以及机器学习/深度学习等几大方向。本申请的方案主要是利用自然语言处理技术来根据用户针对源域中触发用户行为的信息来挖掘用户特征,得到用户的用户特征。并将用户在源域中的用户特征跨领域应用到目标域中,实现跨邻域进行信息推荐。Artificial intelligence technology is a comprehensive subject that involves a wide range of fields, including both hardware-level technology and software-level technology. Artificial intelligence basic technologies generally include technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technology, operation/interaction systems, and mechatronics. Artificial intelligence software technology mainly includes several major directions such as computer vision technology, speech processing technology, natural language processing technology, and machine learning/deep learning. The solution of the present application mainly utilizes natural language processing technology to mine user characteristics according to the user's information triggering user behavior in the source domain, so as to obtain the user characteristics of the user. And the user characteristics of the user in the source domain are applied to the target domain across domains to realize information recommendation across neighborhoods.
在本申请的一些实施例中,目标域可以是与源域中相似的领域或者产品领域。举例来说,若源域和目标域中待推荐的信息均是新闻,目标域和源域可以视为相似的领域;又例如,源域和目标域中待推荐的信息均是博客、公众号文章、视频等聚合信息,目标域和源域可以视为相似的领域。在一些实施例中,若目标域中的信息的类型与源域中信息的类型相同,则目标域可以视为与源域是相似的领域。信息的类型例如文章(博客、新闻、公众号文章等)、小视频、新闻、广告、商品链接等。对应的,源域和目标域中待推荐的信息可以是文章、小视频、新闻、广告等。In some embodiments of the present application, the target domain may be a domain or product domain similar to the source domain. For example, if the information to be recommended in both the source domain and the target domain are news, the target domain and the source domain can be regarded as similar fields; Aggregating information such as articles, videos, etc., the target domain and source domain can be regarded as similar domains. In some embodiments, if the type of information in the target domain is the same as the type of information in the source domain, then the target domain can be regarded as a similar domain to the source domain. Types of information such as articles (blogs, news, public account articles, etc.), small videos, news, advertisements, product links, etc. Correspondingly, the information to be recommended in the source domain and the target domain may be articles, short videos, news, advertisements, and the like.
图1示出了本申请一个示例性实施例提供的信息发送系统的示意图。该信息发送系统包括终端110和发送平台120。Fig. 1 shows a schematic diagram of an information sending system provided by an exemplary embodiment of the present application. The information sending system includes a terminal 110 and a sending
发送平台120可以是独立的物理服务器,也可以是多个物理服务器构成的服务器集群或者分布式系统,还可以是提供云服务、云数据库、云计算、云函数、云存储、网络服务、云通信、中间件服务、域名服务、安全服务、CDN(Content Delivery Network,内容分发网络)、以及大数据和人工智能平台等基础云计算服务的云服务器。The sending
发送平台120用于为支持信息发送的应用程序提供后台服务,可选的,发送平台120承担主要的计算工作,终端110承担次要的计算工作;或者,发送平台120与终端110两者之间采用分布式计算架构进行协同计算。The sending
可选的,发送平台120包括第一发送服务器121和第二发送服务器122。第一发送服务器121用于提供源域中信息发送相关的服务,例如确定向各个用户所要发送的源域中的信息、收集用户针对源域中已发送信息的交互数据等。Optionally, the sending
第二发送服务器122用于提供目标域中信息发送相关的服务,例如确定向各用户所要发送的目标域中的目标信息。在一些实施例中,第二发送服务器122还与第一发送服务器121通信连接。第二发送服务器122上部署了信息处理模型,可以基于第一发送服务器121向用户所发送源域中的信息、以及用户针对源域中已发送信息的交互数据作为信息处理模型的训练数据,来进行信息处理模型训练,以将训练后的信息处理模型用于确定向用户发送的目标域中的目标信息。The
终端110通过无线网络或者有限网络与发送平台120通信连接,终端110可以是智能手机、笔记本电脑、游戏主机、台式计算机、平板电脑、电子书阅读器、智能音箱等电子设备。终端110安装和运行有支持信息发送的应用程序。该应用程序可以是新闻应用程序、即时通讯应用程序、阅读应用程序、基于话题或频道或圈子进行人群聚合的社交类应用程序、基于购物的社交类应用程序、浏览器程序、视频程序中的任意一种。示意性的,终端110是第一用户使用的终端,终端110中运行的应用程序内登录有第一用户帐号。The terminal 110 communicates with the sending
终端110中所运行支持信息发送的应用程序可以是两种应用程序,其中一种应用程序是对应于源域的第一应用程序,另一种应用程序是对应于目标域的第二应用程序。在此基础上,通过第一应用程序可以向用户发送源域中的第一信息;通过第二应用程序可以是向用户推荐目标域中的信息。举例来说,第一应用程序可以是即时通信应用,源域中的信息可以是来自公众号的文章、新闻、博客等;第二应用程序可以是读书应用,目标域中待发送的信息可以是在读书应用中的博客、书评、新闻等。The application programs running on the terminal 110 that support information sending may be two types of application programs, one of which is the first application program corresponding to the source domain, and the other application program is the second application program corresponding to the target domain. On this basis, the first information in the source domain can be sent to the user through the first application program; the information in the target domain can be recommended to the user through the second application program. For example, the first application program can be an instant messaging application, and the information in the source domain can be articles, news, blogs, etc. from official accounts; the second application program can be a reading application, and the information to be sent in the target domain can be Blogs, book reviews, news, etc. in the reading app.
当终端110中可以运行第一应用程序的情况下,当终端110运行第一应用程序的过程中,第一发送服务器121可以向第一应用程序所在用户发送源域中的信息,然后第一应用程序检测用户针对源域中的已发送信息触发的交互行为,并将用户针对源域中已发送信息的交互数据上报到第一发送服务器121,以将所上报用户针对源域中的已发送信息的交互数据作为信息处理模型的训练数据。When the terminal 110 can run the first application program, when the terminal 110 is running the first application program, the first sending
第一发送服务器121可以将众多运行第一应用程序的终端110所上报用户针对源域中的已推荐信息的交互数据,发送到第二发送服务器122,由该第二发送服务器122按照本申请的方法训练信息处理模型,并通过该信息处理模型为第二应用程序中的用户针对性地确定目标域中的目标信息,并向第二应用程序中的用户发送所确定目标域中的目标信息。The
当终端110中还可以运行第二应用程序的情况下,第二发送服务器122可以按照本申请的信息处理方法为第二应用程序所在的用户确定目标域中向该用户待发送的目标信息,并向终端110发送目标域中的目标信息。When the second application program can also be run on the terminal 110, the second sending
可以理解的是,终端110中可以仅运行第一应用程序或者第二应用程序,也可以运行第一应用程序和第二应用程序。图1中所示出信息发送系统中终端的数量还可以更多,该信息发送系统中还可以包括其他终端。It can be understood that, only the first application program or the second application program may be run on the terminal 110, or both the first application program and the second application program may be run. As shown in FIG. 1 , the number of terminals in the information sending system may be more, and the information sending system may also include other terminals.
以下对本申请实施例的技术方案的实现细节进行详细阐述:The implementation details of the technical solutions of the embodiments of the present application are described in detail below:
图2示出了根据本申请的一个实施例示出的信息处理模型的训练方法的流程图,该方法可以由具备处理能力的计算机设备执行,例如服务器等,在此不进行具体限定。参照图2所示,该方法至少包括步骤210至280,详细介绍如下:FIG. 2 shows a flowchart of a method for training an information processing model according to an embodiment of the present application. The method can be executed by a computer device with processing capabilities, such as a server, and is not specifically limited here. Referring to Figure 2, the method at least includes
步骤210,从源域的第一用户行为数据中获取第一用户的交互信息序列和第二用户的交互信息序列。
源域中提供了信息集合,基于该信息集合,可以针对性地向用户进行源域中信息的发送。在向用户发送该信息集合中的信息后,收集用户对源域中已发送信息的交互数据。用户对已推荐信息的交互数据可以通过该用户针对已推荐信息触发的指示对信息感兴趣的交互行为来体现。为便于描述,将指示用户对信息感兴趣的交互行为称为感兴趣用户行为。感兴趣用户行为例如针对已推荐信息触发的点击行为、收藏行为、点赞行为、评论行为等。在一些实施例中,若在用户点击进入已推荐信息的详情页面后,用户才能够针对该已推荐信息触发点赞、收藏或者评论等行为,也可以仅收集反映用户对已推荐信息触发点击行为的交互数据。An information set is provided in the source domain, and based on the information set, the information in the source domain can be sent to the user in a targeted manner. After the information in the information collection is sent to the user, the interaction data of the user on the sent information in the source domain is collected. The user's interaction data on the recommended information may be reflected by the user's interaction behavior indicating interest in the recommended information triggered by the user. For ease of description, an interactive behavior indicating that a user is interested in information is referred to as an interested user behavior. Behaviors of interested users, such as click behaviors, collection behaviors, like behaviors, and comment behaviors triggered by recommended information. In some embodiments, if the user clicks to enter the details page of the recommended information, the user can trigger behaviors such as liking, bookmarking, or commenting on the recommended information. interaction data.
面向大量用户收集多个用户针对源域中的已发送信息的交互数据。根据众多用户针对已发送信息的交互数据,按照时间先后顺序,将同一用户触发交互行为的多个已发送信息进行排列,得到每个用户的交互信息序列。换言之,用户的交互信息序列是指按照时间先后顺序,对用户触发感兴趣用户行为的已发送信息进行排列得到的信息序列。其中,时间先后顺序可以是基于用户对已发送信息触发感兴趣用户行为的时间来确定的。可以理解的是,在用户的交互信息序列中,可以通过已发送信息的信息标识来表示该已发送信息,信息标识例如信息ID等。举例来说,若源域中向用户发送的信息包括A、B、C、D、E、F;其中,用户依次点击的信息包括A、C、E、F,则用户的交互信息序列可以为<A,C,E,F>。For a large number of users, interaction data of multiple users with respect to the sent information in the source domain is collected. According to the interaction data of many users for the sent information, in chronological order, arrange the multiple sent information that triggers the interaction behavior of the same user, and obtain the interaction information sequence of each user. In other words, the user's interaction information sequence refers to the information sequence obtained by arranging the sent information of the user that triggers the user's behavior of interest in chronological order. Wherein, the chronological order may be determined based on the time when the user triggers an interested user behavior on the sent information. It can be understood that, in the user interaction information sequence, the sent information may be represented by an information identifier of the sent information, such as an information ID. For example, if the information sent to the user in the source domain includes A, B, C, D, E, F; among them, the information clicked by the user in turn includes A, C, E, F, then the user's interactive information sequence can be <A, C, E, F>.
进一步的,为了指示交互信息序列所属的领域、该交互信息序列所对应的用户,可以将用户的用户标识、交互信息序列所属领域的领域标识、以及该交互信息序列形成三元数组。从而,源域的第一用户行为数据即为多个三元数组的集合。Further, in order to indicate the field to which the interaction information sequence belongs and the user corresponding to the interaction information sequence, the user identifier of the user, the field identifier of the field to which the interaction information sequence belongs, and the interaction information sequence may form a triplet. Therefore, the first user behavior data in the source domain is a collection of multiple triplets.
源域的第一用户行为数据中包括多个用户在源域中的交互信息序列,在本方案中,由于是通过无监督对比学习的方式来对信息处理模型进行微调训练,因此,每次从源域的第一用户行为数据中获取两个不同的用户在源域中的交互信息序列,为便于区分该两个用户,将其中一个用户称为第一用户,另一个用户成为第二用户。The first user behavior data in the source domain includes the interaction information sequences of multiple users in the source domain. In this scheme, since the information processing model is fine-tuned and trained through unsupervised comparative learning, each time from The interaction information sequences of two different users in the source domain are obtained from the first user behavior data in the source domain. In order to distinguish the two users, one of the users is called the first user, and the other user is called the second user.
步骤220,由第一神经网络模型根据第一用户的交互信息序列输出第一用户的第一用户特征,以及步骤230,由第一神经网络模型根据第二用户的交互信息序列输出第二用户的第一用户特征;第一神经网络模型是根据源域的第二用户历史行为数据进行预训练确定的。In
本申请的方案是通过迁移学习,将源域中学到的知识应用到目标域中的信息发送任务中,因此,借助于源域中的用户交互信息来训练信息处理模型,以使该信息处理模型可以用于确定目标域中待发送的目标信息。The solution of this application is to apply the knowledge learned in the source domain to the information sending task in the target domain through transfer learning. Therefore, the information processing model is trained with the help of user interaction information in the source domain, so that the information processing model Can be used to determine the target information to be sent in the target field.
在本申请的方案中,通过两个训练阶段来训练信息处理模型,该两个训练阶段可以称为预训练(Pre-training)阶段和微调(Fine-tuning)训练阶段,先在预训练阶段进行训练,然后在微调训练阶段进行训练。因此,将源域中的训练数据分成两部分,其中一部分用于预训练阶段,另一部分用于微调训练阶段。在本申请的方案中,将用于预训练阶段的训练数据称为第二用户行为数据,将用于微调训练阶段的训练数据称为第一用户行为数据。In the scheme of this application, the information processing model is trained through two training stages, which can be called the pre-training (Pre-training) stage and the fine-tuning (Fine-tuning) training stage, and the pre-training stage is carried out first training, followed by training in the fine-tuning training phase. Therefore, the training data in the source domain is divided into two parts, one of which is used in the pre-training phase and the other is used in the fine-tuning training phase. In the solution of the present application, the training data used in the pre-training phase is called second user behavior data, and the training data used in the fine-tuning training phase is called first user behavior data.
可以理解的是,即使向不同的用户发送相同的信息,由于不同的用户感兴趣的信息,或者偏好的信息存在差异,因此不同的用户针对相同的信息所触发的互动行为也会存在差异。换言之,用户针对源域中已发送信息触发的互动行为在一定程度上反映的该用户在信息上的偏好特征。基于此,在本申请的方案中,在预训练阶段,通过源域中的第二用户行为数据来训练第一神经网络模型,以使该第一神经网络模型具备根据用户的交互信息序列输出用户的特征表示,为便于描述,将第一神经网络模型所输出的用户的特征表示称为第一用户特征。It is understandable that, even if the same information is sent to different users, different users may have different interactive behaviors triggered by the same information due to differences in the information that different users are interested in or preferred information. In other words, the user's interactive behavior triggered by the information sent in the source domain reflects the user's preference characteristics in information to a certain extent. Based on this, in the solution of this application, in the pre-training stage, the first neural network model is trained through the second user behavior data in the source domain, so that the first neural network model is capable of outputting the user interaction information sequence according to the user The feature representation of the user, for the convenience of description, the feature representation of the user output by the first neural network model is called the first user feature.
第一神经网络模型可以是由神经网络构成的模型。神经网络比如CNN(Convolutional Neural Network,卷积神经网络)、RNN(Recurrent Neural Network,循环神经网络)、LSTM(Long Short-Term Memory Neural Network,长短时记忆神经网络)、全连接神经网络、Transformer模型(转换器神经网络)等。The first neural network model may be a model composed of a neural network. Neural networks such as CNN (Convolutional Neural Network, convolutional neural network), RNN (Recurrent Neural Network, cyclic neural network), LSTM (Long Short-Term Memory Neural Network, long-short-term memory neural network), fully connected neural network, Transformer model (transformer neural network), etc.
在对第一神经网络模型进行预训练后,由于该第一神经网络模型具备根据用户的交互信息序列输出用户的特征表示的能力,因此,该第一神经网络模型可以根据源域中第一用户的交互信息序列对应输出第一用户的第一用户特征,以及可以根据源域中第二用户的交互信息序列输出第二用户的第一用户特征。After pre-training the first neural network model, since the first neural network model has the ability to output the user's feature representation according to the user's interaction information sequence, the first neural network model can be based on the first user in the source domain. The interaction information sequence of correspondingly outputs the first user characteristics of the first user, and can output the first user characteristics of the second user according to the interaction information sequence of the second user in the source domain.
步骤240,由用户特征模型根据第一用户的第一用户特征生成第一用户的局部特征和第一用户的全局特征;以及步骤250,由用户特征模型根据第二用户的第一用户特征生成第二用户的局部特征。
用户特征模型用于基于用户的第一用户特征进行深度特征提取。该用户特征模型可以是通过卷积神经网络、池化神经网络、全连接神经网络等神经网络构建的神经网络模型。The user feature model is used for deep feature extraction based on the first user feature of the user. The user feature model may be a neural network model constructed by a neural network such as a convolutional neural network, a pooled neural network, or a fully connected neural network.
将用户特征模型所输出反映用户局部的特征称为局部特征,将通过用户特征模型所输出反映用户全局的特征称为全局特征。其中,用户的全局特征是基于用户的局部特征进行进一步处理得到的。The features output by the user feature model that reflect the user's local area are called local features, and the features output by the user feature model that reflect the user's overall situation are called global features. Wherein, the global feature of the user is obtained through further processing based on the local feature of the user.
在本申请的一些实施例中,用户特征模型包括至少两个卷积层和池化层,至少两个卷积层中每个卷积层所对应的卷积窗口的长度不同;在该实施例中,该用户特征模型可以通过如下的过程来输出第一用户的局部特征和全局特征:由至少两个卷积层中每个卷积层按照所对应的卷积窗口分别对第一用户的第一用户特征进行卷积处理,得到第一用户的多个中间局部特征;将第一用户的多个中间局部特征进行拼接,得到第一用户的局部特征;由池化层对第一用户的局部特征进行池化处理,得到第一用户的全局特征。In some embodiments of the present application, the user feature model includes at least two convolutional layers and a pooling layer, and the length of the convolution window corresponding to each convolutional layer in the at least two convolutional layers is different; in this embodiment Among them, the user feature model can output the local features and global features of the first user through the following process: Each convolution layer in the at least two convolutional layers is respectively analyzed according to the corresponding convolution window A user feature is subjected to convolution processing to obtain multiple intermediate local features of the first user; multiple intermediate local features of the first user are spliced to obtain the local features of the first user; the local features of the first user are obtained by the pooling layer The features are pooled to obtain the global features of the first user.
同理,该用户特征模型可以通过如下的过程来输出第二用户的局部特征:由至少两个卷积层中每个卷积层按照所对应的卷积窗口分别对第二用户的第一用户特征进行卷积处理,得到第二用户的多个中间局部特征;将第二用户的多个中间局部特征进行拼接,得到第二用户的局部特征。Similarly, the user feature model can output the local features of the second user through the following process: Each convolution layer in the at least two convolutional layers is respectively analyzed by the first user of the second user according to the corresponding convolution window The features are convoluted to obtain multiple intermediate local features of the second user; the multiple intermediate local features of the second user are spliced to obtain the local features of the second user.
对于卷积核而言,卷积窗口的长度不同,则卷积核的感受视野也对应不同。因此,使用卷积窗口不同的卷积层来分别对第一用户的第一用户特征进行卷积处理,所得到第一用户的多个中间局部特征是在不同感受视野下对第一用户的第一用户特征进行提取得到的。每一卷积层输出用户的一中间局部特征。For the convolution kernel, the length of the convolution window is different, and the receptive field of view of the convolution kernel is also correspondingly different. Therefore, the convolutional layers with different convolution windows are used to convolve the first user features of the first user respectively. A user feature is extracted. Each convolutional layer outputs an intermediate local feature of the user.
在本实施例中,在得到用户(第一用户和第二用户)的多个中间局部特征后,将多个中间局部特征进行拼接,将拼接得到的特征作为用户的局部特征。在此基础上,在将用户的拼接得到的局部特征输入到池化层中,由该池化层对用户的局部特征进行池化处理,并输出用户的全局特征。其中,所进行的池化处理可以是平均池化处理、最大池化处理等。In this embodiment, after obtaining the multiple intermediate local features of the user (the first user and the second user), the multiple intermediate local features are concatenated, and the concatenated features are used as the user's local features. On this basis, the local features obtained by the user's concatenation are input into the pooling layer, and the pooling layer performs pooling processing on the user's local features and outputs the user's global features. Wherein, the pooling processing performed may be average pooling processing, maximum pooling processing, and the like.
步骤260,由判别模型根据第一用户的局部特征和第一用户的全局特征计算第一得分;以及由判别模型根据第一用户的全局特征和第二用户的局部特征计算第二得分。In
该判别模型可以是通过卷积神经网络、全连接网络等神经网络构建的模型。通过该判别模型根据第一用户的局部特征和第一用户的全局特征来计算第一得分,由该判别模型根据第一用户的全局特征和第二用户的局部特征计算第二得分,并基于所计算得到的得分来反向调整模型(第一神经网络模型、用户特征模型和判别模型)的参数,从而来缩小同一用户的局部特征与全局特征之间的差距,而拉开第一用户的全局特征和第二用户的局部特征之间的差距,实现最大化用户的全局特征和局部特征之间的信息量,而拉开不同用户的特征之间的区别。The discriminant model may be a model constructed by a neural network such as a convolutional neural network or a fully connected network. The first score is calculated by the discriminant model according to the local features of the first user and the global features of the first user, the second score is calculated by the discriminant model according to the global features of the first user and the local features of the second user, and based on the The calculated score is used to reversely adjust the parameters of the model (the first neural network model, the user feature model, and the discriminant model), thereby narrowing the gap between the local features and the global features of the same user, and widening the global features of the first user. The gap between the features and the local features of the second user maximizes the amount of information between the user's global features and local features, and widens the difference between the features of different users.
由上可以看出,在本申请的方案中,用户特征模型除了用于从第一用户特征中提取特征外,还用于构建无监督对比学习的负样本。具体在本方案中,将第一用户的全局特征和第一用户的局部特征作为正样本,将第二用户的局部特征作为负样本。由于通过该用户特征模型自动构建了对比学习的负样本,因此,在微调训练阶段不需要对数据标注标签,极大减少了模型训练的工作量,而且,由于不需要进行标签标注,可以避免出现因标签标注错误影响模型训练效果的情况。It can be seen from the above that in the solution of this application, the user feature model is not only used to extract features from the first user features, but also used to construct negative samples for unsupervised contrastive learning. Specifically, in this solution, the global features of the first user and the local features of the first user are used as positive samples, and the local features of the second user are used as negative samples. Since the negative samples of comparative learning are automatically constructed through the user feature model, there is no need to label the data during the fine-tuning training stage, which greatly reduces the workload of model training. The situation where the training effect of the model is affected by incorrect labeling.
在一些实施例中,第一得分用于反映第一用户的局部特征与第一用户的全局特征之间的相似度,第二得分用于反映第一用户的全局特征与第二用户的局部特征之间的相似度。从而,在训练过程中,期望第一得分越高越好,第二得分越低越好,以此拉开不同用户的特征之间的区别。In some embodiments, the first score is used to reflect the similarity between the first user's local features and the first user's global features, and the second score is used to reflect the first user's global features and the second user's local features similarity between. Therefore, in the training process, it is expected that the higher the first score, the better, and the lower the second score, the better, so as to widen the difference between the characteristics of different users.
步骤270,根据第一得分和第二得分对第一神经网络模型、用户特征模型和判别模型进行训练,直至达到训练结束条件。
在本申请的一些实施例中,可以分别针对第一神经网络模型、用户特征模型和判别模型设定损失函数,并基于第一神经网络模型的损失函数、用户特征模型的损失函数和判别模型的损失函数设定目标损失函数。在此基础上,根据第一得分和第二得分来分别计算第一神经网络模型、用户特征模型、判别模型的损失函数,进而确定目标损失函数的函数值,按照所计算得到目标损失函数的函数值来调整第一神经网络模型、用户特征模型和判别模型中至少一个模型的参数,并通过调整参数的后第一神经网络模型、用户特征模型和判别模型,再次执行上述步骤220-260的过程,直至所计算得到目标损失函数的函数值使目标函数收敛,然后从第一用户行为数据中获取另外两个用户的交互信息序列来重复上述步骤210-270的过程,直至达到训练结束条件。In some embodiments of the present application, loss functions can be set for the first neural network model, user feature model, and discriminant model, and based on the loss function of the first neural network model, the loss function of the user feature model, and the discriminant model loss function sets the target loss function. On this basis, the loss functions of the first neural network model, the user feature model, and the discriminant model are calculated respectively according to the first score and the second score, and then the function value of the target loss function is determined. According to the calculated function of the target loss function value to adjust the parameters of at least one model in the first neural network model, user feature model and discriminant model, and perform the process of the above steps 220-260 again by adjusting the parameters of the first neural network model, user feature model and discriminant model , until the calculated function value of the target loss function makes the target function converge, and then obtain the interaction information sequence of the other two users from the first user behavior data to repeat the process of the above steps 210-270 until the training end condition is reached.
在一些实施例中,在根据目标损失函数的函数值调整第一神经网络模型、用户特征模型和判别模型的参数的过程中,可以分阶段调整各个模型的参数,一个阶段仅调整一个模型的参数,在下一阶段,再调整其他模型的参数。举例来说,若某一阶段中,固定用户特征模型的参数和固定判别模型的参数,仅调整第一神经网络模型的参数;在下一阶段,固定第一神经网络模型的参数和固定判别模型的参数,调整用户特征模型的参数;在再下一阶段,固定第一神经网络模型的参数和固定用户特征模型的参数,调整判别模型的参数;从而,在微调训练阶段中,按照上述的过程交替调整各个模型的参数。当然,在其他实施例中,还可以设定在一个阶段仅调整两个模型的参数,在此不进行具体限定。由于在一次参数调整过程中仅调整一个或者两个模型的参数,减少了参数调整量,提高模型的训练速度和训练效率。In some embodiments, in the process of adjusting the parameters of the first neural network model, the user feature model, and the discriminant model according to the function value of the target loss function, the parameters of each model can be adjusted in stages, and only the parameters of one model can be adjusted in one stage , and in the next stage, adjust the parameters of other models. For example, if in a certain stage, the parameters of the fixed user feature model and the parameters of the fixed discriminant model are adjusted, only the parameters of the first neural network model are adjusted; in the next stage, the parameters of the fixed first neural network model and the fixed discriminant model are adjusted. Parameters, adjust the parameters of the user feature model; in the next stage, fix the parameters of the first neural network model and the parameters of the fixed user feature model, adjust the parameters of the discriminant model; thus, in the fine-tuning training stage, alternate according to the above process Adjust the parameters of each model. Of course, in other embodiments, it can also be set that only the parameters of the two models are adjusted at one stage, which is not specifically limited here. Since only one or two model parameters are adjusted in a parameter adjustment process, the parameter adjustment amount is reduced, and the training speed and training efficiency of the model are improved.
在一些实施例中,为了保证最大化用户的全局特征和局部特征之间的信息量,拉开不同用户的特征之间的区别,针对判别模型设定的损失函数可以是:In some embodiments, in order to maximize the amount of information between the user's global features and local features, and to separate the differences between the features of different users, the loss function set for the discriminant model can be:
Loss1=log(1+exp-t1)+log(1+expt2);Loss1 = log(1+exp -t1 )+log(1+exp t2 );
其中,Loss1为判别模型的损失函数的函数值,t1表示判别模型计算的第一得分,t2表示判别模型计算的第二得分。Wherein, Loss1 is the function value of the loss function of the discriminant model, t1 denotes the first score calculated by the discriminant model, and t2 denotes the second score calculated by the discriminant model.
步骤280,根据训练后的第一神经网络模型和训练后的用户特征模型确定信息处理模型,信息处理模型在目标域中确定向用户待发送的目标信息。Step 280: Determine an information processing model according to the trained first neural network model and the trained user characteristic model, and the information processing model determines target information to be sent to the user in the target domain.
在本方案中,判别模型可以是用于辅助微调训练的模型,因此,可以是将训练后的第一神经网络模型和训练后的用户特征模型组合,以此确定信息处理模型。在一些实施例中,由于微调训练阶段的任务和目标域中推荐任务的不同,即在微调训练阶段的任务是根据用户的交互信息序列输出用户特征,而在目标域中的推荐任务是预测用户针对目标域中的信息的行为标签,该行为标签用户指示所预测到用户针对信息触发的用户行为,因此,还可以在训练后的第一神经网络模型和训练后的用户特征模型的基础上,加入分类层,通过该分类层来输出所预测到用户针对目标域中信息的行为标签。In this solution, the discriminant model may be a model for auxiliary fine-tuning training, therefore, the information processing model may be determined by combining the trained first neural network model and the trained user feature model. In some embodiments, due to the difference between the task in the fine-tuning training phase and the recommendation task in the target domain, that is, the task in the fine-tuning training phase is to output user features according to the user’s interaction information sequence, while the recommendation task in the target domain is to predict user For the behavior label of the information in the target domain, the behavior label user indicates the predicted user behavior triggered by the user for the information. Therefore, on the basis of the trained first neural network model and the trained user feature model, A classification layer is added, and the predicted user behavior labels for information in the target domain are output through the classification layer.
目标用户泛指目标域中待进行信息发送的用户,在一些实施例中,该目标用户可以是目标域中的新用户,或者在目标域交互数据少于预设阈值的用户。The target user generally refers to a user in the target domain to whom information is to be sent. In some embodiments, the target user may be a new user in the target domain, or a user whose interaction data in the target domain is less than a preset threshold.
在目标域中,由于缺少用户针对目标域中信息的交互数据,或者用户针对目标域中的信息的交互数据较少,因此,在该种情况下,由于通过源域的第一用户行为数据和第二用户行为数据所训练得到的信息处理模型准确地学习到了用户的特征,因此,可以将信息处理模型所学习到的用户特征迁移到目标域的信息发送任务中,由该信息处理模型应用在源域中学习到的用户特征来预测用户针对目标域中的信息的行为标签,然后根据该行为标签来针对性确定向用户待发送的目标信息。In the target domain, due to the lack of user interaction data for information in the target domain, or the lack of user interaction data for information in the target domain, in this case, due to the first user behavior data and The information processing model trained by the second user behavior data has accurately learned the characteristics of the user. Therefore, the user characteristics learned by the information processing model can be transferred to the information sending task of the target domain, and the information processing model can be applied in The user characteristics learned in the source domain are used to predict the user's behavior label for the information in the target domain, and then the target information to be sent to the user is determined according to the behavior label.
在本申请的方案中,在通过源域的第二用户行为数据对第一神经网络模型进行预训练,以使该第一神经网络模型根据用户在源域中的交互信息序列学习用户的用户特征,在此基础上,再结合判别模型,利用源域的第一用户行为数据对第一神经网络模型、用户特征模型进行无监督对比学习,拉开不同用户的特征之间的区别,提高基于交互信息序列所提取到用户特征的准确性。从而,根据训练后的第一神经网络模型和用户特征模型所确定的信息处理模型后,由于该信息处理模型在源域中准确学习到了用户的特征,该用户的特征体现了用户对于信息的偏好,因此,由该信息处理模型根据在源域中所学习到的用户特征来在目标域中确定向用户待发送的目标信息,由于该目标信息的确定结合了用户在源域中的用户特征,因此,可以提高目标信息与用户之间的匹配性,从而可以有效减少与用户之间的交互次数,有效解决现有技术中因信息与用户之间匹配性不高所导致信息发送效率不高的问题。In the solution of this application, the first neural network model is pre-trained through the second user behavior data in the source domain, so that the first neural network model learns the user's user characteristics according to the user's interaction information sequence in the source domain , on this basis, combined with the discriminant model, using the first user behavior data in the source domain to conduct unsupervised comparative learning on the first neural network model and the user feature model, to widen the difference between the features of different users, and improve the interaction-based The accuracy of user features extracted from information sequences. Therefore, after the information processing model determined according to the trained first neural network model and the user feature model, since the information processing model has accurately learned the user's features in the source domain, the user's features reflect the user's preference for information , therefore, the information processing model determines the target information to be sent to the user in the target domain according to the user characteristics learned in the source domain, since the determination of the target information combines the user characteristics of the user in the source domain, Therefore, the matching between the target information and the user can be improved, thereby effectively reducing the number of interactions with the user, and effectively solving the problem of low information sending efficiency caused by the low matching between the information and the user in the prior art. question.
而且,本申请的方案还可以应用于在冷启动场景下为用户确定待发送的目标信息,可有效提升在目标域的冷启动场景下为用户所确定目标信息与用户之间的匹配性,而不需要多次向用户发送信息,提高信息发送效率,对于用户来说,提高了用户获取效率,题,提升了用户体验。Moreover, the solution of the present application can also be applied to determine the target information to be sent for the user in the cold start scenario, which can effectively improve the matching between the target information determined for the user and the user in the cold start scenario of the target domain, and There is no need to send information to users multiple times, which improves the efficiency of information sending. For users, the efficiency of user acquisition is improved, and the user experience is improved.
图3是根据本申请一实施例示出在微调训练进行无监督对比学习的示意图。如图3所示,由预训练后的第一神经网络模型310分别输出第一用户的第一用户特征341和第二用户的第一用户特征342。然后,由用户特征模型320为第一用户生成局部特征和全局特征、以及为第二用户生成局部特征。具体的,如图3所示,由用户特征模型中的不同窗口长度的卷积层对第一用户的第一用户特征341进行N-gram的局部特征提取,得到第一用户的多个中间局部特征351,再将第一用户的多个中间局部特征351进行拼接,得到第一用户的局部特征361,其后,将第一用户的局部特征361进行平均池化处理,得到第一用户的全局特征362。按照相似的过程,用户特征模型320输出第二用户的中间局部特征352、第二用户的局部特征363。Fig. 3 is a schematic diagram showing unsupervised contrastive learning in fine-tuning training according to an embodiment of the present application. As shown in FIG. 3 , the
之后,由判别模型330根据第一用户的局部特征361和第一用户的全局特征362计算第一得分,由判别模型330根据第二用户的局部特征363和第一用户的全局特征362计算第二得分。再根据所计算得到的第一得分和第二得分反向调整第一神经网络模型310、用户特征模型320和判别模型330的参数。Afterwards, the
在本申请的一些实施例中,第一神经网络模型包括基础神经网络模型和插入基础神经网络模型的自适应网络,基础神经网络模型的参数是根据第二训练数据进行预训练确定的;步骤220之前,该方法还包括:固定基础神经网络模型的参数;在本实施例中,步骤270,包括:根据第一得分和第二得分,调整自适应网络、用户特征模型和判别模型的参数并继续训练,直至达到训练结束条件。In some embodiments of the present application, the first neural network model includes a basic neural network model and an adaptive network inserted into the basic neural network model, and the parameters of the basic neural network model are determined by pre-training according to the second training data; step 220 Before, the method also includes: fixing the parameters of the basic neural network model; in this embodiment,
在本方案中,预训练过程中,调整基础神经网络模型的参数;预训练后,固定基础神经网络模型的参数,并在基础神经网络模型中插入自适应网络,从而,在微调训练阶段,对于第一神经网络模型,仅调整自适应网络的参数,而保持基础神经网络模型的参数不变,可以在微调训练阶段保留预训练阶段的学习结果,同时极大减少了参数调整量,有效提高了模型训练速度和训练效果。In this scheme, during the pre-training process, the parameters of the basic neural network model are adjusted; after pre-training, the parameters of the basic neural network model are fixed, and an adaptive network is inserted into the basic neural network model, so that in the fine-tuning training stage, for The first neural network model only adjusts the parameters of the adaptive network while keeping the parameters of the basic neural network model unchanged. It can retain the learning results of the pre-training stage in the fine-tuning training stage, and at the same time greatly reduces the amount of parameter adjustment, effectively improving the Model training speed and training effect.
在本申请的一些实施例中,基础神经网络模型包括嵌入层和多个级联的转换器神经网络。图4是根据本申请的一实施例示出的基础神经网络模型的示意图,在图4中未示出基础神经网络模型400的嵌入层,仅示出了该基础神经网络模型中的多个级联的转换器神经网络410。该嵌入层用于对信息进行转换,得到各信息的嵌入向量。例如,若输入的是源域中一用户的交互信息序列,则该嵌入层可针对用户交互信息序列中每一信息所对应信息ID生成对应的嵌入向量,如图4所示,针对每一信息ID对应的嵌入向量,根据该信息ID在用户交互信息序列的位置,对该信息ID进行位置编码,然后将该信息ID对应的位置编码和该信息ID对应的嵌入向量进行叠加,作为该信息ID的嵌入向量。然后将用户交互信息序列中全部信息ID的嵌入向量输入到初级的转换器神经网络410中,进行特征提取,并将输出的特征输入到下一级的转换器神经网络410中。In some embodiments of the present application, the basic neural network model includes an embedding layer and multiple cascaded converter neural networks. FIG. 4 is a schematic diagram of a basic neural network model according to an embodiment of the present application. The embedding layer of the basic
转换器神经网络又称为Transformer模型,Transformer模型是NLP(NaturalLanguage Processing,自然语言处理)研究领域中用来编码单词、句子的模型结构,是基于注意力机制的模型,能够解决序列长距离依赖的问题,同时也没有像循环神经网络一样严重的梯度弥散及固定的结构限制,也无需像卷积神经网络考虑感受野的范围,总的来说,Transformer模型不仅实作效果好,而且可以并行。The converter neural network is also called the Transformer model. The Transformer model is a model structure used to encode words and sentences in the research field of NLP (Natural Language Processing, Natural Language Processing). It is a model based on the attention mechanism and can solve long-distance dependencies of sequences. The problem is that there is no serious gradient dispersion and fixed structural limitations like the cyclic neural network, and there is no need to consider the range of the receptive field like the convolutional neural network. In general, the Transformer model not only has good implementation effects, but also can be parallelized.
在本申请的一些实施例中,自适应网络可以是插入到转换器网络中,当然,可以是在每一级的转换器神经网络中均插入自适应网络,也可以是在级联的多个转换器神经网络中间隔地插入自适应网络。在另一些实施例中,自适应网络还可以是插入相邻两级转换器网络之间。In some embodiments of the present application, the adaptive network may be inserted into the converter network, of course, the adaptive network may be inserted into each level of the converter neural network, or it may be in cascaded multiple Adaptive networks are inserted at intervals in the transformer neural network. In other embodiments, the adaptive network can also be inserted between adjacent two-stage converter networks.
在本申请的一些实施例中,第一神经网络模型被划分为N个级联的子神经网络,子神经网络包括一转换器神经网络和插入转换器神经网络中的至少一自适应网络;N为大于1的整数;在本实施例中,步骤220,包括:获取k级子神经网络对应的输入信息;其中,若k=1,k级子神经网络对应的输入信息是嵌入层根据第一用户的交互信息序列输出的嵌入向量;若k>1,k级子神经网络对应的输入信息为k-1级子神经网络输出第一用户的k-1级中间用户特征;0<k≤N,k为整数;将k级子神经网络对应的输入信息输入k级子神经网络;由k级子神经网络根据所对应的输入信息进行特征提取,输出第一用户的k级中间用户特征;其中,若k=N,将第一用户的k级中间用户特征作为第一用户的第一用户特征;若k<N,将第一用户的k级中间用户特征作为k+1级子神经网络的输入信息。In some embodiments of the present application, the first neural network model is divided into N cascaded sub-neural networks, and the sub-neural networks include a converter neural network and at least one adaptive network inserted into the converter neural network; N is an integer greater than 1; in this embodiment, step 220 includes: obtaining the input information corresponding to the k-level sub-neural network; wherein, if k=1, the input information corresponding to the k-level sub-neural network is the embedding layer according to the first The embedding vector output by the user's interactive information sequence; if k>1, the input information corresponding to the k-level sub-neural network is the k-1-level intermediate user feature of the first user output by the k-1-level sub-neural network; 0<k≤N , k is an integer; the input information corresponding to the k-level sub-neural network is input into the k-level sub-neural network; the k-level sub-neural network performs feature extraction according to the corresponding input information, and outputs the k-level intermediate user features of the first user; wherein , if k=N, use the k-level intermediate user feature of the first user as the first user feature of the first user; if k<N, use the k-level intermediate user feature of the first user as the k+1-level sub-neural network Enter information.
通过多级的子神经网络来根据用户在源域中的交互信息序列逐级进行特征提取,以保证所提取得到的第一用户特征的准确性。可以理解的是,第二用户的第一用户特征的提取过程与第一用户的第一用户特征的提取过程相似,在此不再赘述第二用户的第一用户特征的提取过程。The multi-level sub-neural network is used to perform feature extraction level by level according to the interaction information sequence of the user in the source domain, so as to ensure the accuracy of the extracted first user feature. It can be understood that the process of extracting the first user features of the second user is similar to the process of extracting the first user features of the first user, and the process of extracting the first user features of the second user will not be repeated here.
图5是根据本申请一实施例示出的子神经网络的示意图,如图5所示,该子神经网络500包括一转换器神经网络和插入该转换器神经网络中的两自适应网络。具体的,如图5所示,该转换器神经网络包括级联的多头注意力层510、第一前馈神经网络层520、第一求和与归一化层530、第二前馈神经网络层540、和第二求和与归一化层550;每一子神经网络500包括插入第一前馈神经网络层520与第一求和与归一化层530之间的自适应网络560,和插入第二前馈神经网络层540与第二求和与归一化层550之间的自适应网络560;子神经网络500还包括从多头注意力层510的输入指向第一求和与归一化层530的输入的恒等映射,和从第二前馈神经网络层540的输入指向第二求和与归一化层550的输入的恒等映射。FIG. 5 is a schematic diagram of a sub-neural network according to an embodiment of the present application. As shown in FIG. 5 , the
一般而言,网络层数越深,能够提取到的不同层次的特征越多,而且不同的层次特征的组合也会越多。但是随着网络层数的不断加深,极容易出现梯度消失和梯度爆炸等问题,会导致网络性能的退化。为了解决这一问题,在本方案中引入残差连接,例如图5中,从多头注意力层的输入指向第一求和与归一化层的输入的恒等映射、从第二前馈神经网络层的输入指向第二求和与归一化层的输入的恒等映射,从而,将浅层的信息应用到到深层的计算中,避免出现梯度消失和梯度爆炸的问题。Generally speaking, the deeper the network layer, the more features of different levels can be extracted, and the more combinations of different levels of features will be. However, as the number of network layers continues to deepen, problems such as gradient disappearance and gradient explosion are extremely prone to occur, which will lead to degradation of network performance. In order to solve this problem, a residual connection is introduced in this scheme. For example, in Figure 5, the identity mapping from the input of the multi-head attention layer to the input of the first summation and normalization layer, and from the second feedforward neural The input of the network layer points to the identity mapping of the input of the second summation and normalization layer, so that the information of the shallow layer is applied to the calculation of the deep layer, and the problems of gradient disappearance and gradient explosion are avoided.
图6是根据本申请一实施例示出的自适应网络的结构示意图,如图6所示,自适应网络(Adapter)560包括级联的首层全连接层610、中间全连接层620、激活层630和末层全连接层640,首层全连接层610中神经元的数量与末层全连接层640中神经元的数量相等,首层全连接层610中神经元的数量大于中间全连接层620中神经元的数量;自适应网络560还包括由首层全连接层610的输入指向末层全连接层640的输出的恒等映射。FIG. 6 is a schematic structural diagram of an adaptive network shown according to an embodiment of the present application. As shown in FIG. 6 , an adaptive network (Adapter) 560 includes a cascaded first fully connected
值得一提的是,在本实施例中是按照数据在自适应网络中的传输方向,将自适应网络(Adapter)中的各全连接层划分为首层全连接层、中间全连接层和末层全连接层。It is worth mentioning that in this embodiment, according to the transmission direction of data in the adaptive network, each fully connected layer in the adaptive network (Adapter) is divided into the first fully connected layer, the middle fully connected layer and the last layer fully connected layer.
对于全连接层而言,该全连接层中神经元的数量决定了该全连接层所输出向量的维度,即若全连接层中神经元的数量为M,则该全连接层所输出的向量为M维。因此,在本方案中,首层全连接层中神经元的数量与末层全连接层的神经元的数量相同,而中间全连接层中神经元的数量小于首层全连接层中神经元的数量,从而使得该自适应网络呈瓶颈样式的MLP(multilayer perceptron,多层感知机)结构。For a fully connected layer, the number of neurons in the fully connected layer determines the dimension of the output vector of the fully connected layer, that is, if the number of neurons in the fully connected layer is M, then the output vector of the fully connected layer It is M dimension. Therefore, in this scheme, the number of neurons in the first fully connected layer is the same as the number of neurons in the last fully connected layer, while the number of neurons in the middle fully connected layer is smaller than the number of neurons in the first fully connected layer. The number, so that the adaptive network is a bottleneck-style MLP (multilayer perceptron, multi-layer perceptron) structure.
激活层630中配置了激活函数,通过激活函数来对上一层的输出进行非线性变换。激活函数可以是ReLU函数、基于高斯误差线性单元(Gaussian Error Linear Units,GELU)的非线性激活函数、双曲正切函数(Tanh函数)等,在此不进行具体限定。由于全连接层层只是对数据进行线性变换(仿射变换),而多个线性变换的叠加仍然是线性变换,因此,通过激活层来增加自适应网络的非线性。值得一提的是,图6仅示例性示出了自适应网络中的一层中间全连接层620,在其他实施例中,自适应网络中可以包括多层中间全连接层。An activation function is configured in the
同理,在自适应网络中,由首层全连接层610的输入指向末层全连接层640的输出的恒等映射,来避免因自适应网络中因网络层数增加出现梯度消失和梯度爆炸等问题,将浅层的首层全连接层的输入引入到位于深层的末层全连接层中进行计算。Similarly, in the adaptive network, the identity mapping from the input of the first fully connected
在本申请的一些实施例中,步骤220之前,如图7所示,该方法还包括:In some embodiments of the present application, before
步骤710,在参考交互信息序列中选取多个目标节点;参考交互信息序列是指第二用户历史行为数据中各用户的交互信息序列。
如上所描述,交互信息序列是多个信息顺序排序形成的,在该交互信息序列中的一信息可以视为一个节点,因此,该参考交互信息序列中包括多个节点,进而在该参考交互信息序列中选取一个或者多个节点作为目标节点。As described above, the mutual information sequence is formed by sequentially sorting a plurality of information, and an information in the mutual information sequence can be regarded as a node. Therefore, the reference mutual information sequence includes multiple nodes, and then the reference mutual information Select one or more nodes in the sequence as the target node.
当然,在其他实施例中,还可以设定预先设定指定数量或者指定比例,从参考交互信息序列中按照该指定数量或者指定比例选取目标节点。其中,指定数量用于指示目标节点的数量,指定比例指示了目标节点的数量在参考交互信息序列中的所占的比例。在具体实施例中,可以是随机选取节点作为目标节点。Of course, in other embodiments, it is also possible to preset a specified number or a specified ratio, and select target nodes from the reference interaction information sequence according to the specified number or specified ratio. Wherein, the specified number is used to indicate the number of target nodes, and the specified ratio indicates the proportion of the number of target nodes in the reference interaction information sequence. In a specific embodiment, a node may be randomly selected as the target node.
步骤720,将参考交互信息序列中的多个目标节点进行遮挡,得到样本交互信息序列。
如上所描述,参考交互信息序列中的各个节点指示了被对应用户触发交互行为的信息,例如被用户触发点击行为的信息等。在本实施例中,将目标节点进行遮挡,从而,让该基础神经网络模型去学习被遮挡起来的目标节点所指示的信息。样本交互信息序列是指将参考交互信息序列中的目标节点进行遮挡后的信息序列。As described above, each node in the reference interaction information sequence indicates the information of the interaction behavior triggered by the corresponding user, such as the information of the click behavior triggered by the user. In this embodiment, the target node is covered, so that the basic neural network model can learn the information indicated by the blocked target node. The sample mutual information sequence refers to the information sequence after occluding the target node in the reference mutual information sequence.
步骤730,将样本交互信息序列输入基础神经网络模型。
步骤740,由基础神经网络模型根据样本交互信息序列,输出对应于目标节点的预测信息。In
如上所描述,该基础神经网络模型需要根据样本交互信息序列,输出该样本交互信息序列中被遮挡的目标节点所指示的信息。具体的,该基础神经网络模型会根据样本交互信息序列中被遮挡节点前后相邻节点所指示的信息来进行预测,换言之,该基础神经网络模型通过学习样本交互信息序列中节点的上下文关系,利用未被遮挡的节点对目标节点所被遮挡的信息进行预测,得到对应于每个目标节点的预测信息。对应于每个目标节点的预测信息用于指示基础神经网络模型所预测到被遮挡的目标节点所指示的信息。As described above, the basic neural network model needs to output the information indicated by the occluded target node in the sample interaction information sequence according to the sample interaction information sequence. Specifically, the basic neural network model will make predictions according to the information indicated by the adjacent nodes before and after the occluded node in the sample interaction information sequence. In other words, the basic neural network model learns the context of the nodes in the sample interaction information sequence. The unoccluded nodes predict the occluded information of the target nodes, and obtain the predicted information corresponding to each target node. The prediction information corresponding to each target node is used to indicate the information indicated by the occluded target node predicted by the basic neural network model.
步骤750,根据预测信息和目标节点,确定预测误差。Step 750: Determine the prediction error according to the prediction information and the target node.
在参考交互信息序列中,目标节点所指示的信息是明确的。因此,可以根据目标节点在参考交互信息序列中所指示的信息和该目标节点所对应的预测信息来确定该预测误差。在一些实施例中,可以预先设定基础神经网络模型的损失函数,该损失函数可以是交叉熵损失函数、均方差损失函数、指数损失函数等,在此不进行具体限定。In the reference mutual information sequence, the information indicated by the target node is clear. Therefore, the prediction error can be determined according to the information indicated by the target node in the reference mutual information sequence and the prediction information corresponding to the target node. In some embodiments, the loss function of the basic neural network model can be preset, and the loss function can be a cross-entropy loss function, a mean square error loss function, an exponential loss function, etc., which are not specifically limited here.
在一些实施例中,基础神经网络模型针对目标节点所输出的是预测信息嵌入向量,在此基础上,可以对应根据获取目标节点在参考交互信息序列中所指示信息的嵌入向量,因此,可以根据对应于目标节点的预测信息的嵌入向量和该目标节点在参考交互信息序列中所指示信息的嵌入向量计算损失函数的函数值,所计算得到损失函数的函数值即为预测误差。In some embodiments, the output of the basic neural network model for the target node is the prediction information embedding vector. On this basis, it can correspond to the embedding vector obtained according to the information indicated by the target node in the reference interaction information sequence. Therefore, it can be obtained according to The function value of the loss function is calculated corresponding to the embedding vector of the prediction information of the target node and the embedding vector of the information indicated by the target node in the reference mutual information sequence, and the calculated function value of the loss function is the prediction error.
步骤760,根据预测误差调整基础神经网络模型的参数并继续训练,直至达到预训练结束条件。
根据所计算得到的预测误差,反向调整基础神经网络模型的参数,然后通过调整参数后的基础神经网络模型重复执行730-750的过程,以使得重新计算得到的预测误差使该基础神经网络模型的损失函数收敛,之后用下一参考交互信息序列按照上述步骤710-750的过程重复对基础神经网络模型进行训练,直至达到预训练结束条件。According to the calculated prediction error, reversely adjust the parameters of the basic neural network model, and then repeat the process of 730-750 through the adjusted basic neural network model, so that the recalculated prediction error makes the basic neural network model The loss function converges, and then use the next reference interaction information sequence to repeat the training of the basic neural network model according to the above steps 710-750 until the pre-training end condition is reached.
在实施例中,将参考交互信息序列中一定比例的目标节点进行遮挡处理,得到样本交互信息序列,由基础神经网络模型根据样本交互信息序列中被遮挡的节点的上下文节点去预测该目标节点在参考交互信息序列中所指示的信息,实现了对基础神经网络模型的自监督训练。In the embodiment, a certain proportion of target nodes in the reference interaction information sequence is occluded to obtain a sample interaction information sequence, and the basic neural network model is used to predict the target node according to the context node of the blocked node in the sample interaction information sequence. With reference to the information indicated in the interaction information sequence, self-supervised training of the underlying neural network model is achieved.
图8是根据本申请一实施例示出的信息处理方法的流程图,如图8所示,该方法包括:步骤810,获取目标域中的候选信息集合。步骤820,通过信息处理模型预测目标用户针对候选信息集合中各信息的行为标签;信息处理模型是按照上述信息处理模型的训练方法的任一实施例训练得到的。FIG. 8 is a flow chart of an information processing method according to an embodiment of the present application. As shown in FIG. 8 , the method includes: Step 810, acquiring a candidate information set in the target domain. Step 820: Predict the target user's behavior label for each information in the candidate information set through the information processing model; the information processing model is trained according to any embodiment of the above information processing model training method.
在一些实施例中,该目标用户可以是目标域中的新用户,而且,该目标用户还是源域中的用户,由于信息处理模型在训练过程中已经学习到源域中用户的用户特征,因此,可以直接将该用户在源域中的用户特征应用到目标域中,作为该目标用户在目标域中的用户特征。在此基础上,该信息处理模型利用目标用户在源域中的用户特征来预测该目标用户针对候选信息集合中各信息的行为标签。In some embodiments, the target user may be a new user in the target domain, and the target user is also a user in the source domain, since the information processing model has learned the user characteristics of the users in the source domain during the training process, therefore , the user characteristics of the user in the source domain can be directly applied to the target domain as the user characteristics of the target user in the target domain. On this basis, the information processing model uses the user characteristics of the target user in the source domain to predict the target user's behavior labels for each information in the candidate information set.
该行为标签用于指示目标用户针对所对应信息触发的用户行为,在一些实施例中,该行为标签用于指示用户是否会点击所对应的信息,对应的,该行为标签包括指示触发点击行为的标签和指示不会触发点击行为的标签。The behavior tag is used to indicate the user behavior triggered by the target user for the corresponding information. In some embodiments, the behavior tag is used to indicate whether the user will click on the corresponding information. Correspondingly, the behavior tag includes Labels and labels that indicate that no click behavior will be triggered.
步骤830,根据行为标签,在候选信息集合中确定向目标用户待发送的目标信息。
根据所预测到候选信息集合中各信息对应的行为标签,从而,可以从候选信息集合中筛选出行为标签指示目标用户会点击的信息,将所筛选出的信息作为向目标用户待发送的目标信息。According to the predicted behavior tags corresponding to each information in the candidate information set, the information that the behavior tag indicates that the target user will click can be filtered out from the candidate information set, and the filtered information can be used as the target information to be sent to the target user .
在本方案中,通过该源域中的第一用户行为数据和第二用户行为数据来对信息处理模型进行训练,以保证该信息处理模型可以准确学习到源域中用户的特征。在此基础上,将信息处理模型在源域中所学习到针对用户的特征应用于目标域中,在源域中所学习到用户的特征来体现用户对于信息的偏好,由于结合了用户在源域中的特征来确定目标域中待发送的目标信息,从而保证了为用户所确定目标信息与目标用户之间的匹配性,因此,可以避免出现多次进行信息筛选和信息发送的情况,有效解决了现有技术因信息与用户之间匹配性不高导致信息发送效率不高的问题。而且,本申请的方案也可以应用于目标域的冷启动场景下向用户确定目标信息。In this solution, the information processing model is trained by using the first user behavior data and the second user behavior data in the source domain, so as to ensure that the information processing model can accurately learn the characteristics of users in the source domain. On this basis, the user-specific features learned by the information processing model in the source domain are applied to the target domain, and the user's features learned in the source domain reflect the user's preference for information. The characteristics in the target domain are used to determine the target information to be sent in the target domain, thereby ensuring the matching between the target information determined for the user and the target user. Therefore, it is possible to avoid multiple times of information screening and information sending, which is effective It solves the problem of low information sending efficiency in the prior art due to low matching between information and users. Moreover, the solution of the present application can also be applied to determine target information to the user in a cold start scenario of the target domain.
在本申请的一些实施例中,信息处理模型包括训练后的第一神经网络模型、训练后的用户特征模型、嵌入查找层和分类层;如图9所示,步骤820,包括:In some embodiments of the present application, the information processing model includes a trained first neural network model, a trained user feature model, an embedded search layer and a classification layer; as shown in FIG. 9,
步骤910,在用户全局特征集合中获取目标用户的全局特征;用户全局特征集合中的全局特征是由训练后的第一神经网络模型和训练后的用户特征模型根据第一用户行为数据生成的。
如上所描述,通过训练后的第一神经网络模型和训练后的用户特征模型,可以根据源域中的第一用户行为数据生成各用户在源域中的全局特征,并将各用户在源域中的全局特征添加到用户全局特征集合中。在将信息处理模型应用到目标域后,在需要对一目标用户进行信息推荐时,若该目标用户即是目标域中的新用户,也是源域中的用户的情况下,可以在用户全局特征集合中查找获取该目标用户的全局特征。As described above, through the trained first neural network model and the trained user feature model, the global features of each user in the source domain can be generated according to the first user behavior data in the source domain, and the global features of each user in the source domain The global features in are added to the user global feature set. After applying the information processing model to the target domain, when it is necessary to recommend information to a target user, if the target user is both a new user in the target domain and a user in the source domain, you can use the user global feature Find the global features of the target user in the collection.
在本申请的一些实施例中,当源域和目标域是按照应用来划分的情况下,假设源域对应于第一应用程序,目标域对应第二应用程序,其中,第一应用程序和第二应用程序可以是登录账号互相开放的应用,例如第一应用程序中的登录账号可以授权在第二应用程序中登录,则该该种情况下,当在第二应用程序中授权登录的账号是第一应用程序中的登录账号的情况下,该用户可以视为既是源域中的用户也是目标域中的用户。从而,将用户的登录账号作为用户标识,在用户全局特征集合中查找获得该用户的全局特征。In some embodiments of the present application, when the source domain and the target domain are divided according to applications, it is assumed that the source domain corresponds to the first application program, and the target domain corresponds to the second application program, wherein the first application program and the second application program The two applications can be applications with login accounts open to each other. For example, the login account in the first application can authorize login in the second application. In this case, when the account authorized to log in in the second application is In the case of a login account in the first application program, the user can be regarded as both a user in the source domain and a user in the target domain. Therefore, the user's login account is used as the user identifier, and the user's global feature is searched in the user's global feature set to obtain the user's global feature.
步骤920,由嵌入查找层生成候选信息集合中各信息的嵌入向量。在一些实施例中,该嵌入查找层可以利用第一神经网络模型中嵌入层对应的嵌入矩阵来生成该候选信息集合中各信息的嵌入向量。In
步骤930,由分类层根据目标用户的全局特征和候选信息集合中各信息的嵌入向量,输出目标用户针对候选信息集合中各信息的行为标签。In
在本申请的一些实施例中,该分类层可以通过softmax函数来根据目标用户的全局特征和候选信息集合中各信息的嵌入向量,来输出目标用户针对候选信息集合中各信息的行为标签。In some embodiments of the present application, the classification layer may use a softmax function to output the target user's behavior labels for each information in the candidate information set according to the target user's global features and the embedding vectors of each information in the candidate information set.
图10是根据本申请一实施例示出的信息处理模型预测行为标签的示意图。如图10所示,第一神经网络模型310根据源域中各用户的交互信息序列输出各用户的第一用户特征,然后由用户特征模型320从各用户的第一用户特征进行深度特征提取,得到各用户在源域上的全局特征,并将各用户在源域上的全局特征添加到用户全局特征集合中。Fig. 10 is a schematic diagram showing an information processing model predicting behavior labels according to an embodiment of the present application. As shown in Figure 10, the first
在此基础上,需要向目标用户确定目标域中的信息时,根据目标用户的用户标识,从用户全局特征集合中获取该目标用户的全局特征。信息处理模型中的嵌入查找层1010根据目标域中各信息所对应的目标域ID(该目标域ID用于标识目标域中的信息)生成该信息的嵌入向量,最后由信息处理模型中的分类层1020根据目标用户的全局特征和目标域中信息所对应的嵌入向量进行分类,输出目标用户针对目标域中信息的行为标签。On this basis, when information in the target domain needs to be determined from the target user, the global features of the target user are obtained from the user global feature set according to the user identifier of the target user. The embedded
在本申请的一些实施例中,步骤820之前,该方法还包括:接收信息请求,信息请求指示了目标用户的用户标识;在本实施例中,步骤820之后,该信息处理方法还包括:根据目标用户的用户标识,向目标用户发送目标信息。In some embodiments of the present application, before
图11是根据本申请一实施例示出在用户界面中显示目标信息的示意图。在图11对应的实施例中,目标域为读书应用所属的领域,其中,该读书应用中“看一看”的子页面中显示该读书应用服务端向用户所发送的信息,具体的,所发送的信息为文章。具体的,在该“看一看”的子页面中可以显示所发送文章的文章标题(如图11中的文章标题I、文章标题II、和文章标题III)、文章作者名称(如图11中的作者I、作者II和作者III),进一步,还可以显示所发送文章的封面图像。Fig. 11 is a schematic diagram illustrating displaying target information in a user interface according to an embodiment of the present application. In the embodiment corresponding to FIG. 11 , the target domain is the field to which the reading application belongs, and the subpage of "Kankan" in the reading application displays the information sent by the reading application server to the user. Specifically, the The information sent is an article. Concrete, can display the article title of sent article (article title I, article title II and article title III among Fig. 11), article author title (as in Fig. 11) in the sub-page of this " take a look " author I, author II and author III), further, the cover image of the sent article can also be displayed.
在具体实施例中,针对本方案的信息处理模型在第一数据集ColdRec-2和第二数据集上进行了严谨的实验,在实验中,将按照本申请的训练方法所训练得到图10所示的信息处理模型与其他推荐模型进行了对比。其中,作为对比参照的推荐模型包括NeuFM、DeepFM(是华为诺亚方舟实验室在2017年提出的推荐模型)、MTL、PeterRec(Parameter-Efficient Transfer from Sequential Behaviors for User Modeling andRecommendation,用于用户建模和推荐的顺序行为的参数高效传输)。在实验中,以HR@5作为参数指标,对各个推荐模型的性能进行了测试。其中,HR@5全称是Hit Ratio Top5,是指召回的前5项的准确率。实验结果如下表1所示。In a specific embodiment, the information processing model of this scheme is rigorously tested on the first data set ColdRec-2 and the second data set. In the experiment, the training method of the present application is used to obtain the data shown in Figure 10. The presented information processing model is compared with other recommendation models. Among them, the recommended models for comparison and reference include NeuFM, DeepFM (recommended model proposed by Huawei Noah's Ark Laboratory in 2017), MTL, PeterRec (Parameter-Efficient Transfer from Sequential Behaviors for User Modeling and Recommendation, used for user modeling and recommended sequential behavior for efficient transfer of parameters). In the experiment, HR@5 is used as the parameter index to test the performance of each recommendation model. Among them, the full name of HR@5 is Hit Ratio Top5, which refers to the accuracy rate of the first 5 items recalled. The experimental results are shown in Table 1 below.
表1Table 1
在如上表1中,ETM模型是按照本申请的训练方法所训练得到如图10所示的信息处理模型。由表1可以看出,相较于其他模型,本申请所训练的得到的信息处理模型在第一数据集ColdRec-2和第二数据集上都取得了超过15%的准确率提升,由此表明了按照本申请的训练方法所得到的信息处理模型有效提升了在目标域中信息确定的准确度。In Table 1 above, the ETM model is trained according to the training method of the present application to obtain the information processing model shown in FIG. 10 . It can be seen from Table 1 that, compared with other models, the information processing model trained by this application has achieved an accuracy rate improvement of more than 15% on both the first data set ColdRec-2 and the second data set. It shows that the information processing model obtained according to the training method of the present application can effectively improve the accuracy of information determination in the target domain.
在本申请的一些实施例中,根据目标用户的用户标识,向目标用户推送目标信息的步骤之后,该信息推荐方法还包括:获取目标用户针对目标信息的交互行为信息;根据目标用户针对目标信息的交互行为信息,对信息处理模型进行更新训练。In some embodiments of the present application, after the step of pushing the target information to the target user according to the user identification of the target user, the information recommendation method further includes: acquiring the target user's interaction behavior information for the target information; The interactive behavior information of the information processing model is updated and trained.
目标用户针对目标信息的交互行为信息用于指示该目标用户针对该目标信息触发的交互行为,例如点击行为等。由服务端可以面向多个目标用户进行目标域中信息的发送,当所收集到目标用户针对目标信息的交互行为信息的数据量足够的情况下,通过根据该目标用户针对目标信息的交互行为信息来对信息处理模型进行更新训练,从而,该信息处理模型可以生成各用户在目标域中的用户特征。进而,在其后的过程中,该信息处理模型可以直接根据用户在目标域中的全局特征来针对性地进行信息发送,而不需要利用用户在源域中的全局特征来进行目标域中信息的发送。The target user's interaction behavior information on the target information is used to indicate the target user's interaction behavior triggered by the target information, such as clicking behavior. The server can send information in the target domain to multiple target users. When the collected data volume of the target user’s interactive behavior information on the target information is sufficient, the target user’s interactive behavior information on the target information can be used to The information processing model is updated and trained, so that the information processing model can generate user features of each user in the target domain. Furthermore, in the subsequent process, the information processing model can send targeted information directly according to the global characteristics of the user in the target domain, without using the global characteristics of the user in the source domain to process information in the target domain. sent.
以下介绍本申请的装置实施例,可以用于执行本申请上述实施例中的方法。对于本申请装置实施例中未披露的细节,请参照本申请上述方法实施例。The following introduces device embodiments of the present application, which may be used to implement the methods in the foregoing embodiments of the present application. For details not disclosed in the device embodiments of the present application, please refer to the foregoing method embodiments of the present application.
图12是根据一实施例示出的信息处理模型的训练装置的框图,如图12所示,该信息处理模型的训练装置包括:第一用户行为数据获取模块1210,用于从源域的第一用户行为数据中获取第一用户的交互信息序列和第二用户的交互信息序列;第一输出模块1220,用于由第一神经网络模型根据第一用户的交互信息序列输出第一用户的第一用户特征,以及第二输出模块1230,用于由第一神经网络模型根据第二用户的交互信息序列输出第二用户的第一用户特征;第一神经网络模型是根据源域的第二用户历史行为数据进行预训练确定的;第一生成模块1240,用于由用户特征模型根据第一用户的第一用户特征生成第一用户的局部特征和第一用户的全局特征;以及第二生成模块1250,用于由用户特征模型根据第二用户的第一用户特征生成第二用户的局部特征;第一得分计算模块1260,用于由判别模型根据第一用户的局部特征和第一用户的全局特征计算第一得分;以及第二得分计算模块1270,用于由判别模型根据第一用户的全局特征和第二用户的局部特征计算第二得分;第一训练模块1280,用于根据第一得分和第二得分对第一神经网络模型、用户特征模型和判别模型进行训练,直至达到训练结束条件;信息处理模型确定模块1290,用于根据训练后的第一神经网络模型和训练后的用户特征模型确定信息处理模型,信息处理模型在目标域中确定向用户待发送的目标信息。FIG. 12 is a block diagram of an information processing model training device according to an embodiment. As shown in FIG. Obtain the interaction information sequence of the first user and the interaction information sequence of the second user from the user behavior data; the first output module 1220 is configured to output the first user's first User features, and a second output module 1230, configured to use the first neural network model to output the first user features of the second user according to the interaction information sequence of the second user; the first neural network model is based on the second user history of the source domain The behavior data is determined by pre-training; the first generation module 1240 is used to generate the local characteristics of the first user and the global characteristics of the first user by the user characteristic model according to the first user characteristics of the first user; and the second generation module 1250 , used to generate the local features of the second user according to the first user features of the second user by the user feature model; the first score calculation module 1260 is used to use the discriminant model to generate the local features of the first user and the global features of the first user Calculate the first score; and the second score calculation module 1270 is used to calculate the second score according to the global features of the first user and the local features of the second user by the discriminant model; the first training module 1280 is used to calculate the second score according to the first score and the local features of the second user. The second score trains the first neural network model, the user feature model and the discriminant model until the training end condition is reached; the information processing
在本申请的一些实施例中,用户特征模型包括至少两个卷积层和池化层,至少两个卷积层中每个卷积层所对应的卷积窗口的长度不同;在本实施例中,第一生成模块1240,包括:第一卷积处理单元,用于由至少两个卷积层中每个卷积层按照所对应的卷积窗口分别对第一用户的第一用户特征进行卷积处理,得到第一用户的多个中间局部特征;第一拼接单元,用于将第一用户的多个中间局部特征进行拼接,得到第一用户的局部特征;第一池化处理单元,用于由池化层对第一用户的局部特征进行池化处理,得到第一用户的全局特征;在本实施例中,第二生成模块1250,包括:第二卷积处理单元,用于由至少两个卷积层中每个卷积层按照所对应的卷积窗口分别对第二用户的第一用户特征进行卷积处理,得到第二用户的多个中间局部特征;第二拼接单元,用于将第二用户的多个中间局部特征进行拼接,得到第二用户的局部特征。In some embodiments of the present application, the user feature model includes at least two convolutional layers and a pooling layer, and the length of the convolution window corresponding to each convolutional layer in the at least two convolutional layers is different; in this embodiment Among them, the
在本申请的一些实施例中,第一神经网络模型包括基础神经网络模型和插入基础神经网络模型的自适应网络,基础神经网络模型的参数是根据源域的第二用户历史行为数据进行预训练确定的;该信息处理模型的训练装置还包括:参数固定模块,用于固定基础神经网络模型的参数;在本实施例中,第一训练模块1280进一步被配置为:根据第一得分和第二得分,调整自适应网络、用户特征模型和判别模型的参数并继续训练,直至达到训练结束条件。In some embodiments of the present application, the first neural network model includes a basic neural network model and an adaptive network inserted into the basic neural network model, and the parameters of the basic neural network model are pre-trained according to the second user historical behavior data in the source domain determined; the training device of the information processing model also includes: a parameter fixing module, used to fix the parameters of the basic neural network model; in this embodiment, the
在本申请的一些实施例中,基础神经网络模型包括嵌入层和多个级联的转换器神经网络,第一神经网络模型被划分为N个级联的子神经网络,子神经网络包括一转换器神经网络和插入转换器神经网络中的至少一自适应网络;N为大于1的整数;第一输出模块1220,包括:输入信息获取单元,用于获取k级子神经网络对应的输入信息;其中,若k=1,k级子神经网络对应的输入信息是嵌入层根据第一用户的交互信息序列输出的嵌入向量;若k>1,k级子神经网络对应的输入信息为k-1级子神经网络输出第一用户的k-1级中间用户特征;0<k≤N,k为整数;第一输入单元,用于将k级子神经网络对应的输入信息输入k级子神经网络;特征提取单元,用于由k级子神经网络根据所对应的输入信息进行特征提取,输出第一用户的k级中间用户特征;其中,若k=N,将第一用户的k级中间用户特征作为第一用户的第一用户特征;若k<N,将第一用户的k级中间用户特征作为k+1级子神经网络的输入信息。In some embodiments of the present application, the basic neural network model includes an embedding layer and a plurality of cascaded converter neural networks, the first neural network model is divided into N cascaded sub-neural networks, and the sub-neural network includes a conversion The neural network of the converter and at least one adaptive network inserted into the neural network of the converter; N is an integer greater than 1; the first output module 1220 includes: an input information acquisition unit, which is used to obtain input information corresponding to the k-level sub-neural network; Among them, if k=1, the input information corresponding to the k-level sub-neural network is the embedding vector output by the embedding layer according to the interaction information sequence of the first user; if k>1, the corresponding input information of the k-level sub-neural network is k-1 The level sub-neural network outputs the k-1 level intermediate user features of the first user; 0<k≤N, k is an integer; the first input unit is used to input the input information corresponding to the k-level sub-neural network into the k-level sub-neural network The feature extraction unit is used to perform feature extraction by the k-level sub-neural network according to the corresponding input information, and output the k-level intermediate user features of the first user; wherein, if k=N, the k-level intermediate user of the first user The feature is used as the first user feature of the first user; if k<N, the k-level intermediate user features of the first user are used as the input information of the k+1-level sub-neural network.
在本申请的一些实施例中,转换器神经网络包括级联的多头注意力层、第一前馈神经网络层、第一求和与归一化层、第二前馈神经网络层、和第二求和与归一化层;每一子神经网络包括插入第一前馈神经网络层与第一求和与归一化层之间的自适应网络,和插入第二前馈神经网络层与第二求和与归一化层之间的自适应网络;子神经网络还包括从多头注意力层的输入指向第一求和与归一化层的输入的恒等映射,和从第二前馈神经网络层的输入指向第二求和与归一化层的输入的恒等映射。In some embodiments of the present application, the converter neural network includes cascaded multi-head attention layers, a first feedforward neural network layer, a first summation and normalization layer, a second feedforward neural network layer, and a second feedforward neural network layer. Two summation and normalization layers; each sub-neural network includes inserting an adaptive network between the first feedforward neural network layer and the first summation and normalization layer, and inserting the second feedforward neural network layer and An adaptive network between the second summation and normalization layer; the sub-neural network also includes an identity mapping from the input of the multi-head attention layer to the input of the first summation and normalization layer, and from the second previous An identity map that feeds the input of the neural network layer to the input of the second summation and normalization layer.
在本申请的一些实施例中,自适应网络包括级联的首层全连接层、中间全连接层、激活层和末层全连接层,首层全连接层中神经元的数量与末层全连接层中神经元的数量相等,首层全连接层中神经元的数量大于中间全连接层中神经元的数量;自适应网络还包括由首层全连接层的输入指向末层全连接层的输出的恒等映射。In some embodiments of the present application, the adaptive network includes a cascaded first fully connected layer, an intermediate fully connected layer, an activation layer, and a final fully connected layer, and the number of neurons in the first fully connected layer is the same as that of the last fully connected layer. The number of neurons in the connection layer is equal, and the number of neurons in the first fully connected layer is greater than the number of neurons in the middle fully connected layer; the adaptive network also includes the input from the first fully connected layer to the last fully connected layer. The identity map of the output.
在本申请的一些实施例中,信息处理模型的训练装置还包括:目标节点选取模块,用于在参考交互信息序列中选取多个目标节点;参考交互信息序列是指第二用户历史行为数据中各用户的交互信息序列;替换模块,用于将参考交互信息序列中的多个目标节点进行遮挡,得到样本交互信息序列;样本交互信息序列输入模块,用于将样本交互信息序列输入基础神经网络模型;预测信息输出模块,用于由基础神经网络模型根据样本交互信息序列,输出对应于目标节点的预测信息;预测误差确定模块,用于根据预测信息和目标节点,确定预测误差;第二训练模块,用于根据预测误差调整基础神经网络模型的参数并继续训练,直至达到预训练结束条件。In some embodiments of the present application, the information processing model training device further includes: a target node selection module, configured to select multiple target nodes in the reference interaction information sequence; the reference interaction information sequence refers to the second user historical behavior data The interaction information sequence of each user; the replacement module is used to block multiple target nodes in the reference interaction information sequence to obtain the sample interaction information sequence; the sample interaction information sequence input module is used to input the sample interaction information sequence into the basic neural network model; the prediction information output module is used to output the prediction information corresponding to the target node by the basic neural network model according to the sample interaction information sequence; the prediction error determination module is used to determine the prediction error according to the prediction information and the target node; the second training The module is used to adjust the parameters of the basic neural network model according to the prediction error and continue training until the pre-training end condition is reached.
图13是根据本申请一实施例示出的信息处理装置的框图,如图13所示,该信息处理装置包括:候选信息集合获取模块1310,用于获取目标域中的候选信息集合;行为标签预测模块1320,用于通过信息处理模型预测目标用户针对候选信息集合中各信息的行为标签;信息处理模型是根据上述任一实施例中信息处理模型的训练方法训练得到的;目标信息确定模块1330,用于根据行为标签,在候选信息集合中确定向目标用户待发送的目标信息。Fig. 13 is a block diagram of an information processing device according to an embodiment of the present application. As shown in Fig. 13 , the information processing device includes: a candidate information set
在本申请的一些实施例中,信息处理模型包括训练后的第一神经网络模型、训练后的用户特征模型、嵌入查找层和分类层;行为标签预测模块1320,包括:全局特征查找单元,用于在用户全局特征集合中获取目标用户的全局特征;用户全局特征集合中的全局特征是由训练后的第一神经网络模型和训练后的用户特征模型根据第一用户行为数据生成的;嵌入向量生成单元,用于由嵌入查找层生成候选信息集合中各信息的嵌入向量;行为标签输出单元,用于由分类层根据目标用户的全局特征和候选信息集合中各信息的嵌入向量,输出目标用户针对候选信息集合中各信息的行为标签。In some embodiments of the present application, the information processing model includes a trained first neural network model, a trained user feature model, an embedding search layer, and a classification layer; the behavior
在本申请的一些实施例中,信息处理装置还包括:接收模块,用于接收信息请求,信息请求指示了目标用户的用户标识;发送模块,用于根据目标用户的用户标识,向目标用户发送目标信息。In some embodiments of the present application, the information processing device further includes: a receiving module, configured to receive an information request, and the information request indicates the user ID of the target user; a sending module, configured to send the target user the target user ID according to the target user ID target information.
在本申请的一些实施例中,信息推荐装置还包括:交互行为信息获取模块,用于获取目标用户针对目标信息的交互行为信息;更新训练模块,用于根据目标用户针对目标信息的交互行为信息,对信息处理模型进行更新训练。In some embodiments of the present application, the information recommending device further includes: an interactive behavior information acquisition module, configured to acquire the target user's interactive behavior information on the target information; an update training module, used to update the target information according to the target user's interactive behavior information on the target information , update and train the information processing model.
图14示出了适于用来实现本申请实施例的电子设备的计算机系统的结构示意图。需要说明的是,图14示出的电子设备的计算机系统1400仅是一个示例,不应对本申请实施例的功能和使用范围带来任何限制。Fig. 14 shows a schematic structural diagram of a computer system suitable for implementing the electronic device of the embodiment of the present application. It should be noted that the
如图14所示,计算机系统1400包括中央处理单元(Central Processing Unit,CPU)1401,其可以根据存储在只读存储器(Read-Only Memory,ROM)1402中的程序或者从存储部分1408加载到随机访问存储器(Random Access Memory,RAM)1403中的程序而执行各种适当的动作和处理,例如执行上述实施例中的方法。在RAM 1403中,还存储有系统操作所需的各种程序和数据。CPU1401、ROM1402以及RAM 1403通过总线1404彼此相连。输入/输出(Input/Output,I/O)接口1405也连接至总线1404。As shown in FIG. 14 , a
以下部件连接至I/O接口1405:包括键盘、鼠标等的输入部分1406;包括诸如阴极射线管(Cathode Ray Tube,CRT)、液晶显示器(Liquid Crystal Display,LCD)等以及扬声器等的输出部分1407;包括硬盘等的存储部分1408;以及包括诸如LAN(Local AreaNetwork,局域网)卡、调制解调器等的网络接口卡的通信部分1409。通信部分1409经由诸如因特网的网络执行通信处理。驱动器1410也根据需要连接至I/O接口1405。可拆卸介质1411,诸如磁盘、光盘、磁光盘、半导体存储器等等,根据需要安装在驱动器1410上,以便于从其上读出的计算机程序根据需要被安装入存储部分1408。The following components are connected to the I/O interface 1405: an
特别地,根据本申请的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本申请的实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信部分1409从网络上被下载和安装,和/或从可拆卸介质1411被安装。在该计算机程序被中央处理单元(CPU)1401执行时,执行本申请的系统中限定的各种功能。In particular, according to the embodiments of the present application, the processes described above with reference to the flowcharts can be implemented as computer software programs. For example, the embodiments of the present application include a computer program product, which includes a computer program carried on a computer-readable medium, where the computer program includes program codes for executing the methods shown in the flowcharts. In such an embodiment, the computer program may be downloaded and installed from a network via
需要说明的是,本申请实施例所示的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(Erasable Programmable Read Only Memory,EPROM)、闪存、光纤、便携式紧凑磁盘只读存储器(Compact Disc Read-Only Memory,CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本申请中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本申请中,计算机可读的信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读的信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:无线、有线等等,或者上述的任意合适的组合。It should be noted that the computer-readable medium shown in the embodiment of the present application may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof. More specific examples of computer-readable storage media may include, but are not limited to, electrical connections with one or more wires, portable computer diskettes, hard disks, random access memory (RAM), read-only memory (ROM), erasable Programmable Read-Only Memory (Erasable Programmable Read Only Memory, EPROM), flash memory, optical fiber, portable compact disk read-only memory (Compact Disc Read-Only Memory, CD-ROM), optical storage device, magnetic storage device, or any suitable The combination. In the present application, a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device. In this application, however, a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, in which computer-readable program codes are carried. Such propagated data signals may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing. A computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can send, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device. . Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wired, etc., or any suitable combination of the above.
附图中的流程图和框图,图示了按照本申请各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。其中,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,上述模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图或流程图中的每个方框、以及框图或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present application. Wherein, each block in the flowchart or block diagram may represent a module, a program segment, or a part of the code, and the above-mentioned module, program segment, or part of the code includes one or more executable instruction. It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or they may sometimes be executed in the reverse order, depending upon the functionality involved. It should also be noted that each block in the block diagrams or flowchart illustrations, and combinations of blocks in the block diagrams or flowchart illustrations, can be implemented by a dedicated hardware-based system that performs the specified function or operation, or can be implemented by a A combination of dedicated hardware and computer instructions.
描述于本申请实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现,所描述的单元也可以设置在处理器中。其中,这些单元的名称在某种情况下并不构成对该单元本身的限定。The units described in the embodiments of the present application may be implemented by software or by hardware, and the described units may also be set in a processor. Wherein, the names of these units do not constitute a limitation of the unit itself under certain circumstances.
作为另一方面,本申请还提供了一种计算机可读存储介质,该计算机可读介质可以是上述实施例中描述的电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。上述计算机可读存储介质承载计算机可读指令,当该计算机可读存储指令被处理器执行时,实现上述任一实施例中的方法。As another aspect, the present application also provides a computer-readable storage medium. The computer-readable medium may be contained in the electronic device described in the above-mentioned embodiments; in the device. The above-mentioned computer-readable storage medium carries computer-readable instructions, and when the computer-readable storage instructions are executed by a processor, the method in any of the above-mentioned embodiments is implemented.
根据本申请的一个方面,还提供了一种电子设备,其包括:处理器;存储器,存储器上存储有计算机可读指令,计算机可读指令被处理器执行时,实现上述任一实施例中的方法。According to one aspect of the present application, there is also provided an electronic device, which includes: a processor; a memory, on which computer-readable instructions are stored, and when the computer-readable instructions are executed by the processor, the implementation of any of the above-mentioned embodiments is realized. method.
根据本申请实施例的一个方面,提供了计算机程序产品或计算机程序,该计算机程序产品或计算机程序包括计算机指令,该计算机指令存储在计算机可读存储介质中。计算机设备的处理器从计算机可读存储介质读取该计算机指令,处理器执行该计算机指令,使得该计算机设备执行上述任一实施例中的方法。According to an aspect of the embodiments of the present application, a computer program product or computer program is provided, the computer program product or computer program includes computer instructions, and the computer instructions are stored in a computer-readable storage medium. The processor of the computer device reads the computer instruction from the computer-readable storage medium, and the processor executes the computer instruction, so that the computer device executes the method in any one of the above embodiments.
应当注意,尽管在上文详细描述中提及了用于动作执行的设备的若干模块或者单元,但是这种划分并非强制性的。实际上,根据本申请的实施方式,上文描述的两个或更多模块或者单元的特征和功能可以在一个模块或者单元中具体化。反之,上文描述的一个模块或者单元的特征和功能可以进一步划分为由多个模块或者单元来具体化。It should be noted that although several modules or units of the device for action execution are mentioned in the above detailed description, this division is not mandatory. Actually, according to the embodiment of the present application, the features and functions of two or more modules or units described above may be embodied in one module or unit. Conversely, the features and functions of one module or unit described above can be further divided to be embodied by a plurality of modules or units.
通过以上的实施方式的描述,本领域的技术人员易于理解,这里描述的示例实施方式可以通过软件实现,也可以通过软件结合必要的硬件的方式来实现。因此,根据本申请实施方式的技术方案可以以软件产品的形式体现出来,该软件产品可以存储在一个非易失性存储介质(可以是CD-ROM,U盘,移动硬盘等)中或网络上,包括若干指令以使得一台计算设备(可以是个人计算机、服务器、触控终端、或者网络设备等)执行根据本申请实施方式的方法。Through the description of the above implementations, those skilled in the art can easily understand that the example implementations described here can be implemented by software, or by combining software with necessary hardware. Therefore, the technical solutions according to the embodiments of the present application can be embodied in the form of software products, which can be stored in a non-volatile storage medium (which can be CD-ROM, U disk, mobile hard disk, etc.) or on the network , including several instructions to make a computing device (which may be a personal computer, server, touch terminal, or network device, etc.) execute the method according to the embodiment of the present application.
本领域技术人员在考虑说明书及实践这里公开的实施方式后,将容易想到本申请的其它实施方案。本申请旨在涵盖本申请的任何变型、用途或者适应性变化,这些变型、用途或者适应性变化遵循本申请的一般性原理并包括本申请未公开的本技术领域中的公知常识或惯用技术手段。Other embodiments of the present application will be readily apparent to those skilled in the art from consideration of the specification and practice of the embodiments disclosed herein. This application is intended to cover any modification, use or adaptation of the application, these modifications, uses or adaptations follow the general principles of the application and include common knowledge or conventional technical means in the technical field not disclosed in the application .
应当理解的是,本申请并不局限于上面已经描述并在附图中示出的精确结构,并且可以在不脱离其范围进行各种修改和改变。本申请的范围仅由所附的权利要求来限制。It should be understood that the present application is not limited to the precise constructions which have been described above and shown in the accompanying drawings, and various modifications and changes may be made without departing from the scope thereof. The scope of the application is limited only by the appended claims.
Claims (16)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202111043012.9A CN115774817B (en) | 2021-09-07 | 2021-09-07 | Training method of information processing model, information processing method and related equipment |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202111043012.9A CN115774817B (en) | 2021-09-07 | 2021-09-07 | Training method of information processing model, information processing method and related equipment |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN115774817A true CN115774817A (en) | 2023-03-10 |
| CN115774817B CN115774817B (en) | 2025-09-19 |
Family
ID=85387605
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202111043012.9A Active CN115774817B (en) | 2021-09-07 | 2021-09-07 | Training method of information processing model, information processing method and related equipment |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN115774817B (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN116975462A (en) * | 2023-09-19 | 2023-10-31 | 西北工业大学 | Information diffusion prediction method based on community Transformer model |
| CN118520599A (en) * | 2024-07-24 | 2024-08-20 | 中国人民解放军军事科学院战争研究院 | Domain adaptive transfer learning-based rotor unmanned aerial vehicle maneuver simulation model generation method |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN107251607A (en) * | 2015-12-22 | 2017-10-13 | 华为技术有限公司 | Transmit method, base station and the user equipment of information |
| CN110297970A (en) * | 2019-05-28 | 2019-10-01 | 北京达佳互联信息技术有限公司 | Information recommendation model training method and device |
| CN110796232A (en) * | 2019-10-12 | 2020-02-14 | 腾讯科技(深圳)有限公司 | Attribute prediction model training method, attribute prediction method and electronic equipment |
| CN110851713A (en) * | 2019-11-06 | 2020-02-28 | 腾讯科技(北京)有限公司 | Information processing method, recommendation method and related equipment |
-
2021
- 2021-09-07 CN CN202111043012.9A patent/CN115774817B/en active Active
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN107251607A (en) * | 2015-12-22 | 2017-10-13 | 华为技术有限公司 | Transmit method, base station and the user equipment of information |
| CN110297970A (en) * | 2019-05-28 | 2019-10-01 | 北京达佳互联信息技术有限公司 | Information recommendation model training method and device |
| CN110796232A (en) * | 2019-10-12 | 2020-02-14 | 腾讯科技(深圳)有限公司 | Attribute prediction model training method, attribute prediction method and electronic equipment |
| CN110851713A (en) * | 2019-11-06 | 2020-02-28 | 腾讯科技(北京)有限公司 | Information processing method, recommendation method and related equipment |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN116975462A (en) * | 2023-09-19 | 2023-10-31 | 西北工业大学 | Information diffusion prediction method based on community Transformer model |
| CN116975462B (en) * | 2023-09-19 | 2025-04-15 | 西北工业大学 | Information diffusion prediction method based on community Transformer model |
| CN118520599A (en) * | 2024-07-24 | 2024-08-20 | 中国人民解放军军事科学院战争研究院 | Domain adaptive transfer learning-based rotor unmanned aerial vehicle maneuver simulation model generation method |
| CN118520599B (en) * | 2024-07-24 | 2024-09-20 | 中国人民解放军军事科学院战争研究院 | Domain adaptive transfer learning-based rotor unmanned aerial vehicle maneuver simulation model generation method |
Also Published As
| Publication number | Publication date |
|---|---|
| CN115774817B (en) | 2025-09-19 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN110807515B (en) | Model Generation Method and Device | |
| CN110287412B (en) | Content recommendation method, recommendation model generation method, equipment, and storage medium | |
| US12488068B2 (en) | Performance-adaptive sampling strategy towards fast and accurate graph neural networks | |
| WO2023065859A1 (en) | Item recommendation method and apparatus, and storage medium | |
| CN107463704A (en) | Searching method and device based on artificial intelligence | |
| CN108920717B (en) | Method and device for displaying information | |
| US11573995B2 (en) | Analyzing the tone of textual data | |
| CN111625715B (en) | Information extraction method and device, electronic equipment and storage medium | |
| US11030402B2 (en) | Dictionary expansion using neural language models | |
| CN112650841A (en) | Information processing method and device and electronic equipment | |
| US20230004862A1 (en) | Method for training ranking learning model, ranking method, device and medium | |
| CN114297417B (en) | Multimedia resource recommendation method and related device | |
| US11914657B2 (en) | Machine learning aided automatic taxonomy for web data | |
| US12111881B2 (en) | Item recommendation with application to automated artificial intelligence | |
| CN108121699A (en) | For the method and apparatus of output information | |
| US12361037B2 (en) | Method for processing question, electronic device and storage medium | |
| US20250356223A1 (en) | Machine-Learning Systems and Methods for Conversational Recommendations | |
| CN115774817B (en) | Training method of information processing model, information processing method and related equipment | |
| CN117235371A (en) | Video recommendation method, model training method and device | |
| US12456072B2 (en) | Method and system for automating scenario planning | |
| CN110222097A (en) | Method and apparatus for generating information | |
| CN113515941B (en) | Named entity recognition method, training method, device, equipment and medium | |
| CN114898184A (en) | Model training method, data processing method and device and electronic equipment | |
| CN114580533A (en) | Method, apparatus, device, medium, and program product for training feature extraction model | |
| US20250209308A1 (en) | Risk Analysis and Visualization for Sequence Processing Models |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |