[go: up one dir, main page]

WO2022148186A1 - Procédé et appareil de traitement de données de séquence comportementale - Google Patents

Procédé et appareil de traitement de données de séquence comportementale Download PDF

Info

Publication number
WO2022148186A1
WO2022148186A1 PCT/CN2021/134635 CN2021134635W WO2022148186A1 WO 2022148186 A1 WO2022148186 A1 WO 2022148186A1 CN 2021134635 W CN2021134635 W CN 2021134635W WO 2022148186 A1 WO2022148186 A1 WO 2022148186A1
Authority
WO
WIPO (PCT)
Prior art keywords
behavior
sample
target
behavior sequence
historical
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2021/134635
Other languages
English (en)
Chinese (zh)
Inventor
牛亚男
宋洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Dajia Internet Information Technology Co Ltd
Original Assignee
Beijing Dajia Internet Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Dajia Internet Information Technology Co Ltd filed Critical Beijing Dajia Internet Information Technology Co Ltd
Publication of WO2022148186A1 publication Critical patent/WO2022148186A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9535Search customisation based on user profiles and personalisation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Definitions

  • the present disclosure relates to the technical field of artificial intelligence, and in particular, to a behavior sequence data processing method, device, electronic device and storage medium.
  • User behavior sequence is the occurrence process of a series of events such as clicks, visits, and purchases generated by users in daily operations. It contains the characteristics of users' fine-grained interests and preferences, and is one of the important feature sources of user-level machine learning models. .
  • behavior sequences including a large number of historical behavior records of users in a long period of time are often directly used as historical data for learning user interest preferences.
  • the present disclosure provides a behavior sequence data processing method, device, electronic device and storage medium.
  • the technical solutions of the present disclosure are as follows:
  • a method for processing behavior sequence data including:
  • the historical behavior sequence of the target object includes a plurality of historical behavior records of the target object
  • position coding information corresponding to each of the historical behavior records is generated, and the position coding information represents the degree of distinction between each of the historical behavior records and other historical behavior records in the plurality of historical behavior records,
  • the degree of discrimination corresponding to each of the historical behavior records is inversely proportional to the time difference corresponding to each of the historical behavior records;
  • the historical behavior sequence is encoded based on the position encoding information to obtain the target behavior sequence feature.
  • the method further includes:
  • the current behavior data representing the behavior data of the recommendation information recommended by the target object to the target object at the current time
  • the encoding processing of the historical behavior sequence based on the position encoding information, and obtaining the target behavior sequence features include:
  • the historical behavior sequence is encoded based on the position encoding information and the current behavior data, to obtain the target behavior sequence feature.
  • the encoding processing of the historical behavior sequence based on the location coding information and the current behavior data, to obtain the target behavior sequence features include:
  • the generating, based on the time difference, the location coding information corresponding to each of the historical behavior records includes:
  • Equal interval classification is performed on the target time difference to obtain a first time difference group corresponding to a plurality of categories
  • the time difference is incrementally classified based on the numerical value of the time difference, and a second time difference group corresponding to multiple categories is obtained, wherein the time difference interval range of the category corresponding to the time difference corresponding to each historical behavior record is the same as that of each
  • the time difference corresponding to the historical behavior record is inversely proportional;
  • One-hot encoding is performed on the second time difference groups corresponding to the multiple categories to obtain the position encoding information.
  • the encoding processing of the historical behavior sequence based on the position encoding information to obtain the target behavior sequence feature includes:
  • the encoding processing of the historical behavior sequence based on the position encoding information to obtain the target behavior sequence feature includes:
  • the position coding information is input to the position coding network for encoding the historical behavior sequence, so as to obtain the target behavior sequence feature.
  • the method further includes:
  • sample behavior sequences of multiple sample objects and multi-task annotation results corresponding to the multiple sample objects where the sample behavior sequence of each sample object includes multiple sample behavior records of each sample object before a preset historical time;
  • sample location coding information corresponding to each of the sample behavior records is generated, and the sample location coding information represents the corresponding value of each of the sample behavior records corresponding to each sample object and each of the sample objects.
  • the degree of distinction between other sample behavior records in the plurality of sample behavior records, the degree of distinction corresponding to each of the sample behavior records is inversely proportional to the sample time difference corresponding to each of the sample behavior records;
  • the first neural network to be trained and the second neural network to be trained are trained based on the target loss to obtain the target encoding network and multitasking network.
  • the method further includes:
  • an apparatus for processing behavior sequence data including:
  • a historical behavior sequence acquisition module configured to execute the acquisition of a historical behavior sequence of a target object, the historical behavior sequence including a plurality of historical behavior records of the target object;
  • a time difference determination module configured to execute and determine the time difference between the behavior time in each historical behavior record and the current time
  • a location coding information generation module configured to generate location coding information corresponding to each of the historical behavior records based on the time difference, the location coding information representing each of the historical behavior records and the plurality of historical behavior records
  • the degree of distinction between other historical behavior records in , the degree of distinction corresponding to each of the historical behavior records is inversely proportional to the time difference corresponding to each of the historical behavior records;
  • the first encoding processing module is configured to perform encoding processing on the historical behavior sequence based on the position encoding information to obtain a target behavior sequence feature.
  • the apparatus further includes:
  • the current behavior data acquisition module is configured to execute and acquire the current behavior data of the target object, the current behavior data representing the behavior data of the recommendation information recommended by the target object to the target object at the current time;
  • the first encoding processing module is further configured to perform encoding processing on the historical behavior sequence based on the position encoding information and the current behavior data to obtain the target behavior sequence feature.
  • the first encoding processing module includes:
  • the first position encoding unit is configured to perform the replacement of the behavior time of each historical behavior record in the historical behavior sequence with the corresponding position encoding information to obtain the target behavior sequence;
  • a first feature extraction processing unit configured to perform feature extraction on the target behavior sequence and the current behavior data, to obtain initial behavior sequence features corresponding to the target behavior sequence and behavior feature information corresponding to the current behavior data ;
  • the first attention learning unit is configured to perform attention learning on the initial behavior sequence feature and the behavior feature information to obtain the target behavior sequence feature.
  • the position coding information generation module includes:
  • a first logarithmic transformation unit configured to perform logarithmic transformation on the time difference to obtain a target time difference
  • a first equal interval classification unit configured to perform equal interval classification on the target time difference to obtain first time difference groups corresponding to multiple categories
  • a first one-hot encoding unit configured to perform one-hot encoding on the first time difference groups corresponding to the multiple categories to obtain the position encoding information
  • the first incremental classification unit is configured to perform incremental classification on the time difference based on the numerical value of the time difference, and obtain a second time difference group corresponding to a plurality of categories, wherein the time difference corresponding to each of the historical behavior records is determined.
  • the time difference interval range of the corresponding category is inversely proportional to the time difference corresponding to each of the historical behavior records;
  • the second one-hot encoding unit is configured to perform one-hot encoding on the second time difference groups corresponding to the multiple categories to obtain the position encoding information.
  • the first encoding processing module includes:
  • the second position encoding unit is configured to perform the replacement of the behavior time of each historical behavior record in the historical behavior sequence with the corresponding position encoding information to obtain the target behavior sequence;
  • a second feature extraction unit configured to perform feature extraction on the target behavior sequence to obtain initial behavior sequence features corresponding to the target behavior sequence
  • the second attention learning unit is configured to perform attention learning on the initial behavior sequence feature to obtain the target behavior sequence feature.
  • the first encoding processing module is further configured to perform encoding processing of inputting the position encoding information into the historical behavior sequence into a position encoding network to obtain the target behavior sequence feature.
  • the apparatus further includes:
  • the training data acquisition module is configured to perform acquisition of sample behavior sequences of multiple sample objects and multi-task annotation results corresponding to the multiple sample objects, and the sample behavior sequence of each sample object includes the sample behavior sequence of each sample object at a preset historical time Multiple previous sample behavior records;
  • a sample time difference determination module configured to execute and determine the sample time difference between the behavior time in each sample behavior record and the preset historical time
  • a sample position coding information generation module configured to generate, based on the sample time difference, sample position coding information corresponding to each of the sample behavior records, the sample position coding information representing each of the samples corresponding to each sample object
  • the degree of distinction between behavior records and other sample behavior records in the plurality of sample behavior records corresponding to each of the sample objects, the degree of discrimination corresponding to each of the sample behavior records and the sample corresponding to each of the sample behavior records The time difference is inversely proportional;
  • the second encoding processing module is configured to input the sample behavior sequence and the sample position encoding information into the first neural network to be trained for encoding processing to obtain the sample behavior sequence features;
  • a second multi-task processing module configured to perform multi-task processing by inputting the sample sequence features into the second neural network to be trained to obtain multi-task prediction results corresponding to the plurality of sample objects;
  • a target loss determination module configured to determine a target loss according to the multi-task prediction result and the multi-task labeling result
  • a network training module configured to perform training of the first neural network to be trained and the second neural network to be trained based on the target loss to obtain the target encoding network and multitasking network.
  • the apparatus further includes:
  • a first multi-task processing module configured to perform multi-task processing by inputting the target behavior sequence feature into a multi-task processing network to obtain a multi-task processing result
  • an information recommendation module configured to perform recommending target information to the target object according to the multitasking result.
  • an electronic device comprising: a processor; a memory for storing instructions executable by the processor; wherein the processor is configured to execute the instructions to achieve The method of any one of the first aspects above.
  • a non-volatile computer-readable storage medium when instructions in the storage medium are executed by a processor of an electronic device, the electronic device can execute the implementation of the present disclosure The method of any one of the first aspects of the examples.
  • a computer program product comprising instructions which, when run on a computer, cause the computer to perform any one of the methods described in the first aspect of the embodiments of the present disclosure.
  • the time difference between the behavior time and the current time in multiple historical behavior records in the historical behavior sequence is combined to generate the position coding information representing the degree of discrimination between each historical behavior record and other historical behavior records.
  • the degree of discrimination corresponding to each historical behavior record is inversely proportional to the time difference corresponding to each historical behavior record, and when encoding and processing the historical behavior sequence, the position encoding information is added, so that the encoding process can better focus on the recent
  • the learning of behavior records ensures that the obtained target behavior sequence features can retain more recent behavior records, which can better reflect the current real interests and preferences of the object, thereby improving the accuracy and effect of subsequent information recommendation.
  • FIG. 1 is a schematic diagram of an application environment according to an exemplary embodiment
  • FIG. 2 is a flowchart of a method for processing behavior sequence data according to an exemplary embodiment
  • FIG. 3 is a schematic flowchart of generating position coding information corresponding to each historical behavior record based on a time difference according to an exemplary embodiment
  • FIG. 4 is a schematic flowchart illustrating a process of encoding a historical behavior sequence based on position coding information to obtain a target behavior sequence feature according to an exemplary embodiment
  • FIG. 5 is a flowchart showing another kind of encoding processing of historical behavior sequences based on position encoding information to obtain characteristics of target behavior sequences according to an exemplary embodiment
  • FIG. 6 is a flowchart of a training target encoding network and a multitasking network according to an exemplary embodiment
  • FIG. 7 is a block diagram of an apparatus for processing behavior sequence data according to an exemplary embodiment
  • Fig. 8 is a block diagram of an electronic device for processing behavior sequence data according to an exemplary embodiment.
  • FIG. 1 is a schematic diagram of an application environment according to an exemplary embodiment.
  • the application environment may include a server 01 and a terminal 02 .
  • Server 01 may be used to train the target encoding network.
  • the server 01 may be an independent physical server, a server cluster or a distributed system composed of multiple physical servers, or a cloud service, cloud database, cloud computing, cloud function, cloud storage, Network services, cloud communications, middleware services, domain name services, security services, CDN (Content Delivery Network), and cloud servers for basic cloud computing services such as big data and artificial intelligence platforms.
  • the terminal 02 may perform behavior sequence data processing in combination with the target coding network trained by the server 01 .
  • the terminal 02 may include, but is not limited to, smartphones, desktop computers, tablet computers, laptop computers, smart speakers, digital assistants, augmented reality (AR)/virtual reality (VR) devices, Electronic devices such as smart wearable devices.
  • the operating system running on the electronic device may include, but is not limited to, an Android system, an IOS system, linux, windows, and the like.
  • FIG. 1 is only an application environment provided by the present disclosure, and other application environments may also be included in practical applications, such as training of target coding networks, which may also be implemented on the terminal 02 .
  • the above-mentioned server 01 and the terminal 02 may be directly or indirectly connected through wired or wireless communication, which is not limited in this disclosure.
  • FIG. 2 is a flowchart of a method for processing behavior sequence data according to an exemplary embodiment. As shown in FIG. 2 , the method for processing behavior sequence data is used in electronic devices such as terminals and edge computing nodes, and includes the following steps:
  • step S201 obtain the historical behavior sequence of the target object
  • step S203 determine the time difference between the behavior time in each historical behavior record and the current time
  • step S205 based on the time difference, the location coding information corresponding to each historical behavior record is generated;
  • step S207 encoding processing is performed on the historical behavior sequence based on the position encoding information to obtain the target behavior sequence feature.
  • step S201 the historical behavior sequence of the target object is acquired.
  • the target object may be the recommendation object of the information in the recommendation system.
  • the target object may be a single user in the recommendation system, or may be a certain group or the like.
  • the historical behavior sequence may include multiple historical behavior records of the target object.
  • the multiple historical behavior records may be historical behavior records of the target object within a period of time before the current time (the period of time may be preset), or may be all historical behavior records of the target object before the current time.
  • each historical behavior record may represent relevant information during the behavior of the target object.
  • the target object often has multiple behaviors in different services, and the target object in the same service can also correspond to multiple behaviors, such as click behavior, video playback behavior, and the like.
  • records corresponding to one or more behaviors may be selected as historical behavior records of the target object in combination with actual application requirements.
  • the user in some scenarios where users passively receive recommended videos, the user often needs to watch (play) the video for a period of time before giving feedback, resulting in a reduction in the user's active choice. It is a historical behavior sequence, which can better reflect the user's interest and preference.
  • each play history record in the historical behavior sequence may include the video id (that is, the video identifier) watched by the target object, and the video author id (that is, author identification), video duration, video tag (topic tag), video viewing duration, video viewing time (action time), etc.
  • step S203 the time difference between the action time in each historical action record and the current time is determined.
  • the long-term behavior of a user reflects the user's multi-interest distribution, while the short-term behavior often reflects the user's current interests.
  • the multiple historical behavior records are generated by the target object performing a certain behavior at different times.
  • the time difference between the behavior time in each historical behavior record and the current time can be determined by determining To distinguish whether the historical behavior record is a short-term behavior record or a long-term behavior record of the target object.
  • step S205 based on the time difference, the location coding information corresponding to each historical behavior record is generated.
  • the above-mentioned location coding information can represent the degree of distinction between each historical behavior record and other historical behavior records in the plurality of historical behavior records, and the time difference between the degree of distinction corresponding to each historical behavior record and the corresponding historical behavior record inversely proportional. That is, the smaller the time difference, the higher the degree of discrimination corresponding to the historical time record; the larger the time difference, the lower the degree of discrimination corresponding to the historical time record.
  • generating the location coding information corresponding to each historical behavior record may include the following steps:
  • step S205 logarithmically transform the time difference to obtain the target time difference
  • step S2053 the target time difference is classified into equal intervals to obtain first time difference groups corresponding to multiple categories;
  • step S2055 one-hot encoding is performed on the first time difference groups corresponding to the multiple categories to obtain position encoding information.
  • an irrational number e when the time difference is logarithmically transformed, an irrational number e may be used as the base, and the time difference is a true number.
  • the smaller the time difference the more discriminating the logarithmically transformed values; ;
  • generating the location coding information corresponding to each historical behavior record may include:
  • the time difference is incrementally classified based on the numerical value of the time difference, and a second time difference group corresponding to a plurality of categories is obtained, wherein the time difference interval range of the category corresponding to the time difference corresponding to each historical behavior record is the same as that of each historical behavior record.
  • the time difference is inversely proportional;
  • One-hot encoding is performed on the second time difference groups corresponding to the multiple categories to obtain position encoding information.
  • the time differences corresponding to the above-mentioned multiple historical time records are divided into four categories, and after the incremental classification, the four categories are: the first category, the time difference is within 0-10 minutes (including 10 minutes); The second category, the time difference is within 10-60 minutes (including 60 minutes); the third category, the time difference is within 60-180 minutes (including 180 minutes); the fourth category, the time difference is greater than 180 minutes.
  • the time difference within 0-10 minutes can be divided into the time difference group corresponding to the first category; the time difference within 10-60 minutes can be divided into the second category.
  • the corresponding time difference group; the time difference within 60-180 minutes can be divided into the time difference group corresponding to the third category; the time difference greater than 180 minutes can be divided into the time difference group corresponding to the fourth category.
  • the time difference corresponding to the multiple historical behavior records is incrementally classified according to the numerical value of the time difference, so that the smaller the time difference, the finer the classification, and the larger the time difference, the coarser the classification, effectively ensuring the location coding information corresponding to the recent historical behavior records.
  • the degree of discrimination between them is higher, thereby effectively ensuring that the recent behavior records have a better degree of discrimination in the subsequent encoding process.
  • the position encoding information corresponding to the time differences in the time difference groups corresponding to the first category may be: 1000;
  • the time difference in the time difference group of , the corresponding position coding information may be: 0100;
  • the time difference in the time difference group corresponding to the third category, the corresponding position coding information may be: 0010;
  • the time difference corresponding to the fourth category For the time difference in the group, the corresponding position coding information may be: 0001.
  • the time difference between the behavior time and the current time in a plurality of historical behavior records in the historical behavior sequence is combined to generate position coding information representing the degree of discrimination between each historical behavior record and other historical behavior records, and each The degree of discrimination corresponding to the historical behavior records is inversely proportional to the time difference corresponding to each historical behavior record, which can effectively ensure that the recent behavior records can be better distinguished in the subsequent encoding process.
  • step S207 the historical behavior sequence is encoded based on the position encoding information to obtain the target behavior sequence feature.
  • encoding the historical behavior sequence based on the position encoding information, and obtaining the target behavior sequence feature may include the following steps:
  • step S401 replace the behavior time of each historical behavior record in the historical behavior sequence with the corresponding position coding information to obtain the target behavior sequence
  • step S403 feature extraction is performed on the target behavior sequence to obtain initial behavior sequence features corresponding to the target behavior sequence;
  • step S405 attention learning is performed on the initial behavior sequence feature to obtain the target behavior sequence feature.
  • the initial behavior sequence feature corresponding to the target behavior sequence may be a feature vector corresponding to the target behavior sequence.
  • feature extraction may be performed on the target behavior sequence, including but not limited to combining with one-hot coding network, N-Gram (Chinese language model) and other feature extraction networks.
  • the initial behavior sequence features may include behavior features corresponding to multiple historical behavior records.
  • performing attention learning on the initial behavior sequence features to obtain the target behavior sequence features may include adding each of the initial behavior sequence features to Dot product each behavior feature with three preset matrices, respectively, to obtain three new eigenvectors corresponding to each behavior feature; perform attention learning based on the three new eigenvectors to obtain the target behavior sequence feature.
  • the dot product of the initial behavior sequence feature and three preset matrices is performed to obtain three corresponding new feature vectors, which may include combining the following formulas:
  • V i X i ⁇ w 3
  • X i represents the i-th behavior feature in the initial behavior sequence features of the target object
  • w 1 , w 2 , and w 3 represent three preset matrices, where w 2 and w 3 may be the same matrix.
  • Q i represents the first eigenvector of the three new eigenvectors corresponding to the ith behavioral feature
  • K i represents the second eigenvector of the three new eigenvectors corresponding to the ith behavioral feature
  • V i represents the th The third eigenvector among the three new eigenvectors corresponding to the i behavioral features.
  • the dot product of each behavior feature in the initial behavior sequence feature and the three preset matrices can add more features, thereby improving the coding effect.
  • performing attention learning based on three new feature vectors to obtain target behavior sequence features may include combining the following formulas:
  • Z i represents the target behavior sequence feature corresponding to the ith behavior feature
  • Q i represents the first feature vector among the three new feature vectors corresponding to the ith behavior feature
  • V i represents the corresponding feature of the ith behavior feature.
  • the third eigenvector in the three new eigenvectors K T represents the second eigenvector corresponding to multiple behavior features in the initial behavior sequence feature
  • d k represents the second eigenvector corresponding to multiple behavior features in the initial behavior sequence feature dimension.
  • the position coding information that can characterize the degree of distinction between the historical behavior record of the target object and other historical behavior records of the target object is combined, and the distinction corresponding to each historical behavior record is combined.
  • the degree is inversely proportional to the time difference corresponding to the historical behavior record, which can effectively ensure that the encoding process can better focus on the learning of recent behavior records, so that the obtained target behavior sequence features retain more recent behavior records, which can better Reflect the object's current real interest preferences, thereby improving the accuracy of subsequent information recommendation.
  • the above method may further include: acquiring current behavior data of the target object, where the current behavior data represents the behavior data of the recommendation information recommended by the target object to the target object at the current time.
  • modules responsible for information recommendation in the recommender system can recommend information to the target object at the current time, and accordingly, can obtain the current behavior data of the target object.
  • the current behavior data represents the target object's recommendation to the target object at the current time.
  • Behavioral data for referral information
  • the historical behavior sequence is encoded based on the location coding information and the current behavior data, and the obtained target behavior sequence features may include:
  • step S501 replace the behavior time of each historical behavior record in the historical behavior sequence with the corresponding position coding information to obtain the target behavior sequence
  • step S503 feature extraction is performed on the target behavior sequence and the current behavior data to obtain initial behavior sequence features corresponding to the target behavior sequence and behavior feature information corresponding to the current behavior data;
  • step S505 perform attention learning on the initial behavior sequence feature and behavior feature information to obtain the target behavior sequence feature.
  • the initial behavior sequence feature corresponding to the target behavior sequence may be a feature vector corresponding to the target behavior sequence
  • the behavior feature information corresponding to the current behavior data may be a feature vector corresponding to the current behavior data.
  • it may include, but is not limited to, feature extraction networks such as one-hot coding network, N-Gram (Chinese language model), etc., to perform feature extraction on the target behavior sequence and current behavior data.
  • performing attention learning on the initial behavior sequence feature and behavior feature information to obtain the target behavior sequence feature may include: performing a dot product on the behavior feature information and a first preset matrix to obtain a fourth feature vector; The behavior sequence features are dot-producted with the second preset matrix and the third preset matrix, respectively, to obtain the fifth feature vector and the sixth feature vector; based on the fourth feature vector, the fifth feature vector and the sixth feature vector, attention learning is performed, Get the target behavior sequence features.
  • obtaining the fourth eigenvector, the fifth eigenvector and the sixth eigenvector above may include combining the following formulas:
  • X represents the initial behavior sequence feature of the target object
  • Y represents the behavior feature information of the target object
  • w 1 , w 2 , w 3 represent the first preset matrix, the second preset matrix and the third preset matrix in sequence, Wherein, w 2 and w 3 may be the same matrix.
  • Q represents the fourth feature vector corresponding to the behavior feature information
  • K represents the fifth sample feature vector corresponding to the initial behavior sequence feature
  • V represents the sixth sample feature vector corresponding to the initial behavior sequence feature.
  • the attention learning is performed based on the fourth feature vector, the fifth feature vector and the sixth feature vector to obtain the target behavior sequence feature, which may include combining the following formulas:
  • Z represents the feature of the target behavior sequence
  • Q represents the fourth feature vector corresponding to the behavior feature information
  • K represents the fifth sample feature vector corresponding to the initial behavior sequence feature
  • V represents the sixth sample feature vector corresponding to the initial behavior sequence feature
  • K T represents the fifth sample feature vector corresponding to the initial behavior sequence feature (because in the self-attention learning process, except for the fifth sample feature vector corresponding to the initial behavior sequence feature, there is no fifth sample feature vector corresponding to its feature)
  • d k represents the dimension of the fifth sample feature vector.
  • the current behavior data of the target object at the current time is added, and more object interest information can be learned.
  • the amount of data in the encoding is small, which can effectively reduce the complexity of the encoding process, thereby improving the processing efficiency.
  • encoding the historical behavior sequence based on the position encoding information to obtain the target behavior sequence feature may include inputting the historical behavior sequence and the position encoding information into the target encoding network for encoding processing to obtain the target behavior sequence feature.
  • the target coding network can be pre-trained. In practical applications, during the training process of the target coding network, the training can be carried out in combination with the task requirements corresponding to the practical application.
  • the above method may further include: pre-training the target encoding network and the multi-tasking network step, in some cases
  • pre-training the target encoding network and the multi-tasking network step in some cases, as shown in Figure 6, it may include:
  • step S601 the sample behavior sequences of the multiple sample objects and the multi-task annotation results corresponding to the multiple sample objects are obtained;
  • step S603 determine the sample time difference between the behavior time in each sample behavior record and the preset historical time
  • step S605 based on the sample time difference, generate sample location coding information corresponding to each sample behavior record;
  • step S607 the sample behavior sequence and the sample position coding information are input into the first neural network to be trained for encoding processing to obtain the sample behavior sequence feature;
  • step S609 the sample sequence features are input into the second neural network to be trained for multi-task processing to obtain multi-task prediction results corresponding to multiple sample objects;
  • step S611 the target loss is determined according to the multi-task prediction result and the multi-task labeling result
  • step S613 the first neural network to be trained and the second neural network to be trained are trained based on the target loss to obtain a target encoding network and a multitasking network.
  • step S601 the sample behavior sequences of the multiple sample objects and the multi-task annotation results corresponding to the multiple sample objects are obtained.
  • the multiple sample objects may be any number of objects in the recommendation system, and the sample behavior sequence of each sample object may include multiple sample behavior records of each sample object before a preset historical time; in some implementations
  • the preset historical time may be a preset historical moment at which information is recommended to the sample object.
  • the historical behavior data of the information recommended by the sample object for the preset historical time may be obtained, and the multi-task annotation result corresponding to the sample object may be determined in combination with the historical behavior data.
  • the subtask annotation results of each task determined in combination with historical behavior data may be used.
  • the historical behavior data of a sample object includes: the object identifier of the sample object, the information identifier of the historical recommendation information, the click information of clicking the historical recommendation information, the historical recommendation information is not forwarded, and it belongs to a long play.
  • a certain task in the multi-task is to predict whether the sample object will click on the historical recommendation information in the historical behavior data; correspondingly, the subtask corresponding to the task is marked as a click.
  • 1 means click, 0 means no click; in some embodiments, in combination with the above historical behavior data, when a certain task in the multi-task is to predict whether the sample object will forward the historical recommendation information in the historical behavior data; correspondingly, The subtask corresponding to the task is marked as unforwarded. In some embodiments, 1 may be used to indicate forwarding, and 0 may be used to indicate unforwarded.
  • multitasking is not limited to the two tasks listed above.
  • it may also include more services in combination with actual business requirements, for example, it may also include duration-related predictions Tasks (such as whether it is an effective play, whether it is a long play, whether it is a short play, and viewing duration prediction), subdivided business prediction tasks (such as whether to download the recommended information, enter the introduction page of whether to enter the recommended information, and stay on the introduction page. duration forecast) and so on.
  • duration-related predictions Tasks such as whether it is an effective play, whether it is a long play, whether it is a short play, and viewing duration prediction
  • subdivided business prediction tasks such as whether to download the recommended information, enter the introduction page of whether to enter the recommended information, and stay on the introduction page. duration forecast
  • step S603 the sample time difference between the behavior time in each sample behavior record and the preset historical time is determined.
  • step S605 based on the sample time difference, the sample position coding information corresponding to each sample behavior record is generated.
  • the sample location coding information represents the degree of distinction between each sample behavior record corresponding to each sample object and other sample behavior records in the multiple sample behavior records corresponding to the sample object, and the corresponding sample behavior record of each sample behavior record.
  • the degree of discrimination is inversely proportional to the sample time difference corresponding to the sample behavior record.
  • generating the sample location coding information corresponding to each sample behavior record based on the sample time difference may include: performing logarithmic transformation on the sample time difference to obtain the target sample time difference; performing equal interval classification on the target sample time difference to obtain multiple The first sample time difference group corresponding to the category; one-hot encoding is performed on the first sample time difference group corresponding to the multiple categories to obtain the sample position encoding information.
  • the one-hot encoding of the sample time difference group can make the position encoding information corresponding to the first sample time difference group in the same category of objects the same, and because the smaller the time difference, the finer the classification, and the sample position encoding corresponding to the recent sample behavior records. The discrimination between the information is higher, thereby effectively ensuring that the recent behavior records have a better discrimination in the subsequent encoding process.
  • generating the sample location coding information corresponding to each sample behavior record based on the sample time difference may include: incrementally classifying the sample time difference based on the numerical value of the sample time difference, and obtaining a second sample time difference group corresponding to multiple categories ; Perform one-hot encoding on the second sample time difference groups corresponding to multiple categories to obtain sample position encoding information.
  • the sample time differences corresponding to multiple sample behavior records are incrementally classified according to the numerical value of the sample time difference, so that the smaller the sample time difference, the finer the classification, and the larger the sample time difference, the coarser the classification, which effectively ensures that the recent sample behavior records correspond to each other.
  • the discriminative degree between the sample position coding information is higher, which effectively ensures that the recent behavior records have a better discrimination degree in the subsequent coding process.
  • the sample time difference between the behavior time in the multiple sample behavior records in the sample behavior sequence and the preset historical time is combined to generate a sample behavior record that can characterize each sample behavior record corresponding to the sample object.
  • the sample location coding information of the discrimination degree between other sample behavior records in each sample behavior record, and the discrimination degree corresponding to each sample behavior record is inversely proportional to the sample time difference corresponding to each sample behavior record, which can effectively ensure that in the subsequent encoding process, Better differentiation of recent behavior records.
  • step S607 the sample behavior sequence and the sample position encoding information are input into the first neural network to be trained for encoding processing to obtain the sample behavior sequence feature.
  • the first neural network to be trained may be an encoding network to be trained.
  • the first neural network to be trained includes: a position encoding layer to be trained, a feature extraction layer to be trained, and an attention learning layer to be trained; correspondingly, the above-mentioned sample behavior sequence and sample position encoding information are input into the first
  • the neural network to be trained performs encoding processing to obtain the sample behavior sequence features, which may include: inputting the sample behavior sequence and sample position encoding information into the position encoding layer to be trained for position encoding to obtain the target sample behavior sequence; inputting the target sample behavior sequence into the to-be-trained feature
  • the extraction layer performs feature extraction to obtain the initial sample behavior sequence features corresponding to the target sample behavior sequence; the initial sample behavior sequence features are input into the attention learning layer to be trained for attention learning, and the sample behavior sequence features are obtained.
  • inputting the sample behavior sequence and the sample position encoding information into the position encoding layer to be trained for position encoding, and obtaining the target sample behavior sequence may include replacing the behavior time in each sample behavior record of the sample behavior sequence with the corresponding The sample position encoding information is used to obtain the target sample behavior sequence.
  • the initial sample behavior sequence feature corresponding to the target sample behavior sequence may be a feature vector corresponding to the target sample behavior sequence.
  • the feature extraction layer to be trained may include, but is not limited to, one-hot (one-hot) encoding network, N-Gram (Chinese language model), and the like.
  • the attention learning layer to be trained may be a self-attention layer in an encoding network in the Transformer.
  • inputting the initial sample behavior sequence features into the attention learning layer to be trained for attention learning, and obtaining the sample behavior sequence features may include: performing a dot product on the initial sample behavior sequence features and three preset matrices to obtain corresponding The three new feature vectors of ; based on the three new feature vectors, the attention learning is performed to obtain the sample behavior sequence features.
  • the initial sample behavior sequence features are input into the attention learning layer to be trained for attention learning, and the specific refinement steps to obtain the sample behavior sequence features can refer to the above-mentioned attention learning of the initial behavior sequence features to obtain the target behavior
  • the specific refinement of the sequence feature will not be repeated here.
  • the three preset matrices may be network parameters.
  • the attention learning layer to be trained may be a multi-head attention learning layer (ie, multiple attention learning layers), and each sample object can obtain a sample after performing attention learning in each attention learning layer Behavior sequence features, correspondingly, by splicing the sample behavior sequence features output by multiple attention learning layers, the sample behavior sequence features of each sample object after being learned by the multi-head attention learning layer can be obtained.
  • a method that can characterize the discrimination between each sample behavior record corresponding to each sample object and other sample behavior records among the multiple sample behavior records corresponding to the sample object is combined.
  • the sample location encodes information, and the degree of discrimination corresponding to each sample behavior record is inversely proportional to the sample time difference corresponding to each sample behavior record, which can effectively ensure that the encoding process can better focus on the learning of recent behavior records, so that the obtained
  • the sample behavior sequence feature retains more recent behavior records, which can better reflect the current real interest and preference of the object, thereby improving the accuracy of subsequent information recommendation.
  • the above method may further include:
  • sample behavior sequence features inputting the sample behavior sequence and the sample position encoding information into the first neural network to be trained for encoding processing, and obtaining the sample behavior sequence features may include:
  • the sample behavior sequence, the sample location coding information and the sample behavior data are input into the first neural network to be trained for encoding processing to obtain the sample behavior sequence feature.
  • inputting the sample behavior sequence, the sample location coding information and the sample behavior data into the first neural network to be trained for encoding processing, and obtaining the sample behavior sequence features may include: inputting the sample behavior sequence and the sample location coding information into the to-be-trained neural network
  • the position encoding layer performs position encoding to obtain the target sample behavior sequence; input the target sample behavior sequence and sample behavior data into the feature extraction layer to be trained for feature extraction, and obtain the initial sample behavior sequence features corresponding to the target sample behavior sequence and the sample behavior data corresponding to Sample behavior feature information; input the initial sample behavior sequence feature and sample behavior feature information into the attention learning layer to be trained for attention learning, and obtain the sample behavior sequence feature.
  • the initial sample behavior sequence features and sample behavior feature information are input into the attention learning layer to be trained for attention learning, and the specific refinement of the sample behavior sequence features can be found in the initial behavior sequence features and behavior features.
  • the specific refinement steps of performing attention learning on the information to obtain the feature of the target behavior sequence will not be repeated here, wherein the first preset matrix, the second preset matrix and the third preset matrix may be network parameters.
  • each sample object when the attention learning layer to be trained is a multi-head attention learning layer (ie, multiple attention learning layers), each sample object can obtain a sample after performing attention learning in each attention learning layer Behavior sequence features, correspondingly, by splicing the sample behavior sequence features output by multiple attention learning layers, the sample behavior sequence features of each sample object after being learned by the multi-head attention learning layer can be obtained.
  • the historical behavior data of the sample object is added, and more object interest information can be learned, and the amount of the historical behavior data is often larger than that in the sample behavior sequence. It can effectively reduce the complexity of the encoding process and improve the processing efficiency.
  • step S609 the sample sequence features are input into the second neural network to be trained for multi-task processing, and multi-task prediction results corresponding to multiple sample objects are obtained.
  • the second neural network to be trained may be a multitasking network to be trained.
  • the multi-task processing network to be trained may be mmoe (Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts, multi-task learning model).
  • multiple sub-feature extraction layers to be trained can extract features from sample sequence features from different angles, but since multiple tasks share multiple sub-feature extraction layers to be trained, in order to highlight the differences between different tasks,
  • the above-mentioned multiple weighted layers of subtasks to be trained may respectively correspond to one task; each weighted layer of subtasks to be trained may be used to weight the subfeature information of multiple sample sequences in combination with the requirements of the corresponding task.
  • the sub-feature information of the sample sequence corresponding to each task can better reflect the feature information concerned by the task.
  • the weights corresponding to the sub-feature information of a plurality of sample sequences may be determined in combination with the degree of attention of different tasks to the sub-feature information of the corresponding sample sequences.
  • sample weighted feature information corresponding to each task can be input into the subtask processing layer corresponding to the task for subtask processing, and the subtask prediction result corresponding to each task can be obtained, and the subtask prediction results corresponding to multiple tasks can be used as.
  • step S611 the target loss is determined according to the multi-task prediction result and the multi-task labeling result.
  • determining the target loss may include calculating the loss between the sub-task prediction result and the sub-task labeling result corresponding to each sample behavior data based on a preset loss function; The losses corresponding to the sample behavior data are added to obtain the above target loss.
  • the preset loss function may include, but is not limited to, a cross-entropy loss function, a logistic loss function, a Hinge (hinge) loss function, an exponential loss function, and the like, and the embodiment of this specification is not limited to the above.
  • step S613 the first neural network to be trained and the second neural network to be trained are trained based on the target loss to obtain the target encoding network and the multitasking network.
  • training the first neural network to be trained and the second neural network to be trained based on the target loss to obtain the target encoding network and the multitasking network may include
  • the target loss is updated based on the updated first neural network to be trained and the second neural network to be trained.
  • the current first neural network to be trained is used as the target coding network
  • the current second neural network to be trained is used as the target encoding network.
  • the network acts as a multitasking network.
  • the target loss meeting the preset condition may be that the target loss is less than or equal to a specified threshold, or the difference between the corresponding target losses in the two training processes before and after is less than a certain threshold.
  • the specified threshold and a certain threshold may be set in combination with actual training requirements.
  • single-task processing can also be performed in combination with the target behavior sequence features output by the target encoding network.
  • the sample time difference between the behavior time in the multiple sample behavior records in the sample behavior sequence and the preset historical time is first combined to generate each sample that can characterize each sample object corresponding to each sample.
  • the sample location coding information of the discrimination degree between the behavior record and other sample behavior records in the multiple sample behavior records corresponding to the sample object, and the discrimination degree corresponding to each sample behavior record is inversely proportional to the sample time difference corresponding to each sample behavior record , and then, in the process of encoding the sample behavior sequence, the target encoding network is added, which can effectively ensure that the encoding process can better focus on the learning of recent behavior records, so that the obtained sample behavior sequence features retain more
  • the recent behavior records can better reflect the current real interests and preferences of the object, improve the prediction accuracy of multi-task processing results, and greatly improve the learning recommendation accuracy and recommendation effect in the recommendation system.
  • the trained target encoding network may include: a feature extraction layer, a position encoding layer and an attention learning layer;
  • the above-mentioned historical behavior sequence and position encoding information are input into the target encoding network for encoding processing, and the characteristics of the target behavior sequence obtained may include:
  • the initial behavior sequence features are input into the attention learning layer for attention learning, and the target behavior sequence features are obtained.
  • inputting the initial behavior sequence feature into the attention learning layer for attention learning, and obtaining the target behavior sequence feature may include: performing a dot product on the initial behavior sequence feature and three preset matrices to obtain corresponding three new The eigenvectors of ; based on three new eigenvectors for attention learning, the target behavior sequence features are obtained.
  • the historical behavior sequence and the position coding information are input into the target coding network for coding processing, and the specific refinement of the relevant steps to obtain the characteristics of the target behavior sequence can refer to the above-mentioned coding processing of the historical behavior sequence based on the position coding information to obtain The specific refinement of the relevant steps of the target behavior sequence feature will not be repeated here.
  • the degree of distinction corresponding to each historical behavior record is combined with each historical behavior record.
  • the time difference corresponding to the behavior records is inversely proportional, which can effectively ensure that in the coding process, the learning of recent behavior records can be better focused, so that the obtained target behavior sequence features retain more recent behavior records, which can better reflect the current behavior of the object. Real interest preferences, thereby improving the accuracy of subsequent information recommendation.
  • the above-mentioned inputting the historical behavior sequence and position encoding information into the target encoding network for encoding processing, and obtaining the target behavior sequence feature may include:
  • the historical behavior sequence, position coding information and current behavior data are input into the target coding network for coding processing, and the target behavior sequence features are obtained.
  • the historical behavior sequence, position encoding information and current behavior data are input into the target encoding network for encoding processing, and obtaining the target behavior sequence feature may include:
  • the initial behavior sequence features and behavior feature information are input into the attention learning layer for attention learning, and the target behavior sequence features are obtained.
  • the historical behavior sequence, location coding information and current behavior data are input into the target coding network for coding processing, and the specific refinement of the relevant steps to obtain the characteristics of the target behavior sequence can refer to the above-mentioned pairing based on the location coding information and the current behavior data.
  • the historical behavior sequence is encoded and processed to obtain the specific refinement of the relevant steps of the feature of the target behavior sequence, which will not be repeated here.
  • the current behavior data of the target object is added, and more object interest information can be learned, and the quantity of the current behavior data is often larger than that in the historical behavior sequence. It can effectively reduce the complexity of the encoding process and improve the processing efficiency.
  • further optimization screening may be performed for subsequent information recommendation based on the above-mentioned target behavior sequence features.
  • the above-mentioned method further includes:
  • the recommendation information whose corresponding task processing result is clicked may be used as the target information, and recommended to the target. object.
  • the information recommendation is performed by the target behavior sequence feature that can effectively reflect the current real interest preference of the target object, which can effectively ensure that the recommended information better meets the user's needs, and improves the recommendation accuracy and recommendation effect.
  • this specification combines the time difference between the behavior time in multiple historical behavior records in the historical behavior sequence and the current time to generate a representation of each historical behavior. Record the location coding information of the degree of discrimination with other historical behavior records. The degree of discrimination corresponding to each historical behavior record is inversely proportional to the time difference corresponding to each historical behavior record, and the location code is added when encoding the historical behavior sequence.
  • Fig. 7 is a block diagram of a behavior sequence data processing apparatus according to an exemplary embodiment.
  • the device includes:
  • the historical behavior sequence acquisition module 710 is configured to execute the acquisition of the historical behavior sequence of the target object, where the historical behavior sequence includes multiple historical behavior records of the target object;
  • a time difference determining module 720 configured to execute and determine the time difference between the behavior time in each historical behavior record and the current time
  • the location encoding information generation module 730 is configured to generate location encoding information corresponding to each historical behavior record based on the time difference, where the location encoding information represents the degree of distinction between each historical behavior record and other historical behavior records in the multiple historical behavior records , the discrimination corresponding to each historical behavior record is inversely proportional to the time difference corresponding to each historical behavior record;
  • the first encoding processing module 740 is configured to perform encoding processing on the historical behavior sequence based on the position encoding information to obtain the target behavior sequence feature.
  • the above-mentioned apparatus further comprises:
  • the current behavior data acquisition module is configured to execute the acquisition of the current behavior data of the target object, and the current behavior data represents the behavior data of the recommendation information recommended by the target object to the target object at the current time;
  • the first encoding processing module 740 is further configured to perform encoding processing on the historical behavior sequence based on the position encoding information and the current behavior data to obtain the target behavior sequence feature.
  • the first encoding processing module 740 includes:
  • the first position encoding unit is configured to perform the replacement of the behavior time of each historical behavior record in the historical behavior sequence with the corresponding position encoding information to obtain the target behavior sequence;
  • a first feature extraction processing unit configured to perform feature extraction on the target behavior sequence and current behavior data, to obtain initial behavior sequence features corresponding to the target behavior sequence and behavior feature information corresponding to the current behavior data;
  • the first attention learning unit is configured to perform attention learning on the initial behavior sequence feature and behavior feature information to obtain the target behavior sequence feature.
  • the position coding information generation module 730 includes:
  • a first logarithmic transformation unit configured to perform logarithmic transformation on the time difference to obtain the target time difference
  • a first equal interval classification unit configured to perform equal interval classification on the target time difference, and obtain first time difference groups corresponding to multiple categories
  • a first one-hot encoding unit configured to perform one-hot encoding on the first time difference groups corresponding to multiple categories to obtain position encoding information
  • the first incremental classification unit is configured to perform incremental classification of the time difference based on the numerical value of the time difference, and obtain a second time difference group corresponding to multiple categories, wherein the time difference interval range of the category corresponding to the time difference corresponding to each historical behavior record Inversely proportional to the time difference corresponding to each historical behavior record;
  • the second one-hot encoding unit is configured to perform one-hot encoding on the second time difference groups corresponding to the multiple categories to obtain position encoding information.
  • the first encoding processing module 740 includes:
  • the second position encoding unit is configured to perform the replacement of the behavior time of each historical behavior record in the historical behavior sequence with the corresponding position encoding information to obtain the target behavior sequence;
  • the second feature extraction unit is configured to perform feature extraction on the target behavior sequence to obtain initial behavior sequence features corresponding to the target behavior sequence;
  • the second attention learning unit is configured to perform attention learning on initial behavior sequence features to obtain target behavior sequence features.
  • the first encoding processing module is further configured to perform encoding processing of inputting the position encoding information into the position encoding network of the historical behavior sequence to obtain the target behavior sequence feature.
  • the above-mentioned apparatus further comprises:
  • the training data acquisition module is configured to perform acquisition of sample behavior sequences of multiple sample objects and multi-task annotation results corresponding to the multiple sample objects.
  • the sample behavior sequence of each sample object includes the sample behavior sequence of each sample object before the preset historical time. Multiple sample behavior records;
  • the sample time difference determination module is configured to execute and determine the sample time difference between the behavior time in each sample behavior record and the preset historical time;
  • the sample position coding information generation module is configured to generate sample position coding information corresponding to each sample behavior record based on the sample time difference, and the sample position coding information represents that each sample behavior record corresponding to each sample object corresponds to each sample object
  • the degree of distinction between other sample behavior records in the multiple sample behavior records, the discrimination degree corresponding to each sample behavior record is inversely proportional to the sample time difference corresponding to each sample behavior record;
  • the second encoding processing module is configured to input the sample behavior sequence and the sample position encoding information into the first neural network to be trained for encoding processing to obtain the sample behavior sequence feature;
  • the second multi-task processing module is configured to perform multi-task processing by inputting the sample sequence features into the second neural network to be trained to obtain multi-task prediction results corresponding to multiple sample objects;
  • the target loss determination module is configured to determine the target loss according to the multi-task prediction result and the multi-task labeling result
  • the network training module is configured to perform training of the first neural network to be trained and the second neural network to be trained based on the target loss to obtain the target encoding network and the multitasking network.
  • the above-mentioned apparatus further includes:
  • the first multi-task processing module is configured to perform multi-task processing by inputting the target behavior sequence feature into the multi-task processing network to obtain a multi-task processing result;
  • the information recommendation module is configured to perform recommending target information to the target object according to the multi-tasking result.
  • FIG. 8 is a block diagram of an electronic device for processing behavior sequence data according to an exemplary embodiment.
  • the electronic device may be a terminal, and its internal structure diagram may be as shown in FIG. 8 .
  • the electronic device includes a processor, a memory, a network interface, a display screen, and an input device connected by a system bus. Among them, the processor of the electronic device is used to provide computing and control capabilities.
  • the memory of the electronic device includes a non-volatile computer-readable storage medium and an internal memory.
  • the computer-readable storage medium stores an operating system and a computer program.
  • the internal memory provides an environment for the execution of the operating system and computer programs in the computer-readable storage medium.
  • the network interface of the electronic device is used to communicate with an external terminal through a network connection.
  • the computer program when executed by a processor, implements a method for processing behavior sequence data.
  • the display screen of the electronic device may be a liquid crystal display screen or an electronic ink display screen
  • the input device of the electronic device may be a touch layer covered on the display screen, or a button, a trackball or a touchpad set on the shell of the electronic device , or an external keyboard, trackpad, or mouse.
  • FIG. 8 is only a block diagram of a part of the structure related to the solution of the present disclosure, and does not constitute a limitation on the electronic device to which the solution of the present disclosure is applied.
  • an electronic device may include more or fewer components than shown in the figures, or combine certain components, or have a different arrangement of components.
  • an electronic device comprising: a processor; a memory for storing instructions executable by the processor; wherein the processor is configured to execute the instructions to implement the present disclosure
  • the behavior sequence data processing method in the example comprising: a processor; a memory for storing instructions executable by the processor; wherein the processor is configured to execute the instructions to implement the present disclosure
  • a non-volatile computer-readable storage medium is also provided, which, when the instructions in the computer-readable storage medium are executed by a processor of the electronic device, enables the electronic device to perform the embodiments of the present disclosure Behavior sequence data processing methods in .
  • the computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, and the like.
  • a computer program product comprising instructions which, when executed on a computer, cause the computer to execute the behavior sequence data processing method in the embodiment of the present disclosure.
  • any reference to memory, storage, database or other medium used in the various embodiments provided in this application may include non-volatile and/or volatile memory.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Biology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

La présente invention concerne le domaine technique de l'intelligence artificielle, et concerne un procédé et un appareil de traitement de données de séquence comportementale, un dispositif électronique et un support de stockage. Le procédé consiste à obtenir une séquence comportementale historique d'un objet cible, la séquence comportementale historique comprenant une pluralité d'enregistrements comportementaux historiques de l'objet cible; déterminer une différence temporelle entre un moment de comportement dans chaque enregistrement comportemental historique et le moment actuel; générer, sur la base de la différence temporelle, des informations de codage d'emplacement correspondant à chaque enregistrement comportemental historique, les informations de codage d'emplacement représentant la netteté entre chaque enregistrement comportemental historique et d'autres enregistrements comportementaux historiques dans la pluralité d'enregistrements comportementaux historiques, et la netteté correspondant à chaque enregistrement comportemental historique étant inversement proportionnelle à la différence temporelle correspondant à chaque enregistrement comportemental historique; et effectuer un traitement de codage sur la séquence comportementale historique sur la base des informations de codage d'emplacement pour obtenir une caractéristique de séquence comportementale cible.
PCT/CN2021/134635 2021-01-11 2021-11-30 Procédé et appareil de traitement de données de séquence comportementale Ceased WO2022148186A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110034304.X 2021-01-11
CN202110034304.XA CN112883257B (zh) 2021-01-11 2021-01-11 行为序列数据处理方法、装置、电子设备及存储介质

Publications (1)

Publication Number Publication Date
WO2022148186A1 true WO2022148186A1 (fr) 2022-07-14

Family

ID=76045059

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/134635 Ceased WO2022148186A1 (fr) 2021-01-11 2021-11-30 Procédé et appareil de traitement de données de séquence comportementale

Country Status (2)

Country Link
CN (1) CN112883257B (fr)
WO (1) WO2022148186A1 (fr)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115438787A (zh) * 2022-09-26 2022-12-06 支付宝(杭州)信息技术有限公司 行为预测系统的训练方法及装置
CN115660105A (zh) * 2022-10-28 2023-01-31 支付宝(杭州)信息技术有限公司 一种模型训练的方法、业务风控的方法及装置
CN115935085A (zh) * 2022-11-15 2023-04-07 中国工商银行股份有限公司 一种基于用户轨迹的服务推荐方法及装置
CN116029422A (zh) * 2022-12-05 2023-04-28 杭州海康威视数字技术股份有限公司 一种行为序列预测方法、装置、电子设备及介质
CN116596093A (zh) * 2023-05-29 2023-08-15 支付宝(杭州)信息技术有限公司 推荐模型的训练方法及装置

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112883257B (zh) * 2021-01-11 2024-01-05 北京达佳互联信息技术有限公司 行为序列数据处理方法、装置、电子设备及存储介质
CN115705583A (zh) * 2021-08-09 2023-02-17 财付通支付科技有限公司 多目标预测方法、装置、设备及存储介质
CN113641835B (zh) * 2021-08-27 2022-10-11 北京达佳互联信息技术有限公司 多媒体资源推荐方法、装置、电子设备及介质
CN114722238B (zh) * 2022-05-13 2022-09-30 北京达佳互联信息技术有限公司 视频推荐方法、装置、电子设备、存储介质及程序产品

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106997360A (zh) * 2016-01-25 2017-08-01 阿里巴巴集团控股有限公司 用户行为数据的处理方法和装置
US20200090056A1 (en) * 2018-09-19 2020-03-19 Tata Consultancy Services Limited Systems and methods for real time configurable recommendation using user data
CN111274462A (zh) * 2020-01-16 2020-06-12 珠海格力电器股份有限公司 数据的处理方法和装置
CN111274501A (zh) * 2020-02-25 2020-06-12 支付宝(杭州)信息技术有限公司 推送信息的方法、系统和非暂时性存储介质
CN111859097A (zh) * 2019-04-30 2020-10-30 北京达佳互联信息技术有限公司 数据处理方法、装置、电子设备及存储介质
US10824940B1 (en) * 2016-11-30 2020-11-03 Amazon Technologies, Inc. Temporal ensemble of machine learning models trained during different time intervals
CN112883257A (zh) * 2021-01-11 2021-06-01 北京达佳互联信息技术有限公司 行为序列数据处理方法、装置、电子设备及存储介质

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106997360A (zh) * 2016-01-25 2017-08-01 阿里巴巴集团控股有限公司 用户行为数据的处理方法和装置
US10824940B1 (en) * 2016-11-30 2020-11-03 Amazon Technologies, Inc. Temporal ensemble of machine learning models trained during different time intervals
US20200090056A1 (en) * 2018-09-19 2020-03-19 Tata Consultancy Services Limited Systems and methods for real time configurable recommendation using user data
CN111859097A (zh) * 2019-04-30 2020-10-30 北京达佳互联信息技术有限公司 数据处理方法、装置、电子设备及存储介质
CN111274462A (zh) * 2020-01-16 2020-06-12 珠海格力电器股份有限公司 数据的处理方法和装置
CN111274501A (zh) * 2020-02-25 2020-06-12 支付宝(杭州)信息技术有限公司 推送信息的方法、系统和非暂时性存储介质
CN112883257A (zh) * 2021-01-11 2021-06-01 北京达佳互联信息技术有限公司 行为序列数据处理方法、装置、电子设备及存储介质

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115438787A (zh) * 2022-09-26 2022-12-06 支付宝(杭州)信息技术有限公司 行为预测系统的训练方法及装置
CN115660105A (zh) * 2022-10-28 2023-01-31 支付宝(杭州)信息技术有限公司 一种模型训练的方法、业务风控的方法及装置
CN115935085A (zh) * 2022-11-15 2023-04-07 中国工商银行股份有限公司 一种基于用户轨迹的服务推荐方法及装置
CN116029422A (zh) * 2022-12-05 2023-04-28 杭州海康威视数字技术股份有限公司 一种行为序列预测方法、装置、电子设备及介质
CN116596093A (zh) * 2023-05-29 2023-08-15 支付宝(杭州)信息技术有限公司 推荐模型的训练方法及装置

Also Published As

Publication number Publication date
CN112883257B (zh) 2024-01-05
CN112883257A (zh) 2021-06-01

Similar Documents

Publication Publication Date Title
WO2022148186A1 (fr) Procédé et appareil de traitement de données de séquence comportementale
US10958748B2 (en) Resource push method and apparatus
CN111291266B (zh) 基于人工智能的推荐方法、装置、电子设备及存储介质
CN114730383B (zh) 使用循环神经网络融合多模态数据
US20230222341A1 (en) Targeted crowd sourcing for metadata management across data sets
CN113641835B (zh) 多媒体资源推荐方法、装置、电子设备及介质
US10650311B2 (en) Suggesting resources using context hashing
US11694165B2 (en) Key-value memory network for predicting time-series metrics of target entities
US11373117B1 (en) Artificial intelligence service for scalable classification using features of unlabeled data and class descriptors
WO2021027256A1 (fr) Procédé et appareil de traitement de données de séquence interactive
US20180101617A1 (en) Ranking Search Results using Machine Learning Based Models
CN114491093B (zh) 多媒体资源推荐、对象表征网络的生成方法及装置
CN114912030A (zh) 权益模型训练方法、推荐方法及电子终端和计算机介质
WO2023024408A1 (fr) Procédé de détermination d'un vecteur de caractéristiques d'un utilisateur, dispositif associé et support
CN112883256B (zh) 多任务处理方法、装置、电子设备及存储介质
US11558471B1 (en) Multimedia content differentiation
CN118570691A (zh) 提取目标帧的方法、电子设备和计算机程序产品
CN115481312B (zh) 应用推荐方法、装置、计算机设备和存储介质
Lakshmi Chetana et al. CF-AMVRGO: collaborative filtering based adaptive moment variance reduction gradient optimizer for movie recommendations
CN114510627B (zh) 对象推送方法、装置、电子设备及存储介质
CN114528472B (zh) 资源推荐模型训练方法、资源信息推荐方法和装置
CN115033793B (zh) 产品推荐模型训练、信息推送方法、装置、设备及介质
CN119719504B (zh) 资源展示方法、装置、电子设备及存储介质
KR102650587B1 (ko) 정량적 기준 및 정성적 기준을 고려한 계약 중개인에 대한 평가 및 매칭 솔루션 제공 방법, 장치 및 시스템
US20250348779A1 (en) Predictive assistance in digital channels using contextual data

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21917227

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 20/10/2023)

122 Ep: pct application non-entry in european phase

Ref document number: 21917227

Country of ref document: EP

Kind code of ref document: A1