US10715869B1 - Deep reinforcement learning for personalized screen content optimization - Google Patents
Deep reinforcement learning for personalized screen content optimization Download PDFInfo
- Publication number
- US10715869B1 US10715869B1 US16/228,123 US201816228123A US10715869B1 US 10715869 B1 US10715869 B1 US 10715869B1 US 201816228123 A US201816228123 A US 201816228123A US 10715869 B1 US10715869 B1 US 10715869B1
- Authority
- US
- United States
- Prior art keywords
- content
- categories
- content items
- requested
- recommended
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/06—Buying, selling or leasing transactions
- G06Q30/0601—Electronic shopping [e-shopping]
- G06Q30/0631—Recommending goods or services
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/466—Learning process for intelligent management, e.g. learning user preferences for recommending movies
- H04N21/4662—Learning process for intelligent management, e.g. learning user preferences for recommending movies characterized by learning algorithms
- H04N21/4666—Learning process for intelligent management, e.g. learning user preferences for recommending movies characterized by learning algorithms using neural networks, e.g. processing the feedback provided by the user
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/088—Non-supervised learning, e.g. competitive learning
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/466—Learning process for intelligent management, e.g. learning user preferences for recommending movies
- H04N21/4668—Learning process for intelligent management, e.g. learning user preferences for recommending movies for recommending content, e.g. movies
Definitions
- the present disclosure relates to improved selection and presentation of content items, and more particularly, to techniques for training a neural net to select content items for presentation by incentivizing selection of diverse explorative content and dis-incentivizing selection of content that is already likely to be requested.
- Computerized content delivery and presentation systems commonly face a task of generating recommendation screens.
- OTT over-the-top
- video delivery systems and online music delivery services commonly need to generate such recommendation screens.
- Generation of appropriate and engaging screens is highly important because it allows such systems to inform the users about relevant content offerings and promote views.
- One approach to selecting content for the recommendation screen is to display content recommendations curated by human editors. However, this approach does not take into account preferences of each individual user and will inevitably result in presentation of irrelevant or redundant content recommendations.
- Another approach is to display groups of content items that are similar to content items most commonly requested by that user. However, this approach commonly results in presentation of homogenous content item recommendations that user was already likely to request.
- Such recommendation techniques are inefficient because they fail to provide new information and fail to expose a user to new relevant content.
- a content recommendation application uses a neural net trained to incentivize selection of relevant content while dis-incentivizing selection of content that is already likely to be requested.
- the content recommendation application performs an intelligent selection of content categories from a superset of all available categories by using a category section neural net that was trained specifically to recommend relevant content to a particular user while avoiding recommendation of content that the particular user would request even without seeing the recommendation screen.
- the content recommendation application identifies a set of content items that are likely to be requested by the particular user. For example, the content recommendation application may identify such a set based on a history of past content item requests from the particular user. Then, the content recommendation application selects a first plurality of content categories using a category selection neural net.
- the neural net may be a set of neurons connecting features (e.g., used features) with a superset of all available content categories. In such a case, the neural net accepts features (e.g., user features) as input and selects some of the possible content categories via the neural connections.
- the content recommendation application selects a first set of recommended content items for the first plurality of content categories. For example, the content recommendation application may query a traditional or custom-built recommendation engine for available content items in the selected categories.
- the content recommendation application may (e.g., without displaying such recommendations) monitor actual content requests received from the particular user over a certain time period (e.g., a week or ten days).
- a request is received from a user for a content item that is included in the first set of recommended content items
- the reward score for the first plurality of content categories is increased (e.g., by a value of 1).
- the reward score for the first plurality of content categories is decreased (e.g., by 1 or by 0.5). This decrease in the reward score allows for training the neural net to avoid selecting content that the user would have watched regardless.
- the content recommendation application uses the computed reward score to train (e.g., adjust) the neural net.
- the connection between neurons may be adjusted to reinforce reward score maximization. For example, neural connections that led to a high reward score are likely to remain, while neural connections that led to a low reward score are likely to be removed or changed.
- the content recommendation application selects a second plurality of content categories using the trained category selection neural net. Then, the content recommendation application selects a second set of recommended content items for the second plurality of content categories. The content recommendation application may then actually display identifiers for recommended content items of the second set of recommended content items.
- such technique is likely to result in presentation of identifiers for content items that are relevant to the particular user but are not the type of items that would have been watched by the user regardless. In this way, the aforementioned technique provides an improved user interface for a computerized content delivery system that provides explorative recommendations to a user while avoiding recommendation of content that is not new to the user.
- FIG. 1 shows an example of selecting content item identifiers for display, in accordance with some embodiments of the disclosure
- FIG. 2 shows an illustrative example of a category selection neural net, in accordance with some embodiments of the disclosure
- FIG. 3 shows a diagram of an illustrative media device, in accordance with some embodiments of the disclosure
- FIG. 4 is a diagram of an illustrative content recommendation application, in accordance with some embodiments of the disclosure.
- FIG. 7 is flowchart of another detailed illustrative process of training a neural network, in accordance with some embodiments of the disclosure.
- FIG. 1 shows an example 100 of selecting content item identifiers for display.
- FIG. 1 shows the operation of a content recommendation application.
- the content recommendation application is executing on a server of an OTT content provider (e.g., Netflix, YouTube, Pandora) that is generating a recommendation screen for a particular use of the OTT system.
- the content recommendation application is executing on a media device of the user.
- the content recommendation application generates a feature set 102 , e.g., based on a stored user profile.
- the user profile may include a history of content requests received from the user in the past.
- the content requests may include requests for recorded video, requests to tune to a video that is being broadcast, requests for music playback, requests for a multimedia presentation, or any combination of the above.
- the content recommendation application extracts features (e.g., used features) including: a list of TV series watched by the user, a list of TV categories watched by the user, a list of movie categories watched by the user, a list content types requested by the user (e.g., content lengths, language, etc.), and a list of collaborative filtering vectors (e.g., filtering vectors as described in U.S. Pat. No. 8,239,412, which is hereby incorporated by reference herein in its entirety).
- features e.g., used features
- features e.g., used features
- a list of TV series watched by the user e.g., a list of TV categories watched by the user, a list of movie categories watched by the user
- a list content types requested by the user e.g., content lengths, language, etc.
- collaborative filtering vectors e.g., filtering vectors as described in U.S. Pat. No. 8,239,412, which is hereby incorporated by reference
- the content recommendation application proceeds to select 110 content items for each of the selected categories 108 .
- the content recommendation application may make an application program interface (API) call to a recommendation engine for each category (e.g., using the feature set as an input).
- API application program interface
- the content recommendation application may use any of recommendations engines and recommendations techniques as described in U.S. Pat. No. 6,388,714 and in U.S. Pat. No. 6,756,997 (both of which are hereby incorporated by reference herein in their entirety).
- the content recommendation application may select several content items for each of the categories as depicted in table 112 .
- the content recommendation application may (without displaying the selected content items) track 114 user's content requests for a certain period of time (e.g., for 10 days). Whenever the content recommendation application detects that the user requested a content item that matches an item in the table 112 , the content recommendation application increases a reward score assigned to the selected set of categories 108 . In some embodiments, the score for each category is tracked separately (as shown in table 116 ), while in other embodiments only the total score is tracked. For example, each time a “hit” occurs, the content recommendation application increments the reward score by 1.
- the content recommendation application also identifies 118 content items that are likely to be requested in the future based on a history of content item requests. For example, the content recommendation application may make an API call to a prediction engine (e.g., using the feature set as input). The content recommendation application may use any of predictive engines or techniques as described in U.S. Pat. No. 8,613,024 and in U.S. Patent Application Publication No. 2018/0189826 (both of which are hereby incorporated by reference herein in their entirety). In one implementation, the content recommendation application decreases a reward score for each “hit” that occurred at block 114 that also matches the content items that were already likely to be requested in the future by the user.
- the content recommendation application may subtract 1 or 0.5 from the reward score. This allows the content recommendation application to train the neural net to select explorative content rather than conventional content recommendations that merely display content that the user is already likely to watch.
- the content recommendation application may adjust the model (e.g., the neural net) using the reward score.
- the model e.g., the neural net
- several sets of categories may have been selected at 106 and independently tracked at 114 , thus allowing the neural net to be trained using multiple training examples. Further, steps 106 - 118 may be repeated several times further refining the model.
- FIG. 2 shows an illustrative example 200 of a neural net 204 , in accordance with some embodiments of the disclosure.
- neural net 204 may be the same as the neural net that was used as a selection model in FIG. 1 .
- neural net 204 may include multiple neurons (e.g., as represented with circular shapes in FIG. 1 ) and connections between neurons (represented as straight lines in FIG. 1 ).
- Each neuron may be a data structure with two states (e.g., ⁇ 1 ⁇ or ⁇ ON ⁇ state, and ⁇ 0 ⁇ or ⁇ OFF ⁇ state).
- Each neuron may have positive or negative connections to neurons of the previous the layer and/or to neurons of the next layer.
- Each connection may be used to communicate the state of the neuron to other neurons.
- Neural net 204 may utilize a feature set (e.g., user feature set for an individual user as described in block 102 ) as input 202 .
- neural net 204 may also use calendar data (e.g., time of day) as one of the inputs 202 .
- the content recommendation application may maintain and train a neural net for each individual user and use the feature set of that user for input layer 202 .
- the content recommendation application converts the feature set to binary vectors that are then used as a starting points of input layer 202 .
- the content recommendation application converts the feature set into principle components using PCA analysis. In that case, the principle components are used as starting points of input layer 202 .
- the input layer 202 accepts the feature set, while the first layer of neural net 204 is defined by PCA-generated principle component vectors.
- each bit of the feature set may be mapped to one of the neurons of layer 202 .
- a value of ⁇ 1 ⁇ in input layer 202 may cause the corresponding neuron of input layer 202 to be set to the ⁇ ON ⁇ state and a value of ⁇ 0 ⁇ in input layer 202 may cause the corresponding neuron to be set to the ⁇ OFF ⁇ state.
- the connections between neurons may then determine the state of the next layer, and so on, until the states of neurons of output layer 206 are determined.
- some or all of the neurons may have a variable weight score.
- signals from neurons with higher weight scores may count more when determining a state of the next neuron. For example, if a neuron has a weight of 2, the input from that neuron may be weighted the same as inputs from two neurons with weight 1.
- deep deterministic policy gradients training is improved by removing any delayed rewards.
- the reward sampling of the environment may be performed beforehand (e.g., as described in FIG. 1 ) making the model more likely to converge and removing the added complexity of trying to maximize exploration/optimal screens in real time.
- I/O path 302 may provide content (e.g., broadcast programming, on demand programming, Internet content, content available over a local area network (LAN) or wide area network (WAN), and/or other content) and data to control circuitry 304 , which includes processing circuitry 306 , display generator circuitry 307 , and storage 308 .
- Control circuitry 304 may be used to send and receive commands, requests, and other suitable data using I/O path 302 .
- I/O path 302 may connect control circuitry 304 (and specifically processing circuitry 306 ) to one or more communications paths (described below). I/O functions may be provided by one or more of these communications paths, but are shown as a single path in FIG. 3 to avoid overcomplicating the drawing.
- control circuitry 304 executes instructions for a user equipment device and/or application stored in memory (i.e., storage 308 ). Specifically, control circuitry 304 may be instructed by a user equipment device and/or application to perform the functions discussed above and below.
- Control circuitry 304 may include communications circuitry suitable for communicating with a data source and processing server or other networks or servers.
- Communications circuitry may include a cable modem, an integrated services digital network (ISDN) modem, a digital subscriber line (DSL) modem, a telephone modem, Ethernet card, or a wireless modem for communications with other equipment, or any other suitable communications circuitry.
- ISDN integrated services digital network
- DSL digital subscriber line
- Such communications may involve the Internet or any other suitable communications networks or paths.
- communications circuitry may include circuitry that enables peer-to-peer communication of media devices, or communication of media devices in locations remote from each other.
- Memory may be an electronic storage device provided as storage 308 that is part of control circuitry 304 .
- the phrase “electronic storage device” or “storage device” should be understood to mean any device for storing electronic data, computer software, or firmware, such as random-access memory, hard drives, optical drives, solid state devices, quantum storage devices, gaming consoles, gaming media, or any other suitable fixed or removable storage devices, and/or any combination of the same.
- Nonvolatile memory may also be used (e.g., to launch a boot-up routine and other instructions).
- Cloud-based storage described in relation to FIG. 4 , may be used to supplement storage 308 or instead of storage 308 .
- a user may send instructions to control circuitry 304 using user input interface 310 of media device 300 or user input interface 310 of media device 301 .
- User input interface 310 may be any suitable user interface, such as a remote control, mouse, trackball, keypad, keyboard, touchscreen, touchpad, stylus input, joystick, voice recognition interface, or other user input interfaces.
- Display 310 may be a touchscreen or touch-sensitive display. In such circumstances, user input interface 310 may be integrated with or combined with display 312 .
- Display 312 may be provided as a stand-alone device or integrated with other elements of media device 301 .
- Speakers 314 may be provided as integrated with other elements of media device 300 .
- speakers 318 may be stand-alone units (e.g., smart speakers).
- the audio component of videos and other content displayed on display 312 may be played through speakers 318 .
- the audio may be distributed to a receiver (not shown), which processes and outputs the audio via speakers 318 .
- Media device 402 may be coupled to communications network 404 .
- Communications network 404 may be one or more networks including the Internet, a mobile phone network, mobile voice or data network (e.g., a 4G or LTE network), cable network, public switched telephone network, or other types of communications network or combinations of communications networks.
- Data source 406 , processing server 408 , and media device 402 may be connected to communication path 404 via one or more communications paths, such as, a satellite path, a fiber-optic path, a cable path, a path that supports Internet communications (e.g., IPTV), free-space connections (e.g., for broadcast or other wireless signals), or any other suitable wired or wireless communications path or combination of such paths.
- Content recommendation applications may be, for example, stand-alone applications implemented on media device 402 , on processing server 408 or any combination of the two.
- a content recommendation application may be implemented as software or a set of executable instructions which may be stored in storage 308 , and executed by control circuitry 304 of a media device 402 .
- content recommendation applications may be client-server applications where only a client application resides on the media device, and a server application resides on processing server 408 .
- content recommendation applications may be implemented partially as a client application on control circuitry 304 of media device 402 and partially on processing server 408 as a server application running on control circuitry of processing server 408 .
- the content recommendation application may instruct the control circuitry to generate the content recommendation application output (e.g., the expiatory recommendation screen) and transmit the generated output to media device 402 .
- the client application may instruct control circuitry of the receiving media device 402 to generate the content recommendation application output. e.g., by using a neural net stored and trained on one of media device 402 or processing server 408 .
- FIG. 5 is a flowchart of an illustrative process for selecting content item identifiers for display using a trained neural net, in accordance with some embodiments of the disclosure.
- each step of process 500 can be performed by media device 402 (e.g., via control circuitry 304 ) or any of the system components shown in FIG. 3 .
- each step of process 500 can be performed by processing server 408 (e.g., via control circuitry 304 of processing server 408 ).
- each step of process 500 may be performed by control circuitry 304 executing instructions of the content recommendation application based on instructions stored in memory 308 to generate an explorative recommendation screen for a particular user (e.g., user of media device 402 ).
- control circuitry 304 selects a first plurality of content categories using a category selection neural net. In some embodiments, this step is performed as described in relation to block 106 of FIG. 1 .
- a neural net 204 may be used where a feature set of the particular user is used as input 202 to select a set of content categories as output 206 .
- the feature set may include one or more of content item requests data, content category requests data, time data, and collaborative filtering vectors.
- the neural net is stored on media device 402 .
- the neural net may be stored on a processing server 408 .
- control circuitry 304 may select a set of recommended content items for one of the first plurality of content categories. For example, control circuitry 304 may make an API call to recommendation engine 460 via network 404 (e.g., where the category and user history are used as function inputs). The control circuitry 304 may receive content item recommendation for the requested content category from the recommendation engine. At 508 , control circuitry 304 checks if there are more categories in the first plurality of content categories. If so, control circuitry 304 proceeds back to step 506 and repeats that step for the next category of the first plurality of content categories. Newly received recommended content items may be concentrated into a first set of recommended content items.
- control circuitry 304 proceeds to step 510 .
- control circuitry 304 may reduce the first set of recommended content items by removing at least one content item that also belongs to the set of content items that are likely to be requested in the future (e.g., as determined in step 506 ). For example, some content items may be so likely to be requested as to present no value for training the neural net. For example, if the user watches a news program every day at 9:00PM, there is no value in recommending such programming to the user.
- control circuitry 304 may increase a reward score for the first plurality of content categories. For example, the score may be increased by 1.
- control circuitry 304 performs an additional check to determine whether the requested content item is included in the set of content items that are likely to be requested in the future, as identified at step 502 . If so, control circuitry 304 proceeds to 518 where control circuitry 304 decreases the reward score for the first plurality of content categories (e.g., by 1 or by 0.5). Otherwise, step 518 is skipped and process 500 proceeds to step 520 .
- steps 510 - 518 may be repeated multiple times over a certain time period (e.g., 10 days) every time a request for content item is received. In this way, the reward score may be adjusted during the entire time period.
- the final reward score (e.g., at the end of the time period) is used in step 520 to train the category selection neural net (e.g., neural net 204 ) to reinforce reward score maximization.
- the category selection neural net e.g., neural net 204
- a reward is score is calculated for each plurality. Each such reward score may be used to the train the neural net as described in relation to FIG. 2 . Because the neural net training process does not require actually showing the recommendations to the user, the speed at which the neural net is trained is greatly increased.
- control circuitry 304 uses the trained neural net to select a second plurality of content categories (e.g., in the same way as described in regard to step 504 ).
- control circuitry 304 may select a second set of recommended content items for the second plurality of content categories (e.g., by making appropriate API calls to the recommendation engine 450 ).
- control circuitry 304 may choose to repeat step 524 if more categories remand for which the content items needs to be selected; otherwise control circuitry 304 may proceed to block 528 .
- control circuitry 304 may generate for display identifiers for recommended content items of the second set of recommended content items. For example, control circuitry 304 may generate an explorative display where the recommended content items are identified by titles and arranged using horizontal category carousels (e.g., as arranged in table 112 ). In some embodiments, control circuitry 304 may generate an explorative display where the recommended content items are identified by titles and arranged in a list. In other embodiments, control circuitry 304 may generate an explorative display as part of an EPG screen (e.g., an EPG screen where the recommended items are highlighted).
- EPG screen e.g., an EPG screen where the recommended items are highlighted.
- FIG. 6 is a flowchart of an illustrative process of training a neural network, in accordance with some embodiments of the disclosure.
- each step of process 600 can be performed by media device 402 (e.g., via control circuitry 304 ) or any of the system components shown in FIG. 3 .
- each step of process 600 can be performed by processing server 408 (e.g., via control circuitry 304 of processing server 408 ).
- each step of process 600 may be performed by control circuitry 304 executing instructions of content recommendation application based on instructions stored in memory 308 (e.g., user of media device 402 ).
- steps of process 600 are performed in addition to steps of process 500 .
- a further goal of the content recommendation application may be to present content item recommendations that are not homogenous.
- control circuitry 304 may perform steps of process 600 to train the content category selection neural net differently. For example, reward scores may be calculated differently to encourage selection of non-homogenous content item recommendations.
- control circuitry 304 may receive a request from the user for a requested content item (e.g., a request to watch or record the content item) that is included in the first set of recommended content items (selected at step 504 ). This “hit” may lead to an increased reward score (e.g., at step 514 ).
- control circuitry 304 may receive a subsequent request for a another content item that is also included in the first set of recommended content items.
- control circuitry 304 may check whether requested content items from steps 602 and 604 belong to different categories. If they do, control circuitry 304 may further increase the reward score for the first plurality of content categories at step 608 . Alternatively, if the requested content items from steps 602 and 604 are from the same category, control circuitry 304 may decrease the reward score at step 610 . Such a decrease in reward score will help to better train the neural net (e.g., neural net 204 ) to select categories that will lead to non-homogenous content item recommendations.
- neural net e.g., neural net 204
- FIG. 7 is a flowchart of an illustrative process of training a neural network, in accordance with some embodiments of the disclosure.
- each step of process 700 can be performed by content device 402 (e.g., via control circuitry 304 ) or any of the system components shown in FIG. 3 .
- each step of process 700 can be performed by processing server 408 (e.g., via control circuitry 304 of processing server 408 ).
- each step of process 600 may be performed by control circuitry 304 executing instructions of the content recommendation application based on instructions stored in memory 308 .
- steps of process 700 are performed in addition to steps of process 500 .
- a further goal of the content recommendation application may be to present content item recommendations that are not homogenous.
- control circuitry 304 may perform steps of process 700 to train the content category selection neural net differently. For example, reward scores may be calculated differently to encourage selection of non-homogenous content item recommendations.
- steps of process 700 are performed in addition to steps of process 500 .
- control circuitry 304 may receive a request from the user to access a content item from one of the categories in the first plurality of content categories (selected at step 504 .)
- control circuitry 304 may receive a request from the user to access (e.g., watch or record) a content item from another of the categories in the first plurality of content categories (selected at step 504 .)
- control circuitry 304 may check whether the category accessed in step 702 and the category accessed in step 704 are sufficiently similar. For example, control circuitry 304 may maintain a list of relatedness between content categories. For example, categories “horror” and “suspense” may be highly related, while categories “romance” and “education” may be deemed as not related. If the categories are not related, control circuitry 304 may further increase the reward score for the first plurality of content categories at step 708 . Alternatively, if the categories are related, control circuitry 304 may decrease the reward score at step 710 . Such a decrease in reward score will help to better train the neural net (e.g., neural net 204 ) to select categories that will lead to non-homogenous content item recommendations.
- neural net e.g., neural net 204
- processes 500 - 700 or any step thereof could be performed on, or provided by, any of the devices shown in FIGS. 1-4 .
- the processes may be executed by control circuitry 304 ( FIG. 3 ) as instructed by a content recommendation application.
- one or more steps of a process may be omitted, modified, and/or incorporated into or combined with one or more steps of any other process or embodiment (e.g., steps from process 500 may be combined with steps from processes 600 and 700 ).
- steps and descriptions described in relation to FIGS. 5-7 may be done in alternative orders or in parallel to further the purposes of this disclosure. For example, each of these steps may be performed in any order or in parallel or substantially simultaneously to reduce lag or increase the speed of the system or method.
- a content recommendation application may be a stand-alone application implemented on a media device or a server.
- the content recommendation application may be implemented as software or a set of executable instructions.
- the instructions for performing any of the embodiments discussed herein of the content recommendation application may be encoded on non-transitory computer-readable media (e.g., a hard drive, random-access memory on a DRAM integrated circuit, read-only memory on a BLU-RAY disk, etc.) or transitory computer-readable media (e.g., propagating signals carrying data and/or instructions).
- the instructions may be stored in storage 308 and executed by control circuitry 304 of a computer device 301 .
Landscapes
- Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Business, Economics & Management (AREA)
- General Physics & Mathematics (AREA)
- Accounting & Taxation (AREA)
- Finance (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Marketing (AREA)
- Development Economics (AREA)
- Economics (AREA)
- General Business, Economics & Management (AREA)
- Strategic Management (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
Claims (20)
Priority Applications (6)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/228,123 US10715869B1 (en) | 2018-12-20 | 2018-12-20 | Deep reinforcement learning for personalized screen content optimization |
| PCT/US2019/067973 WO2020132518A1 (en) | 2018-12-20 | 2019-12-20 | Deep reinforcement learning for personalized screen content optimization |
| US16/893,054 US11265609B2 (en) | 2018-12-20 | 2020-06-04 | Deep reinforcement learning for personalized screen content optimization |
| US17/581,490 US11706491B2 (en) | 2018-12-20 | 2022-01-21 | Deep reinforcement learning for personalized screen content optimization |
| US18/204,497 US12149789B2 (en) | 2018-12-20 | 2023-06-01 | Deep reinforcement learning for personalized screen content optimization |
| US18/917,798 US20250047944A1 (en) | 2018-12-20 | 2024-10-16 | Deep reinforcement learning for personalized screen content optimization |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/228,123 US10715869B1 (en) | 2018-12-20 | 2018-12-20 | Deep reinforcement learning for personalized screen content optimization |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/893,054 Continuation US11265609B2 (en) | 2018-12-20 | 2020-06-04 | Deep reinforcement learning for personalized screen content optimization |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20200204862A1 US20200204862A1 (en) | 2020-06-25 |
| US10715869B1 true US10715869B1 (en) | 2020-07-14 |
Family
ID=69187968
Family Applications (5)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/228,123 Active US10715869B1 (en) | 2018-12-20 | 2018-12-20 | Deep reinforcement learning for personalized screen content optimization |
| US16/893,054 Active US11265609B2 (en) | 2018-12-20 | 2020-06-04 | Deep reinforcement learning for personalized screen content optimization |
| US17/581,490 Active US11706491B2 (en) | 2018-12-20 | 2022-01-21 | Deep reinforcement learning for personalized screen content optimization |
| US18/204,497 Active US12149789B2 (en) | 2018-12-20 | 2023-06-01 | Deep reinforcement learning for personalized screen content optimization |
| US18/917,798 Pending US20250047944A1 (en) | 2018-12-20 | 2024-10-16 | Deep reinforcement learning for personalized screen content optimization |
Family Applications After (4)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/893,054 Active US11265609B2 (en) | 2018-12-20 | 2020-06-04 | Deep reinforcement learning for personalized screen content optimization |
| US17/581,490 Active US11706491B2 (en) | 2018-12-20 | 2022-01-21 | Deep reinforcement learning for personalized screen content optimization |
| US18/204,497 Active US12149789B2 (en) | 2018-12-20 | 2023-06-01 | Deep reinforcement learning for personalized screen content optimization |
| US18/917,798 Pending US20250047944A1 (en) | 2018-12-20 | 2024-10-16 | Deep reinforcement learning for personalized screen content optimization |
Country Status (2)
| Country | Link |
|---|---|
| US (5) | US10715869B1 (en) |
| WO (1) | WO2020132518A1 (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11170036B2 (en) * | 2019-06-27 | 2021-11-09 | Rovi Guides, Inc. | Methods and systems for personalized screen content optimization |
| US12149789B2 (en) | 2018-12-20 | 2024-11-19 | Adeia Guides Inc. | Deep reinforcement learning for personalized screen content optimization |
Families Citing this family (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11281734B2 (en) * | 2019-07-03 | 2022-03-22 | International Business Machines Corporation | Personalized recommender with limited data availability |
| US11397786B2 (en) * | 2019-12-12 | 2022-07-26 | Yahoo Assets Llc | Method and system of personalized blending for content recommendation |
| US11902622B2 (en) * | 2020-05-28 | 2024-02-13 | Comcast Cable Communications, Llc | Methods, systems, and apparatuses for determining viewership |
| US11623140B2 (en) * | 2020-07-01 | 2023-04-11 | Sony Interactive Entertainment LLC | High-speed save data storage for cloud gaming |
| CN111899728B (en) * | 2020-07-23 | 2024-05-28 | 海信电子科技(武汉)有限公司 | Training method and device for intelligent voice assistant decision strategy |
| US12505341B2 (en) * | 2021-01-06 | 2025-12-23 | Capital One Services, Llc | Methods and systems for dynamically selecting alternative content based on real-time events during device sessions using cross-channel, time-bound deep reinforcement machine learning |
| CN114741586B (en) * | 2021-01-08 | 2025-06-20 | 腾讯科技(深圳)有限公司 | Project recommendation method and device, project recommendation model, medium and electronic device |
| CN115407872B (en) | 2022-08-12 | 2023-10-13 | 北京津发科技股份有限公司 | Evaluation method, device and storage medium for intelligent man-machine cooperative system |
| US12307338B2 (en) | 2022-12-31 | 2025-05-20 | Maplebear Inc. | Counterfactual policy evaluation of model performance |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20150052561A1 (en) | 2011-08-24 | 2015-02-19 | Inview Technology Limited | Audiovisual content recommendation method and device |
| US20160379123A1 (en) * | 2015-06-24 | 2016-12-29 | Comcast Cable Communications, Llc | Entertainment Prediction Favorites |
| US20170228385A1 (en) * | 2016-02-08 | 2017-08-10 | Hulu, LLC | Generation of Video Recommendations Using Connection Networks |
| US20180075137A1 (en) | 2016-09-09 | 2018-03-15 | Yandex Europe Ag | Method and apparatus for training a machine learning algorithm (mla) for generating a content recommendation in a recommendation system and method and apparatus for generating the recommended content using the mla |
| US20180225710A1 (en) * | 2017-02-03 | 2018-08-09 | Adobe Systems Incorporated | User segment identification based on similarity in content consumption |
| US20190327139A1 (en) * | 2018-04-20 | 2019-10-24 | Spotify Ab | Methods and systems for provisioning settings of a media playback device |
| US20190332920A1 (en) * | 2016-11-04 | 2019-10-31 | Google Llc | Unsupervised detection of intermediate reinforcement learning goals |
Family Cites Families (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6388714B1 (en) | 1995-10-02 | 2002-05-14 | Starsight Telecast Inc | Interactive computer system for providing television schedule information |
| US6177931B1 (en) | 1996-12-19 | 2001-01-23 | Index Systems, Inc. | Systems and methods for displaying and recording control interface with television programs, video, advertising information and program scheduling information |
| US7461892B2 (en) | 2003-12-01 | 2008-12-09 | W.E.T. Automotive Systems, A.C. | Valve layer for a seat |
| US8613024B2 (en) | 2005-12-13 | 2013-12-17 | United Video Properties, Inc. | Cross-platform predictive popularity ratings for use in interactive television applications |
| US8239412B2 (en) | 2010-05-05 | 2012-08-07 | Rovi Technologies Corporation | Recommending a media item by using audio content from a seed media item |
| US10438114B1 (en) * | 2014-08-07 | 2019-10-08 | Deepmind Technologies Limited | Recommending content using neural networks |
| US10060929B2 (en) | 2015-01-23 | 2018-08-28 | Purdue Research Foundation | Difference gel electrophoresis of phosphoproteome |
| US10423979B2 (en) | 2016-12-29 | 2019-09-24 | Rovi Guides, Inc. | Systems and methods for a framework for generating predictive models for media planning |
| US10715869B1 (en) | 2018-12-20 | 2020-07-14 | Rovi Guides, Inc. | Deep reinforcement learning for personalized screen content optimization |
-
2018
- 2018-12-20 US US16/228,123 patent/US10715869B1/en active Active
-
2019
- 2019-12-20 WO PCT/US2019/067973 patent/WO2020132518A1/en not_active Ceased
-
2020
- 2020-06-04 US US16/893,054 patent/US11265609B2/en active Active
-
2022
- 2022-01-21 US US17/581,490 patent/US11706491B2/en active Active
-
2023
- 2023-06-01 US US18/204,497 patent/US12149789B2/en active Active
-
2024
- 2024-10-16 US US18/917,798 patent/US20250047944A1/en active Pending
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20150052561A1 (en) | 2011-08-24 | 2015-02-19 | Inview Technology Limited | Audiovisual content recommendation method and device |
| US20160379123A1 (en) * | 2015-06-24 | 2016-12-29 | Comcast Cable Communications, Llc | Entertainment Prediction Favorites |
| US20170228385A1 (en) * | 2016-02-08 | 2017-08-10 | Hulu, LLC | Generation of Video Recommendations Using Connection Networks |
| US20180075137A1 (en) | 2016-09-09 | 2018-03-15 | Yandex Europe Ag | Method and apparatus for training a machine learning algorithm (mla) for generating a content recommendation in a recommendation system and method and apparatus for generating the recommended content using the mla |
| US20190332920A1 (en) * | 2016-11-04 | 2019-10-31 | Google Llc | Unsupervised detection of intermediate reinforcement learning goals |
| US20180225710A1 (en) * | 2017-02-03 | 2018-08-09 | Adobe Systems Incorporated | User segment identification based on similarity in content consumption |
| US20190327139A1 (en) * | 2018-04-20 | 2019-10-24 | Spotify Ab | Methods and systems for provisioning settings of a media playback device |
Non-Patent Citations (7)
| Title |
|---|
| Anonymous: Application programming interface—Wikipedia Retrieved from the Internet: URL:https://en.wikipedia.org/w/index.php?title=Application_programming_interface&old_id=836341145 Apr. 14, 2018 (Apr. 14, 2018), XP055615138. |
| Fakhfakh et al., "Deep learning-based recommendation: current issues and challenges," International Journal of Advanced Computer Science and Applications, vol. 8, No. 12, pp. 59-68 (2017). |
| International Search Report and Written Opinion of PCT/US2019/067973 dated Feb. 27, 2020. |
| Kunaver et al., "Diversity in recommender system—A Survey," Knowledge-Based Systems, vol. 123, pp. 154-162 (2017). |
| Nilashi et al., "A new method for collaborative filtering recommender systems: The case of Yahoo! Movies and TipAdvisor datasets," An International Journal JSCDSS, 3(5):44-46 (2016). |
| Qusai Shambour, "A user-based multi-criteria recommendation approach for personalized recommendations," Int'l J. of Computer Science and Information Security, 14(12):657-663 (2016). |
| Sergey Levin, "Actor-Critic Algorithms, CS 294-112: Deep Reinforcement Learning," http://rail.eecs.berkeley.edu/deeprlcourse-fa17/f17docs/lecture_5_actor_critic_pdf.pdf Nov. 14, 2018 (28 pages). |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US12149789B2 (en) | 2018-12-20 | 2024-11-19 | Adeia Guides Inc. | Deep reinforcement learning for personalized screen content optimization |
| US11170036B2 (en) * | 2019-06-27 | 2021-11-09 | Rovi Guides, Inc. | Methods and systems for personalized screen content optimization |
| US11741148B2 (en) | 2019-06-27 | 2023-08-29 | Rovi Guides, Inc. | Methods and systems for personalized screen content optimization |
Also Published As
| Publication number | Publication date |
|---|---|
| US20200204862A1 (en) | 2020-06-25 |
| WO2020132518A1 (en) | 2020-06-25 |
| US20250047944A1 (en) | 2025-02-06 |
| US11265609B2 (en) | 2022-03-01 |
| US20220150589A1 (en) | 2022-05-12 |
| US20230319357A1 (en) | 2023-10-05 |
| US11706491B2 (en) | 2023-07-18 |
| US20200304873A1 (en) | 2020-09-24 |
| US12149789B2 (en) | 2024-11-19 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12149789B2 (en) | Deep reinforcement learning for personalized screen content optimization | |
| US20240202213A1 (en) | Feedback loop content recommendation | |
| KR102254579B1 (en) | System and method for streaming personalized media content | |
| US12265897B2 (en) | Systems and methods for automated content curation using signature analysis | |
| US20230345077A1 (en) | Churn analysis and methods of intervention | |
| US11657080B2 (en) | Methods and systems for generating and presenting content recommendations for new users | |
| US20240251135A1 (en) | Systems and methods for providing content recommendations | |
| US11741148B2 (en) | Methods and systems for personalized screen content optimization | |
| US20140149326A1 (en) | Post-processed content recommendation | |
| US20250203141A1 (en) | Parameter modification for selecting personalized item sets | |
| US11238287B2 (en) | Systems and methods for automated content curation using signature analysis | |
| US12430569B2 (en) | Systems and methods for providing media content recommendations | |
| US20180139501A1 (en) | Optimized delivery of sequential content by skipping redundant segments | |
| US20140074858A1 (en) | Percent-consumed weighted content recommendation | |
| CN116137677A (en) | Topic recommendation method for smart TV, computer equipment and readable storage medium | |
| US20260037405A1 (en) | Supplemental Content Load Tuning | |
| US20250234070A1 (en) | Content sharing platform channel review using a virtual assistant |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: HPS INVESTMENT PARTNERS, LLC, AS COLLATERAL AGENT, NEW YORK Free format text: SECURITY INTEREST;ASSIGNORS:ROVI SOLUTIONS CORPORATION;ROVI TECHNOLOGIES CORPORATION;ROVI GUIDES, INC.;AND OTHERS;REEL/FRAME:051143/0468 Effective date: 20191122 |
|
| AS | Assignment |
Owner name: MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT, MARYLAND Free format text: PATENT SECURITY AGREEMENT;ASSIGNORS:ROVI SOLUTIONS CORPORATION;ROVI TECHNOLOGIES CORPORATION;ROVI GUIDES, INC.;AND OTHERS;REEL/FRAME:051110/0006 Effective date: 20191122 |
|
| AS | Assignment |
Owner name: ROVI GUIDES, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MILLER, KYLE;REEL/FRAME:051346/0128 Effective date: 20191219 |
|
| AS | Assignment |
Owner name: BANK OF AMERICA, N.A., NORTH CAROLINA Free format text: SECURITY INTEREST;ASSIGNORS:ROVI SOLUTIONS CORPORATION;ROVI TECHNOLOGIES CORPORATION;ROVI GUIDES, INC.;AND OTHERS;REEL/FRAME:053468/0001 Effective date: 20200601 |
|
| AS | Assignment |
Owner name: VEVEO, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:HPS INVESTMENT PARTNERS, LLC;REEL/FRAME:053458/0749 Effective date: 20200601 Owner name: ROVI GUIDES, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:HPS INVESTMENT PARTNERS, LLC;REEL/FRAME:053458/0749 Effective date: 20200601 Owner name: TIVO SOLUTIONS, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:HPS INVESTMENT PARTNERS, LLC;REEL/FRAME:053458/0749 Effective date: 20200601 Owner name: ROVI SOLUTIONS CORPORATION, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:HPS INVESTMENT PARTNERS, LLC;REEL/FRAME:053458/0749 Effective date: 20200601 Owner name: ROVI TECHNOLOGIES CORPORATION, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:HPS INVESTMENT PARTNERS, LLC;REEL/FRAME:053458/0749 Effective date: 20200601 Owner name: ROVI SOLUTIONS CORPORATION, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:053481/0790 Effective date: 20200601 Owner name: ROVI GUIDES, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:053481/0790 Effective date: 20200601 Owner name: ROVI TECHNOLOGIES CORPORATION, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:053481/0790 Effective date: 20200601 Owner name: VEVEO, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:053481/0790 Effective date: 20200601 Owner name: TIVO SOLUTIONS, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:053481/0790 Effective date: 20200601 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
| AS | Assignment |
Owner name: ADEIA GUIDES INC., CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:ROVI GUIDES, INC.;REEL/FRAME:069003/0960 Effective date: 20220815 |