|
| 1 | +# Model Update |
| 2 | + |
| 3 | +In a real-world setting, it is crucial for these systems to routinely |
| 4 | +update their model parameters. However, maintaining seamless user |
| 5 | +experience while concurrently updating ultra-large-scale recommendation |
| 6 | +models becomes a daunting task, especially when catering to a large |
| 7 | +number of online users. This section primarily addresses why continuous |
| 8 | +updates to model parameters are necessary for recommender systems, |
| 9 | +before delving into the offline update methodology and a recommender |
| 10 | +system capable of online updates. |
| 11 | + |
| 12 | +## Necessity for Model Updates |
| 13 | + |
| 14 | +Recommendation models employed in online services encounter highly |
| 15 | +dynamic factors in their datasets: |
| 16 | + |
| 17 | +1. **New users and items**: Both the users targeted by a recommendation |
| 18 | + model and the items encompassed within the model are in constant |
| 19 | + flux. New users and items can emerge at any given time. As depicted |
| 20 | + in Figure [1](#fig:embedding missing){reference-type="ref" |
| 21 | + reference="fig:embedding missing"}, the recommender model will be |
| 22 | + unable to cater to a new user if the embedding table lacks the |
| 23 | + corresponding embedding item. Similarly, if a newly added item is |
| 24 | + absent from the model's embedding table, it will not surface in the |
| 25 | + recommendation pipeline, rendering it impossible to recommend to the |
| 26 | + intended users. |
| 27 | + |
| 28 | +2. **Shifting user interests**: Recommendation models must adapt to |
| 29 | + continually shifting user interests. Should a model fail to swiftly |
| 30 | + adjust its weights to mirror evolving user interests, the quality of |
| 31 | + its recommendations may suffer. As an example, in a news |
| 32 | + recommendation application, trending news topics alter daily. If the |
| 33 | + model consistently suggests outdated trends, the click-through rate |
| 34 | + (CTR) of users will progressively decline. |
| 35 | + |
| 36 | +<figure id="fig:embedding missing"> |
| 37 | +<p> </p> |
| 38 | +<figcaption>Examples of Missing Embedding Items</figcaption> |
| 39 | +</figure> |
| 40 | + |
| 41 | +While empirical rule formulation (such as directly incorporating new or |
| 42 | +statistically popular items into the recommendation results) can |
| 43 | +potentially address the above dynamic factors, it only provides a |
| 44 | +temporary and partial solution. Such empirical methods cannot entirely |
| 45 | +rectify the problems, mainly due to the substantial performance gap |
| 46 | +between learning-based recommendation models and experience-based |
| 47 | +recommendation rules. |
| 48 | + |
| 49 | +## Offline Process of Model Update |
| 50 | + |
| 51 | +To keep up with the evolving dynamics of datasets, recommender systems |
| 52 | +need to consistently accumulate new training data and harness this |
| 53 | +information to refine their models. This process, known as model |
| 54 | +updating, is integral to improving the efficacy of these systems. |
| 55 | +Typically, the trained model is saved as a checkpoint, which is then put |
| 56 | +through an offline validation process. This updated model is only |
| 57 | +deployed onto the inference servers if it can effectively enhance the |
| 58 | +quality of recommendation. A depiction of this offline model update |
| 59 | +process can be seen in Figure |
| 60 | +[2](#fig:offline update){reference-type="ref" |
| 61 | +reference="fig:offline update"}. |
| 62 | + |
| 63 | + |
| 64 | +:label:`offline` |
| 65 | + |
| 66 | +To break it down, when a model completes a training period, the |
| 67 | +subsequent steps unfold: |
| 68 | + |
| 69 | +1. **Checkpoint step**: The model checkpoint is retrieved from the |
| 70 | + parameter server in the training data center and stored on the disk. |
| 71 | + |
| 72 | +2. **Validation step**: This checkpoint undergoes validation against an |
| 73 | + offline dataset. If it does not pass this validation, the model |
| 74 | + continues its training. |
| 75 | + |
| 76 | +3. **Dissemination step**: If the model checkpoint passes the offline |
| 77 | + validation, it is then disseminated to the parameter servers which |
| 78 | + host the replicas of this model in different inference data centers. |
| 79 | + |
| 80 | +## Needs for Low-latency Model Updates |
| 81 | + |
| 82 | +Using an offline process of updating models described above can vary in |
| 83 | +duration, ranging from several minutes to hours. Nevertheless, some |
| 84 | +systems have streamlined the procedures for storing and propagating |
| 85 | +checkpoints, which allows updates to occur on a minute-by-minute basis. |
| 86 | +Despite this, the existing minute-level latency in model updates still |
| 87 | +falls short of the requirements for several key scenarios that involve |
| 88 | +recommender systems. |
| 89 | + |
| 90 | +### Incorporating Fresh Content |
| 91 | + |
| 92 | +There are needs in which applications prioritize the delivery of fresh |
| 93 | +content. For example, in the context of short video recommendations, a |
| 94 | +content creator might generate videos relevant to the latest trending |
| 95 | +topics. If these videos are not recommended in a timely manner, the |
| 96 | +topics may become obsolete, resulting in a lower-than-anticipated number |
| 97 | +of views. |
| 98 | + |
| 99 | +### Servicing Anonymous Users |
| 100 | + |
| 101 | +There are often needs where user features are either unavailable or |
| 102 | +scarce. Users are increasingly opting for anonymous usage of |
| 103 | +applications and sharing minimal data due to heightened privacy concerns |
| 104 | +(e.g., Internet browsers prohibit recommender systems from collecting |
| 105 | +user cookies) and stricter data protection regulations (e.g., the |
| 106 | +General Data Protection Regulation -- GDPR -- in Europe). As a result, |
| 107 | +recommender systems are compelled to learn user interests online within |
| 108 | +a very narrow timeframe. |
| 109 | + |
| 110 | +### Adopting Online Machine Learning Techniques |
| 111 | + |
| 112 | +There are needs that call for the application of online machine learning |
| 113 | +techniques. Traditional recommender systems often employ offline |
| 114 | +training where data gathered over a specified timeframe (e.g., one day) |
| 115 | +is used to train a model, which is subsequently deployed online during |
| 116 | +off-peak periods (e.g., early morning). However, recent research and |
| 117 | +practice suggest that increasing the training frequency can notably |
| 118 | +improve the quality of recommendations. The logical endpoint of |
| 119 | +increasing training frequency is online training, where data is |
| 120 | +processed in a streaming fashion and fed to the model. The model then |
| 121 | +continually fine-tunes its parameters based on these online samples. |
0 commit comments