Next Article in Journal
Cross-Layer Influence of Multiple Network Embedding on Venture Capital Networks in China: An ERGM-Based Analysis
Previous Article in Journal
Machine Learning-Powered Dynamic Fleet Routing Towards Real-Time Fuel Economy with Smart Weight Sensing and Intelligent Traffic Reasoning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Capturing Dynamic User Preferences: A Recommendation System Model with Non-Linear Forgetting and Evolving Topics

1
School of Management, Nanjing University of Posts and Telecommunications, Nanjing, 210003, China
2
Research Center for Education Artificial Intelligence, Nanjing University of Posts and Telecommunications, Nanjing, 210023, China
3
School of Information Management, Nanjing University, Nanjing, 210023, China
4
School of Computer Science, Nanjing Audit University, Nanjing, 210023, China
*
Authors to whom correspondence should be addressed.
Systems 2025, 13(11), 1034; https://doi.org/10.3390/systems13111034
Submission received: 31 August 2025 / Revised: 1 November 2025 / Accepted: 14 November 2025 / Published: 19 November 2025

Abstract

Though recommendation systems can help users save time while shopping online, their performance is significantly limited by sparse user data and the inability to capture temporal dynamics of user preferences, such as interest forgetting and topic evolution in reviews. Existing studies primarily focus on static user–item interactions or partial temporal signals (e.g., rating timestamps) but fail to comprehensively model two critical aspects: the non-linear decay of user interests over time, where users gradually forget historical preferences, and the semantic evolution of review topics, which reflects implicit shifts in user preferences across different periods. To address these limitations, we propose a Temporal Dynamic Latent Review-aware Preference Model with Matrix Factorization. Our model integrates an adaptive forgetting-weight function to simulate users’ interest decay and a multi-interval latent topic model to extract evolving preference features from review semantics. Specifically, we design a joint optimization framework that dynamically weights user ratings based on temporal forgetting patterns and decomposes review texts into latent topic factors to alleviate data sparsity. Finally, the experiments employ five baseline methods on six datasets to test the recommendation performance, validating its effectiveness in tracking users’ temporal interest drift and improving recommendation accuracy.

1. Introduction

With the rapid growth of e-commerce platforms, temporal recommendation systems face two critical challenges. The first is sparsity of user–item interactions, in that users’ explicit feedback (e.g., ratings) is often limited, leading to inaccurate preference modeling [1,2]. The second is dynamic preference evolution, which means users’ interests exhibit non-linear decay (e.g., forgetting historical preferences) and semantic shifts (e.g., changes in review topics over time). The existing studies suffer from three key limitations. First, most temporal recommendation methods primarily rely on rating timestamps but ignore non-linear interest forgetting [3,4]. For instance, they assume linear decay of user preferences, which contradicts psychological studies on human memory. Second, review-based approaches extract static topic features from reviews but fail to capture temporal topic dynamics [5]. For example, a user’s focus on “battery life” in electronics reviews may shift to “design aesthetics” over months, yet existing models treat all reviews as equally relevant. Third, hybrid methods attempt to combine ratings and reviews but lack a unified framework to jointly optimize temporal decay and topic evolution. This results in suboptimal performance under data sparsity [6,7].
Although previous studies have separately explored temporal decay or comment topic modeling, no work has simultaneously modeled nonlinear interest decay and the semantic evolution of multi-period comment topics within a unified optimization framework. To overcome these limitations, the proposed TDLRP-MF model not only integrates these two types of dynamic signals but also achieves fine-grained perception of user interest drift through joint matrix factorization and latent topic word confidence modeling. Our model introduces three contributions:
  • We proposed a novel multi-feature adaptive time weighting approach. It comprehensively considers the impact of user interest drift features on recommendation generation to enhance the accuracy of user preference prediction.
  • We constructed a novel latent topic model for multi-interval individual topic reviews. The model can more accurately perceive hot topics in user reviews at different periods and capture the dynamic evolution of user preferences in time steps.
  • We built a feature fusion model of multi-dimensional latent factors. It can learn user ratings’ synergistic temporal evolution features and multiple latent preference features of the topics to achieve high-accuracy user preference perception in data-sparse scenarios.

2. Related Work

2.1. Recommendation Based on Temporal Matrix Factorization

Traditional recommendation methods can be recommended by collaborative filtering [8], but with the rapid increase in the variety and number of products, the sparsity of data problem produces trouble, so people consider the time factor of the matrix decomposition of the matrix dimensionality reduction to improve the data sparsity problem. TimeSVD++ captures the evolving nature of user preferences by probability matrix [9]. This approach introduces time-evolving discrepancies for people and goods within the latent space. However, it overlooks the interconnections among shifts in user preferences at specific times, potentially missing crucial time-specific preference aspects. Moreover, it lacks consideration for auxiliary data, reducing its effectiveness in scenarios with sparse data. BPTF employs Bayesian probabilistic tensor decomposition to estimate ratings [10]. It offers insights into the evolution of latent features and ratings by utilizing the latent vectors. Nonetheless, it might not be sufficiently sensitive to detect subtle, localized variations in user preferences. TCMF is designed to address data sparsity by incorporating temporal dynamics and multimodal data [11]. However, it heavily relies on implicit feedback from user reviews and overlooks the implicit implications present in supplementary data. In TMRevCo, temporal dynamics, and auxiliary data are considered to improve performance [12]. This is achieved by using item relevance measures in the auxiliary factors and correlating them with the matrix decomposition item factors. While this method considers user preferences across different periods, it does not consider the potential effects of varying time stages on users. This could result in significant deviations between predicted and actual user behavior, ultimately reducing the effectiveness of recommendations.

2.2. Recommendation Based on Deep Learning with Interest Drift

In addition to matrix decomposition methods, deep learning is also an important branch of research in the field of recommendation, and in recent years, a large number of deep learning-based methods have emerged to mine the evolutionary characteristics of user preferences. HITUCF is a hybrid recommendation algorithm that integrates similarity and forgetting curves [13]. While the method partially emulates the forgetting patterns of human interests, it does not account for the temporal evolution of user review information. FSTS aims to address user interest drift, low accuracy, and long tail issues by extracting item features using stability variables and time-sensitive factors [14]. It also uses feature vectors and time-sensitive factors in the prediction process to enhance recommendation performance with minimal time complexity. However, its time-sensitive factor may not be sufficient to capture dynamic shifts in user interest accurately. TSCIA is a temporal collaborative interest awareness model that explicitly considers similar interests across adjacent sessions to model overall collaborative interests [15]. It addresses the issue of user interest drift and captures patterns of user preference changes more accurately. However, it does not take into account the time-varying characteristics of items. ADARIFT uses personalized learning rates to handle user interest drift and incorporates item features in adaptive recommender systems [16]. However, it does not utilize auxiliary information. ALSTM is a sentiment rating model utilizing long and short-term memory that addresses data sparsity and imbalance by combining sentiment bias in user ratings and reviews. While it effectively models the drift in user and item features, it does not delve into mining latent semantics embedded in review texts [17]. Additionally, some scholars have considered the influence of forgetting laws on item selection by studying the forgetting characteristics of users’ interests. TPNE is a module with temporal positive and negative incentive modeling built through the static-dynamic interest learning framework for more accurate sequential recommendations [18]. TDADLFM is a time decay adaptive latent factor model, which assumes that people’s preferences usually change over time [19]. It adds an attention mechanism to adjust the user’s embedding vectors dynamically and combines a time decay function with an attention model for item score prediction. Although these methods take into account user interest drift, it does not consider the interaction between user comments and ratings, which may affect the prediction of user preference evolution.

2.3. Recommendation Based on Review Mining

Some studies have been conducted to identify user preference features for items by inferring user interest changes from semantic features of textual reviews. These reviews serve as a reference information source for recommender systems. ADR is a recommendation approach focusing on deep reviews, supplementing item representation with auxiliary vectors, and performing user-side modeling based on a user attention mechanism [20]. MRMRP is a multivariate review-based rating prediction model. It can extract valuable features from the supplemental reviews to enhance recommendation performance using deep learning methods [21]. ERP is an enhanced review-based rating prediction model that distinguishes the impact of reviews generated by different users through user influence. It also uses auxiliary and review information to learn item-perceived user preferences [22]. LUAR is a method that employs a neural attention mechanism to automatically identify and prioritize crucial auxiliary reviews in the dataset [23]. PSAR is a persona-driven sentiment-focused recommendation model that aims to learn segment-level and sequence-level personalized sentiment representations from user reviews [24]. DARMH is a multi-task dual-attention recommendation model that leverages reviews and their usefulness [25]. The method uses interactive and local attention mechanisms separately to capture personalized preferences for specific items by specific users. DeepCGSR is a deep recommendation model that combines textual review sentiment and rating matrices to conduct cross-granularity sentiment analysis by treating the set of user reviews of items as a corpus [26]. This allows the method to gain sentiment feature vectors of both users and items by associating fine-grained and coarse-grained level information. ERUR is a review-based user representation model based on learning social graphs, which enhances user modeling by integrating user reviews and social relationships to achieve better recommendation results [27]. Prior research on recommendation systems that employ review topics has shown that considering various factors, such as semantics and context of user reviews, can enhance recommendation effectiveness. However, these studies do not explore the temporal dynamics of the recommendation process, which may lead to the loss of temporal interest drift characteristics of users. MTUPD is a method that incorporates multiple leaps of user interest drift [28]. The process utilizes multiple leap factors and adaptive temporal weights to gauge the relevance of user preferences across various periods using a forgetting curve function. However, this method only considers the topic model with non-negative matrix decomposition and does not explain the user review topic evolution law at a deeper level.

3. Preliminary

The paper introduces a temporal recommendation model with a latent factor that utilizes users’ ratings and reviews. In this section, we will discuss the initial steps and define and describe the fundamental concepts of the model. Additionally, the main symbols and parameters used in this paper are listed in Table 1.

3.1. Problem Promotion

User rating preference features change dynamically over time and show forgetfulness or indifference to specific item interests over time, while users also do reviews based on their interest points, which can contain richer topic preference features than considering only rating information. In contrast, conventional collaborative filtering recommendation methods struggle to thoroughly gauge how changes in users’ various interests affect the recommendation system. Hence, this paper aims to closely mimic the temporal changes in users’ multiple features, striving to enhance the system’s recommendation accuracy. The proposed recommendation model is illustrated in Figure 1.
Let t denote the current period requiring the recommendation, while t T represents a historical period relative to t, determined by the frequency of recommendation updates on the social media platform. To illustrate, in a total span of 16 months with T = 6 , if t corresponds to the 16th month, then t T pertains to the timeframe spanning from the 10th to the 15th month before t. We give the function definitions as follows:

3.1.1. Adaptive Temporal Forgetting-Weight

Ebbinghaus demonstrates that human forgetting does not decrease linearly, and the rate of forgetting is typically faster at first and then slows down. While various methods are used in previous studies to simulate human forgetting patterns [29,30], in reality, people are often affected by multiple factors that influence their perception and forgetting process. Thus, this paper aims to develop new forgetting weights based on memory and forgetting laws, considering the impact of real-life environments on individuals’ forgetting and simulating the relationship between users’ interest in items over time. From this, we propose a new time-weighted formulation that is defined as shown in Equation (1).
H t T ( r l i j ) = μ e δ T + ε
The r l i j denotes the viewing relationship of the ith user’s jth item, the t be the time at which user i rates item j, and the t-moment rating time weight of user i on item j. The fitting parameters μ , δ , and  ε are used to calculate the memory residual rate H t T ( r l i j ) for an item from the t T moment to the t moment. A higher value of the H function indicates that users have a stronger memory for the item’s content. The main focus of the forgetting function in this paper is to accurately reflect the change in users’ interest in the item. To simulate the forgetting characteristics of users’ interest, the forgetting curve tracing points of Ebbinghaus were selected as (0, 100), (1/3, 58), (1, 44), (9, 36), (24, 33), (48, 28), (144, 25), and (744, 21). The horizontal coordinates of these data points represent time in hours, while the vertical coordinates represent the memory retention rate in percentage. Using MATLAB 2024a’s Curve Fitting Tool, the data was fitted according to the definition of the H function. The most recent item rating is then used to represent the user’s current interest. After normalizing the fitting results, the final interest forgetting weighting function is defined as shown in Equation (2):
H t T ( r l i j ) = μ e δ 1 T | t m a x t 0 | + ε
where μ = 0.6509 , δ = 0.8023 , ε = 0.3481 , t 0 denotes the most recent evaluation time of the item, t m a x denotes the longest evaluation time of the item by the user, and T denotes the time interval between the current moment and the initial viewing moment.
Users often retain a degree of memory regarding items due to their enduring preferences [31]. For instance, if users show consistent interest in items with literary qualities, they are likely to continuously seek out such items over time. However, solely relying on the weight of interest and forgetting an item might diminish the lasting impact of specific features on sustained interest. Therefore, this article integrates the temporal progression of user preferences for item reviews across various time frames, as depicted in Equation (3):
s i t T = p = 1 P i ( t ) q = 1 P j ( t T ) Sim w v r l p , w v r l q P i ( t ) P j ( t T )
where P i ( t ) and P j ( t T ) represent the sets of items preferred by users at moments t and t T , respectively. The symbol ⊝ denotes the set containing elements unique to each set and does not overlap. Sim ( w v r l p , w v r l q ) indicates the similarity between the word vectors of user reviews w v r l p and w v r l q . To eliminate the influence of deactivated words, the initial step involves utilizing the deactivated word lexicon from the text. Subsequently, the words are tokenized using Jieba. Finally, Word2vec is employed to train the word vectors and determine their similarity, as demonstrated in Equations (4) and (5):
θ = w v r l p w v r l q | | w v r l p | | × | | w v r l q | |
S i m ( V r l p , V r l q ) = φ ( a w v θ + b w v )
The word vector cosine similarity is denoted by θ , the logistic sigmoid function is denoted by φ , and the weight matrix and bias vector are denoted by a w v and b w v , respectively. The review weights s i t T reflect the stability of the evolution of user interest in item review topics over time for user i. Larger values of s i t T indicate more stable changes in topic interest over time.
Through the integration of forgetting weights and the evolution of review weights, the ultimate model can be depicted using the temporal weights as indicated in Equation (6):
W t T = α H t T + 1 α s t T
Here, α signifies the weight balance coefficient employed for self-adjustment to modulate the influence of various weights.

3.1.2. User-Item Eigen Decomposition

The original NMF algorithm only considers the decomposition of static data, but in reality, the user’s preference may keep changing over time, so this paper considers each user’s latent factor dynamic change characteristics for improvement.
Initially, an auxiliary matrix is formulated to establish connections between the latent factors of users who have transitioned across various periods. Subsequently, temporal weights are introduced to accommodate the diminishing impact of factor matrix weights over time. This aims to capture the temporal changes in user preferences more accurately and enhance the precision of predicting evolving dynamic preferences. Moreover, we explore integrating an influence factor that signifies the temporal effect on a user’s characteristic factor matrix at different time points, contingent upon the duration of the time interval T up to moment t. Building upon the concept above and integrating it with the decomposition paradigm by Lee [32], the formula for temporal drift matrix decomposition is defined as presented in Equation (7):
R ( t ) U R ( t ) V R ( t ) R ( t ) U R ( t T ) D R ( t T ) V R ( t )
The dimension τ is set to 2, meaning that time is divided into two parts: the current moment t and the past moment t T . The transformation matrix D represents the relationship between these two moments, showing how the current matrix U t is transformed into the past matrix U t T .

4. Methodology

4.1. Temporal Latent Rating Matrix Prediction Model

The objective function is framed as a minimization problem to minimize the variance between rating prediction matrices and R ( t ) . This approach assesses the efficacy of the dimensionality reduction transformation applied to the rating matrix, as depicted in Equations (8) and (9):
arg min U R ( t ) , V R ( t ) R ( t ) U R ( t ) V R ( t ) F 2
arg min U R ( t T ) , V R ( t ) R ( t T ) U R ( t T ) V R ( t ) F 2
Through the correlation of Equations (7)–(9), while considering the duration of the time interval up to time T, we can formulate the objective minimization relationship as expressed in Equation (10):
arg min U , Q , V R ( t ) γ U R ( t ) V R ( t ) F 2 + R ( t ) ( 1 γ ) U R ( t T ) D R ( t T ) V R ( t ) F 2
s . t . U R ( t ) , U R ( t T ) , D R ( t T ) , V R ( t ) 0
The γ ( 0 < γ < 1 ) represents the time regularisation factor, accounting for varying time influences on user preferences.

4.2. The Latent Topic Model

To delve deeper into unexplored user preference topics using the rating model, this section introduces a latent topic model to describe the potential association between user reviews and topic words. This model will be used to construct a joint optimization model for rating and topic decomposition in the section.

4.2.1. Topic Corpus Processing

In this study, the topic words refer to all the terms that accurately reflect the content of the review topic. These words are derived from the review collection, differentiating them from general terms. Unlike narrative words, which are manually cited, these topic headings are extracted directly from the original review content. This approach is beneficial for automatically identifying the validity of the topic, making it particularly useful for processing large amounts of review data. Unlike keywords, which only include words that appear in the reviews, topic words also encompass other relevant terms that may not be explicitly mentioned in the reviews. Consequently, there has been a notable increase in the quantity of topic words, offering a more extensive coverage within the review. In addition, to stop words such as “of,” “a,” “he,”, etc., and some symbols, we use the NLTK tool to remove them from the text.

4.2.2. Definition of the Basic Concept

The chosen topic words may reflect a connection to the review topic, but it is complex. On the one hand, it is impossible to cover all relevant topics in a review due to the existence of synonyms and differences in users’ interests in word choice. On the other hand, topic words that do not appear in the reviews may also reflect the review topic. Therefore, we refer to the topic words that appear in the review content as explicit topic words and other words that do not appear in the current review content as latent topic words. Latent topic words related to the current review topic are called implicitly relevant topic words, while latent topic words unrelated to the review topic are called implicitly irrelevant topic words. To ensure that the keywords fully represent the content of the comments, then we adopted the following pre-process. First, we trained a Word2Vec model using the entire dataset of comment corpora to obtain semantic vectors for words. Next, we calculated the TF-ICF values for all nouns and adjectives, selecting the top 20 words with the highest TF-ICF values in each comment as candidate explicit keywords. Finally, we manually sample and validate high-frequency keywords across multiple categories to ensure their semantic relevance. This process aims to minimize lexical selection bias and guarantee that the chosen words represent the core content of the reviews.
Here, we propose the notion of relevance probability to ascertain the likelihood that a latent topic word is implicitly associated with the reviews. A higher relevance probability indicates a greater likelihood that the latent topic word is implicitly related to the reviews.
First, the class term frequency-inverse class frequency (CTF-ICF) is applied to construct the weight matrix G t = g j k for the review data at time t. It can address the issue of a topic word appearing frequently in a shorter text but not having a high word frequency. Simply considering topic word frequency is insufficient in determining a term’s importance. It considers the user’s historical preferences for the same item category by aggregating them into a longer text and then calculating the frequency of the topic word. If the term appears frequently in all texts within the category, it is considered more important for that category. Here, g j k represents the weight of topic word k in review j. Next, a probability matrix Z t = z j k is constructed, where all terms that have already been identified are considered positive samples and z j k is set to 1. On the other hand, all latent topic words are considered negative samples and assigned a correlation probability of 0, as shown in Equation (11):
z j k = 0 , g j k = 1 , g j k
We have preference weight moments P t = P j k based on the richness of reviews and the popularity of topic words at moment t. The corresponding P j k is set to 1 for review topic words that already have a weight value. For latent topic words, their weights represent the confidence level of the related topic words. A higher weight indicates a higher probability that it is a related topic word.
In determining the weight assigned to latent topic words, Zhao et al. examined a non-uniform weighting method for missing items based on item popularity. They observed that items with more incredible popularity are more likely to be viewed by users. However, the absence of user interaction with these items may suggest a lack of interest [33]. Consequently, we propose Hypothesis 1: the absence of reviews for trendy topic words better signifies the insignificance of both reviews and those particular topic words. Hypothesis 1 primarily considers the impact of topic word popularity and does not account for the symmetric factor of review richness. Zhang et al. contend that more active users tend to be less interested in items that do not generate interactions, as indicated in [34]. From this perspective, we propose Hypothesis 2: topic words absent from wealthy reviews are more likely to be irrelevant to the topic discussed in the review. ’Richness’ refers to the extent of topic information within a review, wherein a higher abundance of topic words signifies more extraordinary richness.

4.2.3. The Objective Latent Topic Model

Based on the basic concept, we proposed a latent factor factorization model that utilizes the richness and popularity weights of review topic words to enhance the topic model further. This model aims to infer the relationship between reviews and latent topic words through objective function optimization.
According to Hypothesis 1 and Hypothesis 2, we posit that when the popularity or richness of words associated with a latent topic increases, the probability of these words being implicitly related to the review decreases. Consequently, the preference weight P j k for such latent topic words can be adjusted to a smaller value.
Then we define the richness of review j, denoted as X j , as the total number of topic words present in that review, i.e.,  X j = k = 1 n z j k . Similarly, the popularity of a topic word k is represented by Y k , which is the total number of times that word appears in different comments, i.e.,  Y k = j = 1 l z j k . To normalize the values of X j and Y k , we use the Z-score normalization method, expressed in the following Equations (12) and (13):
X ^ j = X j X ¯ S D ( X )
Y ^ k = Y k Y ¯ S D ( Y )
Linear weighting of Equations (12) and (13), the final weighting equation is Equation (14)
P j k = β 1 e X ^ j + ( 1 β ) 1 e Y ^ k , z j k = 0 1 , z j k = 1
β 0 , 1 is a weighting factor to adjust the share of review richness and topic word popularity.
As illustrated in Figure 2, we use Equation (14) to construct the preference coefficient matrix P t . Then, we utilize the correlation probability matrix Z t and the preference weight matrix P t to perform a weighted matrix decomposition of the associated probability matrix Z t . This decomposition results in two low-dimensional feature matrices, M t and Q t , where M t represents the topic word correlation latent factor matrix and Q t represents the topic word correlation probability confidence latent factor matrix. Based on our previous hypotheses, we apply the preference weights P ( t ) to the word relevance latent factor matrices to ensure accuracy. We then calculate the loss function Θ using alternating least square error (ALS) as shown in Equation (15):
argmin Θ = j k P j k ( t ) | | Z j k ( t ) Z ^ j k ( t ) | F 2 = j k P j k ( t ) | | Z ( t ) M j ( t ) Q k ( t ) T | F 2
The P ˜ ( t ) denotes a diagonal matrix composed of the corresponding elements in P ( t ) . To prevent overfitting, we add regularisation terms to the loss function Θ , and the final loss function Θ is defined as Equation (16)
argmin Θ = j k P j k ( t ) | | Z j , k ( t ) M j ( t ) Q k ( t ) T | F 2 + σ | M j ( t ) | F 2 + | Q k ( t ) | F 2

4.2.4. The Optimization of Topic Model

Firstly, the process of taking partial derivatives of M t is as follows: first, take the derivative of a single element as shown in Equation (17):
1 2 M ( t ) , Q ( t ) M j m ( t ) = k P j k ( t ) M j ( t ) Q k ( t ) T Z j , k ( t ) Q k m ( t ) + σ k P j k ( t ) M j m ( t )
The ( 0 m d ) , then derive the vector element as Equation (18):
1 2 M ( t ) , Q ( t ) M j ( t ) = 1 2 M ( t ) , Q ( t ) M j 1 ( t ) , , M ( t ) , Q ( t ) M j d ( t ) = M j ( t ) Q ( t ) T P j ( t ) Q ( t ) + σ k P j k ( t ) I Z j ( t ) P j ( t ) Q ( t )
The I is the unit matrix. Taking partial derivatives of M t and derive Θ ( M , Q ) M = 0 to obtain the updated M t , which is calculated as Equation (19)
M ( t ) M ( t ) z j ( t ) P j ( t ) Q ( t ) Q ( t ) T P j ( t ) Q ( t ) + σ k P j k ( t ) I
Likewise, taking partial derivatives of Q t ,and derive Θ ( M , Q ) Q = 0 to obtain the updated Q t , which is calculated as Equation (20)
Q ( t ) Q ( t ) z k ( t ) T P j ( t ) M ( t ) M ( t ) T P j ( t ) M ( t ) + σ j P j k ( t ) I
The matrices M and Q, which represent latent factors, undergo iterative updates using Equations (19) and (20) until the loss function reaches a local optimum. At this stage, we define Z ^ = M Q T , where each element Z ^ j k in the matrix Z ^ signifies the probability that review j, when weighted, is associated with topic word k. Our objective is to approximate Z ^ to Z, wherein a higher Z ^ j k value indicates a greater likelihood of the topic word k being relevant to review j.

4.3. Joint Optimization Loss Function

The paper aims to transform the association between user rating and latent topics into a task of finding a joint optimal solution. Optimizing the regularisation factor Z allows the model to extract topic word features from users’ various topics over time. This process enables the reflection of the evolving relationship between each latent review topic associated with user i and item j ratings. So we give the objective functions in Equations (10) and (16) as Equation (21):
arg min L = R ( t ) γ U R ( t ) V R ( t ) F 2 + R ( t ) ( 1 γ ) U R ( t T ) D R ( t T ) V R ( t ) F 2 + σ 1 D R ( t T ) I F 2 + j k P j k ( t ) | | Z ( t ) M j ( t ) Q k ( t ) T | F 2 + σ 2 U R ( t ) 1 + U R ( t T ) 1 + V R ( t ) 1 + | M j ( t ) | F 2 + | Q k ( t ) | F 2
In Equation (21), the function indicates that the error between the predicted and actual data is minimized. The bias level of U ( t T ) is governed by the time regularization term σ 1 | | D R t T I | | F 2 . Here, I represents the identity matrix, while σ 1 and σ 2 are regularisation parameters. A higher σ 1 value implies that the model U R t T is more sensitive to the impact of the auxiliary matrix D R ( t T ) on its influence. The model’s structure is depicted in Figure 3.
For the objective function, the optimization problem is solved using the first-order necessary KKT condition equation constraints [35]. Especially, each rating matrix R i t of user i is updated according to the temporal weight of Equation (6) with Equation (22):
R i ( t ) W ( t T ) R i ( t )
The ultimate goal of this model is to ascertain the optimal solution for the parameters U, V, D, M, and Q, respectively, to link the user rating with the review information. The algorithm is presented in Algorithm 1.    
Algorithm 1: The proposed algorithm
Systems 13 01034 i001

4.4. Computational Complexity

The model’s primary considerations involve the computational complexity of the objective function and topic decomposition. Regarding the objective function L , the algorithm’s maximum complexity during iteration is O ( I i j τ | T | ) , where I represents the iterations needed for matrix decomposition, and | T | signifies the total time dimension. The maximum time complexity for the factorization iterations of the latent topics in Equation (14) is O ( I j k ) . Consequently, the overall complexity is denoted as T ( n ) = O ( I i j τ | T | + I j k ) . As τ remains a constant term in this context, its complexity is O(1). Thus, the maximum time complexity of the model algorithm stands at O ( I i j | T | ) .

5. Experiment and Analysis

5.1. Datasets

The experiment uses six datasets to assess the model’s performance. These datasets are sourced from the Amazon compiled by Stanford University [36]. Each dataset comprises user ratings, item details, associated user reviews, temporal information on review behavior, and rating values ranging from integers 1 to 5. The Table 2 displays the sizes of each dataset.

5.2. Evaluation Indicators

In this paper, we utilized four evaluation metrics, precision, recall, NDCG, and F1, for our experiments. The objective of item recommendation involves generating Top-N item recommendation lists for the specified users in the test dataset. The computations for these metrics are depicted in Equations (23)–(27). The F1 metric’s calculation can be performed using Equation (23).
F 1 = 2 × Precision ( N ) × Recall ( N ) Precision ( N ) + Recall ( N )
where precision and recall are defined as shown in Equations (24) and (25):
Precision ( N ) = Top ( N ) BL N
Recall ( N ) = Top ( N ) BL BL
where BL represents the user’s purchase list, Top(N) represents the top N recommended item list, and NDCG is a measure used to evaluate the quality of recommendation results within the top N items. The formula for NDCG@N is expressed in Equation (26):
NDCG @ N = 1 IDCG @ N i = 1 N 2 k i 1 log 2 i + 1
Here, k i represents whether item i is present in the recommended result set (1 if true, 0 if false). The formula for IDCG@N is defined as presented in Equation (27):
IDCG @ N = i = 1 RLT 2 k i 1 log 2 i + 1
The RLT in Equation (27) denotes the collection of the most pertinent outcomes within the suggested results under optimal conditions. It refers to the assortment comprising the top N results arranged in descending order of recommendation relevance. Hence, as the relevance of the recommended results increases, so does the NDCG value.

5.3. Experimental Parameter Setting

We compare the proposed TDLRP-MF with baseline models such as TPNE [18], TCMF [11], TMRevCo [12], TDADLFM [19], and MTUPD [28] which are described in Section 2.
In the experiment, we use the state-of-the-art parameters of the corresponding model for comparison. The TPNE model’s head number of layers is set as 4 and the number of Transformer layers is set as 2. For the TCMF model experimental parameters, we set λ = β = γ = 1 , ϵ = 10 4 , and maxIter = 1000. The TMRevCo model parameter settings are λ c = 0.001 and λ r = 0.05 . For the TDADLFM model, we set the history size, number of epochs, and batch size as 30, 20, and 1500, respectively. The parameter settings for the MTUPD model use α = 1 , K = 20, λ 1 = 0.5 , and λ 2 = 0.5 . To ensure training stability and convergence, all latent factor matrices of the proposed model were initialized using the Xavier method to prevent gradient vanishing or explosion. Then the optimization process employs Alternating Least Squares (ALS), which effectively enhances training stability by decomposing non-convex problems into a series of convex subproblems. All experiments were completed within 48 h on a server equipped with a 14,900 CPU and a 4090 GPU.

5.4. Parameter Analysis

In the parametric experiments, we configured the maximum number of iterations (maxIter) to 10 3 and the maximum number of sub-iterations (subMaxIter) to 10 2 . Additionally, we set the decomposition dimension to 20. We employed the random initialization method to initialize the weight matrix a w v and the bias vector b w v . We updated the weight and bias parameters using gradient descent. Based on this setup, we analysed the model’s required parameters through experimental investigation, namely α , β , γ , σ 1 , and σ 2 .
This study examined the impact of the review topic hotness and richness regulation parameter β . We varied β between 0.05 and 0.95, with a time fragment of T = 12 months and a step size of 0.2. These experiments were performed across different datasets and evaluated using three metrics: Precision, F1, and NDCG. The results, depicted in Figure 4, showcase the relationship between the model’s β values and changes in F1 performance. Notably, we observed that the model’s performance improves as β increases, reaching its peak at β = 0.55 . However, beyond β = 0.55 , the accuracy starts to decline. This observation suggests that tremendous and small values of β can harm the accuracy of predicting user topic preferences. Therefore, we identify β = 0.55 as the optimal regulation parameter for review topic hotness and richness in this study.
The impact of regularisation parameters σ 1 and σ 2 on model performance was investigated in our experiments. These parameters were varied in the range of [ 0 , + ) , and the optional values are 0.01 , 0.1 , 0.2 , 0.5 , 1 , 5 , 10 , 20 for the Books dataset. In Figure 5, the horizontal axis represents the values of the regularisation parameters, and the vertical axis shows the changes in model NDCG performance. As the values of σ 1 and σ 2 increase, the model performance also improves. The highest overall performance is achieved when the values of σ 1 and σ 2 are set to 0.5, after which the accuracy decreases. This suggests that a too-small value of the regularisation parameter leads to underfitting. At the same time, a too-large value leads to overfitting, resulting in a decrease in model accuracy and an inability to predict user preferences accurately. Therefore, we have chosen a value of 0.5 for the regularisation parameters σ 1 and σ 2 in this paper.
We investigated the impact of the time-weight balance factor, represented by alpha ( α ), and the time-evolution balance factor, represented by gamma ( γ ), on the model’s performance. Section 3 and Section 4 provided the definitions for α and γ to determine time segments ranging from 24 to 72 months. We adjusted the values of α and γ within the range of 0 to 1 in increments of 0.25 to assess their effects. The results, depicted in Figure 6, show the model’s precision index across different values. Lighter colours in the graph denote a higher precision index, while darker shades indicate lower values. Our findings reveal that the model’s performance improves with increasing α and γ values. The accuracy declines when α surpasses 0.5 and γ exceeds roughly 0.75. This suggests that extreme values of α or γ inadequately express the influence of forgetting weight and review evolution weight on users’ preferences or ignore the crucial evolutionary relationship between users’ current and past preferences, leading to reduced accuracy. Consequently, based on our experiments, we identified optimal balance factors of α = 0.5 and β = 0.75 for this model. These outcomes underscore the significance of comprehensively incorporating the forgetting law and considering the combined impact of users’ current and historical preferences on recommendations to enhance outcomes effectively.
To validate the robustness of the training process, we conducted 10 additional independent experiments on the Books dataset using different random seeds. The results show that the NDCG@5 mean metric on the test set is 0.7941 with a standard deviation of 0.0023 (relative error < 0.3%). This indicates that despite different starting points, the model converges stably to locally optimal points with comparable performance. Additionally, we implemented an early stopping mechanism that terminates training if the validation set loss fails to decrease for 10 consecutive epochs. This effectively enhances training efficiency and prevents overfitting.

5.5. Ablation Results

To delve deeper into the impact of each component’s enhancement on performance, we partitioned the ablation model into three categories: experiments for ablation of the topic model, experiments isolating the temporal evolution factor, and experiments altering topic weights. These ablation experiments were conducted across each dataset, and their results are detailed in Table 3.
Four topic models were compared in the initial experiments with those utilized in this paper, as depicted in Table 3. LSA and pLSA employ matrix decomposition to create matrices for document topics and topic words. A Bayesian statistics-based LDA generates distributions for document topics and topic words. NMF is another matrix decomposition-based method, that forms matrices for document and word topics. The approach adopted in this paper incorporates non-negative decomposition, accounting for temporal evolution, increased emphasis on topic popularity, and richness to enhance recommendation effectiveness. Bold results indicate that while combining these four ablated topic models attained the highest performance within their respective categories based on F1 and NDCG metrics across various datasets (identified as the best metric among these experiments), the overall performance still fell short compared to the final optimized model T D L R P o p t presented in this paper. The proposed model can mine the popular topics in user comments in different periods and capture the dynamic evolution law of user topic preferences through the potential features of popularity and abundance of topic terms to improve the effectiveness of recommendations.
In the second group of TDW ablation experiments, we analyzed four scenarios using the temporal weights defined by Equations (8) and (12). These scenarios were TDW1, TDW2, TDW3, and TDW4. In TDW1, the factor values were set to α = 1 and γ = 0 , which means that the temporal weight only considers the user forgetting weight and the interest preference of the past period t T . In TDW2, the factor values were set to α = 0 and γ = 1 , which means that the temporal weight only considers the user review weight and the interest preference of the current period t. In TDW3, the factor values were set to α = 0 and γ = 0 , which means that the temporal weight only considers the user review weight and the interest preference of the period t. Finally, in TDW4, the factor values were set to α = 1 and γ = 1 , which means that the temporal weight only considers the user forgetting weight and the interest preference of the current period t. In Table 3, the symbol † indicates the best performance of the NDCG indicator, 0.7549, and the best performance of the F1 indicator for TDW2, 0.6946. However, in each data test, the indicator results were lower than the final model T D L R P o p t , indicating that only by fully considering the temporal weights can we better reflect the temporal pattern of user interests and improve the recommendation effectiveness.
The third group conducted ablation experiments regarding weight LRP in their study. These experiments were split into two scenarios: LRP1, where β = 0 and the weight solely accounts for popularity, and LRP2, where β = 1 and the weight solely considers richness. The outcomes are detailed in Table 3. Across each dataset, the symbol ‡ highlights the highest performance of the NDCG indicator at 0.7649, and for TDW2, the best F1 indicator performance reached 0.6989. However, these dataset-specific results fell short compared to the indicator outcomes of the ultimate model T D L R P o p t . The result suggests that neglecting temporal topic weighting may lead to the loss of crucial user preference information, potentially impacting the accuracy of recommendation outcomes.
The enhanced model in this study comprehensively incorporates multiple users’ temporal interests, showcasing the most vital capability in capturing the evolving aspects of users’ diverse interests. This enhancement significantly contributes to improving the effectiveness of recommendations.

5.6. Performance Analysis and Case Study

Based on the optimized parameters mentioned earlier, we expanded the number of recommendations from 10 to 50 in increments of 10. The accuracy evaluation for Top@N recommendations across various datasets and recommendation methods is depicted in Figure 7. Our findings illustrate the superiority of the TDLRP-MF model over other compared models. This underscores its exceptional capacity to accommodate users’ evolving temporal preferences.
Table 4 presents a more in-depth dataset analysis detailing each model’s performance metrics (where Top@N = 5 or @N = 10). Table 4 compares these metrics across models, revealing that the TDLRP-MF model outperforms other baseline methods across all three metrics. To validate the statistical significance of the performance improvement achieved by the TDLRP-MF model, we conducted five independent experiments using different random seeds and performed paired t-tests on the NDCG@5 results against the strongest baseline model MTUPD. The results indicate that across all six datasets, the p-values for improvements achieved by TDLRP-MF relative to MTUPD were consistently below 0.01. This strongly demonstrates that the performance gains obtained by our model are statistically significant. The result emphasizes our model’s effectiveness in capturing users’ dynamic preferences over time, considering user forgetting patterns and review evolution, thereby capturing temporal interest shifts. Consequently, our model significantly enhances recommendation accuracy compared to other temporal approaches.
Moreover, we compared NDCG@5 among different time-aware models at every period, as illustrated in Figure 8. Compared to the baseline model, our model can stay ahead of the curve at each time step, meanwhile, as the time step increases, our model can remain stable in terms of performance. This result demonstrates that leveraging user ratings and item review evolution not only effectively enhances recommendation performance, but also can learn the evolutionary pattern of user preferences through a history of longer time steps to improve the perception of sparse data so that the predicted items are close to the user’s real preferences.
We selected a real anonymized user with long-term active behavior in the Electronics category to qualitatively demonstrate how TDLRP-MF tracks their interest drift. Phase 1 (2016): The user exhibited strong long-term memory for items (slower decay of forgetting weights), with review keywords centered on “battery life” and “good price.” Phase 2 (2018): The user’s interests undergo a noticeable shift. The decay weight indicates accelerated forgetting of old interests, while the core comment keywords evolve to “camera quality” and “screen resolution.” Our model successfully captured this dynamic: In Phase 1, it recommended cost-effective devices with long battery life. By Phase 2, the recommendation list shifted to flagship models emphasizing camera and screen quality as core selling points. This case clearly demonstrates how the collaborative work of the forgetting mechanism and topic evolution module enables the model to adapt to natural shifts in user interests, thereby delivering recommendations better aligned with their current preferences.

6. Conclusions

The proposed model aims to seek the dynamic attributes of users’ temporal preferences, encompassing aspects such as the fading of interests over time and the evolution of comments across various stages. This model not only tracks the shifting preferences of users throughout time but also accounts for the influence of comment evolution on user preferences. Consequently, it enables accurate, personalized recommendations from a substantial corpus of published information. The research reveals that current user preferences are shaped by their historical inclinations. Hence, considering popularity and depth, the model establishes a topic preference weight matrix for each user. Utilizing potential topic matrix decomposition, it captures the evolving interests of users in various subjects. Empirical findings demonstrate that this recommendation model surpasses existing baseline methods in effectively extracting user temporal characteristics. However, the study has several limitations that provide directions for future work, such as the forgetting function employed in the model using globally fixed parameters, failing to capture variations in interest decay rates across different users or item types.
In future work, one potential improvement area is focusing on developing an adaptive mechanism that learns personalized forgetting laws from users’ historical interaction data. In addition, the introduction of large language models (LLMs) for user topic representation learning is also a new research avenue for the future.

Author Contributions

Conceptualization, H.D.; methodology, H.D.; software, H.D.; validation, H.D.; formal analysis, H.D.; investigation, H.D.; resources, H.D.; data curation, H.D.; writing—original draft preparation, H.D.; writing—review and editing, H.D., W.Z., G.H. and Z.B.; visualization, H.D.; supervision, H.D.; project administration, H.D.; funding acquisition, H.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by China Postdoctoral Science Foundation grant number 2025M773202.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The online survey data have been presented in this paper and the others are securely protected by the researchers.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, X.; Liu, C. Design of Personalized News Recommendation System Based on an Improved User Collaborative Filtering Algorithm. Mob. Inf. Syst. 2023, 2023, 9898337. [Google Scholar] [CrossRef]
  2. Wang, Y.; Wu, J.; Wu, Z.; Adomavicius, G. Efficient and flexible long-tail recommendation using cosine patterns. INFORMS J. Comput. 2025, 37, 446–464. [Google Scholar] [CrossRef]
  3. Jain, G.; Mahara, T.; Sharma, S.C. Effective time context based collaborative filtering recommender system inspired by Gower’s coefficient. Int. J. Syst. Assur. Eng. Manag. 2023, 14, 429–447. [Google Scholar] [CrossRef]
  4. Tanveer, J.; Lee, S.W.; Rahmani, A.M.; Aurangzeb, K.; Alam, M.; Zare, G.; Alamdari, P.M.; Hosseinzadeh, M. PGA-DRL: Progressive graph attention-based deep reinforcement learning for recommender systems. Inf. Fusion 2025, 121, 103167. [Google Scholar] [CrossRef]
  5. Parmezan, A.R.S.; Souza, V.M.; Batista, G.E. Evaluation of statistical and machine learning models for time series prediction: Identifying the state-of-the-art and the best conditions for the use of each model. Inf. Sci. 2019, 484, 302–337. [Google Scholar] [CrossRef]
  6. Pang, N. A Personalized Recommendation Algorithm for Semantic Classification of New Book Recommendation Services for University Libraries. Math. Probl. Eng. 2022, 2022, 8740207. [Google Scholar] [CrossRef]
  7. Wang, X.J.; Liu, T.; Fan, W. TGVx: Dynamic personalized POI deep recommendation model. INFORMS J. Comput. 2023, 35, 786–796. [Google Scholar] [CrossRef]
  8. Alabduljabbar, R.; Alshareef, M.; Alshareef, N. Time-Aware Recommender Systems: A Comprehensive Survey and Quantitative Assessment of Literature. IEEE Access 2023, 11, 45586–45604. [Google Scholar] [CrossRef]
  9. Koren, Y. Collaborative filtering with temporal dynamics. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, 28 June–1 July 2009; pp. 447–456. [Google Scholar]
  10. Xiong, L.; Chen, X.; Huang, T.K.; Schneider, J.; Carbonell, J.G. Temporal collaborative filtering with bayesian probabilistic tensor factorization. In Proceedings of the 2010 SIAM International Conference on Data Mining, Columbus, OH, USA, 29 April–1 May 2010; pp. 211–222. [Google Scholar]
  11. Rafailidis, D.; Kefalas, P.; Manolopoulos, Y. Preference dynamics with multimodal user-item interactions in social media recommendation. Expert Syst. Appl. 2017, 74, 11–18. [Google Scholar] [CrossRef]
  12. Wu, T.; Feng, Y.; Sang, J.; Qiang, B.; Wang, Y. A novel recommendation algorithm incorporating temporal dynamics, reviews and item correlation. IEICE Trans. Inf. Syst. 2018, 101, 2027–2034. [Google Scholar] [CrossRef]
  13. Li, T.; Jin, L.; Wu, Z.; Chen, Y. Combined recommendation algorithm based on improved similarity and forgetting curve. Information 2019, 10, 130. [Google Scholar] [CrossRef]
  14. Pang, S.; Yu, S.; Li, G.; Qiao, S.; Wang, M. Time-Sensitive Collaborative Filtering Algorithm with Feature Stability. Comput. Inform. 2020, 39, 141–155. [Google Scholar] [CrossRef]
  15. Lv, Y.; Zhuang, L.; Luo, P.; Li, H.; Zha, Z. Time-sensitive collaborative interest aware model for session-based recommendation. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo (ICME), London, UK, 6–10 July 2020; pp. 1–6. [Google Scholar]
  16. José, E.F.; Enembreck, F.; Barddal, J.P. Adadrift: An adaptive learning technique for long-history stream-based recommender systems. In Proceedings of the 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, ON, Canada, 11–14 October 2020; pp. 2593–2600. [Google Scholar]
  17. Rabiu, I.; Salim, N.; Da’u, A.; Nasser, M. Modeling sentimental bias and temporal dynamics for adaptive deep recommendation system. Expert Syst. Appl. 2022, 191, 116262. [Google Scholar] [CrossRef]
  18. Huang, C.; Wang, S.; Wang, X.; Yao, L. Modeling temporal positive and negative excitation for sequential recommendation. In Proceedings of the ACM Web Conference 2023, Austin, TX, USA, 30 April–4 May 2023; pp. 1252–1263. [Google Scholar]
  19. Wu, C.C.; Chen, Y.L.; Yeh, Y.H. A Deep Recommendation Model Considering the Impact of Time and Individual Diversity. IEEE Trans. Comput. Soc. Syst. 2023, 11, 2558–2569. [Google Scholar] [CrossRef]
  20. Sun, Y.T.; Guo, G.B.; Chen, X.; Zhang, P.H.; Wang, X.W. Exploiting review embedding and user attention for item recommendation. Knowl. Inf. Syst. 2020, 62, 3015–3038. [Google Scholar] [CrossRef]
  21. Wang, X.; Xiao, T.; Tan, J.; Ouyang, D.; Shao, J. MRMRP: Multi-source review-based model for rating prediction. In Proceedings of the Database Systems for Advanced Applications: 25th International Conference, DASFAA 2020, Jeju, Republic of Korea, 24–27 September 2020; Proceedings, Part II 25. Springer: Berlin/Heidelberg, Germany, 2020; pp. 20–35. [Google Scholar]
  22. Wu, S.; Zhang, Y.; Zhang, W.; Bian, K.; Cui, B. Enhanced review-based rating prediction by exploiting aside information and user influence. Knowl.-Based Syst. 2021, 222, 107015. [Google Scholar] [CrossRef]
  23. Hyun, D.; Park, C.; Cho, J.; Yu, H. Learning to utilize auxiliary reviews for recommendation. Inf. Sci. 2021, 545, 595–607. [Google Scholar] [CrossRef]
  24. Wang, P.P.; Li, L.; Wang, R.; Zheng, X.H.; He, J.X.; Xu, G.D. Learning persona-driven personalized sentimental representation for review-based recommendation. Expert Syst. Appl. 2022, 203, 117317. [Google Scholar] [CrossRef]
  25. Liu, Z.; Yuan, B.X.; Ma, Y. A multi-task dual attention deep recommendation model using ratings and review helpfulness. Appl. Intell. 2022, 52, 5595–5607. [Google Scholar] [CrossRef]
  26. Cai, Y.; Ke, W.M.; Cui, E.; Yu, F. A deep recommendation model of cross-grained sentiments of user reviews and ratings. Inf. Process. Manag. 2022, 59, 102842. [Google Scholar] [CrossRef]
  27. Liu, H.; Chen, Y.; Li, P.; Zhao, P.; Wu, X. Enhancing review-based user representation on learned social graph for recommendation. Knowl.-Based Syst. 2023, 266, 110438. [Google Scholar] [CrossRef]
  28. Wangwatcharakul, C.; Wongthanavasu, S. A novel temporal recommender system based on multiple transitions in user preference drift and topic review evolution. Expert Syst. Appl. 2021, 185, 115626. [Google Scholar] [CrossRef]
  29. Milovančević, N.S.; Gračanac, A. Time and ontology for resource recommendation system. Phys. A Stat. Mech. Its Appl. 2019, 525, 752–760. [Google Scholar] [CrossRef]
  30. Cheng, Y.; Qiu, G.; Bu, J.; Liu, K.; Han, Y.; Wang, C.; Chen, C. Model bloggers’ interests based on forgetting mechanism. In Proceedings of the 17th International Conference on World Wide Web, Beijing, China, 21–25 April 2008; pp. 1129–1130. [Google Scholar]
  31. Yu, Z.; Lian, J.; Mahmoody, A.; Liu, G.; Xie, X. Adaptive User Modeling with Long and Short-Term Preferences for Personalized Recommendation. IJCAI 2019, 7, 4213–4219. [Google Scholar]
  32. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef]
  33. Zhao, Z.; Chen, J.; Zhou, S.; He, X.; Cao, X.; Zhang, F.; Wu, W. Popularity bias is not always evil: Disentangling benign and harmful bias for recommendation. IEEE Trans. Knowl. Data Eng. 2022, 35, 9920–9931. [Google Scholar] [CrossRef]
  34. Zhang, Y.; Feng, F.; He, X.; Wei, T.; Song, C.; Ling, G.; Zhang, Y. Causal intervention for leveraging popularity bias in recommendation. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 11–15 July 2021; pp. 11–20. [Google Scholar]
  35. Kuhn, H.W.; Tucker, A.W. Nonlinear programming. In Traces and Emergence of Nonlinear Programming; Springer: Basel, Switzerland, 2013; pp. 247–258. [Google Scholar]
  36. McAuley, J.; Leskovec, J. Hidden factors and hidden topics: Understanding rating dimensions with review text. In Proceedings of the 7th ACM Conference on Recommender Systems, Hong Kong, China, 12–16 October 2013; pp. 165–172. [Google Scholar]
Figure 1. The proposed TDLRP-MF model diagram.
Figure 1. The proposed TDLRP-MF model diagram.
Systems 13 01034 g001
Figure 2. The latent topic words model diagram.
Figure 2. The latent topic words model diagram.
Systems 13 01034 g002
Figure 3. Graphical illustration of the proposed TDLRP-MF model.
Figure 3. Graphical illustration of the proposed TDLRP-MF model.
Systems 13 01034 g003
Figure 4. The impact of topic popularity and richness parameter β on model performance.
Figure 4. The impact of topic popularity and richness parameter β on model performance.
Systems 13 01034 g004
Figure 5. The impact of different values of regularization parameter σ 1 , σ 2 on the model.
Figure 5. The impact of different values of regularization parameter σ 1 , σ 2 on the model.
Systems 13 01034 g005
Figure 6. The impact of time weight adjustment parameter α , γ on the model. (Top-5 recommendation, T = 24, 48, 72 respectively).
Figure 6. The impact of time weight adjustment parameter α , γ on the model. (Top-5 recommendation, T = 24, 48, 72 respectively).
Systems 13 01034 g006
Figure 7. Results of precision comparison.
Figure 7. Results of precision comparison.
Systems 13 01034 g007
Figure 8. Comparison of NDCG@5 at each time-step.
Figure 8. Comparison of NDCG@5 at each time-step.
Systems 13 01034 g008
Table 1. Main symbols and parameters.
Table 1. Main symbols and parameters.
NotationDescription
TTime periods.
τ Time decomposition dimension.
R ( t ) R i × j The ratings matrix at time t.
R ( t T ) R i × j The ratings matrix at time t T .
U R ( t ) R i × τ User latent factor matrix.
V R ( t ) R τ × j Item latent factor matrix.
D R ( t T ) R τ × τ Drift transition matrix.
W t T The enhanced time weighting.
lNumber of review topics.
cNumber of review topic words.
G t R l × c Review topic words weight matrix.
Z t R l × c Review topic words probability matrix.
P t R l × c Review topic words preference weight matrix.
S D ( ) Standard deviation.
| | | | p The p-norm.
| | | | F The Frobenius norm.
α Time weighting balance parameter.
β Topic words feature weight adjustment coefficient.
γ Temporal decomposition adjustment parameter.
σ 1 , σ 2 Regularization parameters.
Table 2. Datasets description.
Table 2. Datasets description.
DatasetsUsersItemsRatings & ReviewsSparsityTime Range
Kindle Store346,45729,3571,767,61299.9826%April 1999–October 2018
Books578,85861,3794,159,48099.9883%December 1997–October 2018
Magazine Subscriptions438,64537,2502,638,21999.9839%March 1999–October 2019
Electronics9,838,676786,86820,994,35399.9997%December 1997–October 2018
Home&Kitchen9,767,6061,301,22521,928,56899.9998%November 1999–October 2018
Toys&Games4,204,994634,4148,201,23199.9997%October 1999–October 2018
Table 3. Ablation results.
Table 3. Ablation results.
BooksKindle StoreMagazine Subscriptions
Object NDCG F1 NDCG F1 NDCG F1
@Top5 @Top10 @Top5 @Top10 @Top5 @Top10 @Top5 @Top10 @Top5 @Top10 @Top5 @Top10
+NMF0.68780.68250.63920.61600.73530.71630.69360.67090.71680.70950.65080.6361
+LSA0.69940.68730.64190.63130.71470.72370.68410.64500.70310.69310.63870.6345
+pLSA0.73370.71900.64180.62550.74590.71790.67090.65980.72580.71630.65350.6435
+LDA0.71580.70580.64610.61020.73270.73960.67940.65350.72370.71740.65670.6472
TDW 1 ff = 1 , fl = 0 0.69790.69100.66090.66870.7549 †0.75430.68830.68560.72010.72730.65400.6544
TDW 2 ff = 0 , fl = 1 0.69310.6946 †0.65240.65710.74170.74270.67410.67500.71320.71790.65820.6598
TDW 3 ff = fl = 0 0.66780.66830.61230.61330.73800.73900.68410.68880.69890.70390.63710.6435
TDW 4 ff = fl = 1 0.67620.67620.62080.62120.74960.74640.67090.67930.71680.71260.65140.6592
LRP 1 fi = 0 0.70260.70390.63400.63920.7649 ‡0.76370.68290.68770.73070.73150.66140.6621
LRP 2 fi = 1 0.69890.69680.63760.63550.76070.76640.6989 ‡0.69090.72050.72480.65300.6613
TDLRP opt 0.78070.78330.70470.70830.85150.85090.80290.80360.81930.81970.79130.7942
Table 4. Performance evaluation on Top-N recommendation.
Table 4. Performance evaluation on Top-N recommendation.
Comparison of Precision
DataSets/MethodsTPNEMTUPDTCMFTDADLFMTMRevCoTDLRP-MF
P@5P@10P@5P@10P@5P@10P@5P@10P@5P@10P@5P@10
Books0.55970.53970.74350.72180.72660.70980.64910.61670.71730.69460.77910.7652
Kindle Store0.54710.54590.73710.71830.76270.74830.63730.62430.71340.69410.79860.7659
Magazine Subscriptions0.53990.53340.79290.78340.74720.75740.61440.60380.71460.73740.83150.8191
Electronics0.61260.59820.81720.76780.78240.75010.63850.61170.80810.79100.84370.8321
Home&Kitchen0.56170.56360.72150.71560.75120.73900.61370.61050.71240.69170.74610.7435
Toys&Games0.56310.56570.82610.80790.80210.78490.76310.75280.72240.69500.85940.8453
Comparison of F1
DataSets/MethodsTPNEMTUPDTCMFTDADLFMTMRevCoTDLRP-MF
F1@5F1@10F1@5F1@10F1@5F1@10F1@5F1@10F1@5F1@10F1@5F1@10
Books0.58110.58050.70580.69880.69600.69460.65810.65820.66980.63750.72320.7128
Kindle Store0.57810.57660.79710.77540.71640.70520.73800.75530.60440.59720.80720.7984
Magazine Subscriptions0.67440.65190.71750.71390.71130.77780.65370.66100.70240.68360.73610.7215
Electronics0.65710.65810.79890.79920.77020.76670.72650.73410.75720.73360.82350.8196
Home&Kitchen0.63090.62760.70570.69920.69870.67050.67260.67210.64210.63730.72380.7219
Toys&Games0.66720.65130.78640.77270.71520.70790.73610.71740.59680.60080.80150.7952
Comparison of NDCG (N)
DataSets/MethodsTPNEMTUPDTCMFTDADLFMTMRevCoTDLRP-MF
N@5N@10N@5N@10N@5N@10N@5N@10N@5N@10N@5N@10
Books0.65360.65240.77210.75040.74280.73100.74670.74570.68650.66240.79840.7769
Kindle Store0.66590.62380.80200.79460.80670.83290.65980.66790.73090.71850.82690.8139
Magazine Subscriptions0.63690.62910.73170.72330.75440.73180.65450.64720.69280.66200.83760.8273
Electronics0.64780.63930.81020.79830.80320.80560.65770.64160.72150.70890.83470.8131
Home&Kitchen0.61960.61570.70080.71040.69720.69160.62120.61900.65540.64900.74920.7413
Toys&Games0.63380.63150.83220.82490.79930.78050.69570.69310.66790.65050.86130.8504
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ding, H.; Zhu, W.; Hu, G.; Bu, Z. Capturing Dynamic User Preferences: A Recommendation System Model with Non-Linear Forgetting and Evolving Topics. Systems 2025, 13, 1034. https://doi.org/10.3390/systems13111034

AMA Style

Ding H, Zhu W, Hu G, Bu Z. Capturing Dynamic User Preferences: A Recommendation System Model with Non-Linear Forgetting and Evolving Topics. Systems. 2025; 13(11):1034. https://doi.org/10.3390/systems13111034

Chicago/Turabian Style

Ding, Hao, Weiwei Zhu, Guangwei Hu, and Zhan Bu. 2025. "Capturing Dynamic User Preferences: A Recommendation System Model with Non-Linear Forgetting and Evolving Topics" Systems 13, no. 11: 1034. https://doi.org/10.3390/systems13111034

APA Style

Ding, H., Zhu, W., Hu, G., & Bu, Z. (2025). Capturing Dynamic User Preferences: A Recommendation System Model with Non-Linear Forgetting and Evolving Topics. Systems, 13(11), 1034. https://doi.org/10.3390/systems13111034

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop