Next Article in Journal
A Low-Complexity Channel Estimation Based on a Least-Squares Algorithm in OFDM Systems
Next Article in Special Issue
ROS-Based Unmanned Mobile Robot Platform for Agriculture
Previous Article in Journal
On the Absorption and Photoluminescence Properties of Pure ZnSe and Co-Doped ZnSe:Eu3+/Yb3+ Crystals
Previous Article in Special Issue
Development of Priority Index for Intelligent Vessel Traffic Monitoring System in Vessel Traffic Service Areas
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Logit Averaging: Capturing Global Relation for Session-Based Recommendation

Department of Artificial Intelligent, SungKyunKwan University, Suwon 16419, Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2022, 12(9), 4256; https://doi.org/10.3390/app12094256
Submission received: 18 March 2022 / Revised: 12 April 2022 / Accepted: 20 April 2022 / Published: 22 April 2022
(This article belongs to the Special Issue Advances in Intelligent Systems)

Abstract

:
Session-based recommendation predicts an anonymous user’s next action, whether she or he is likely to purchase based on the user’s behavior in the current session as sequences. Most recent research on session-based recommendations makes predictions based on a single-session without incorporating global relationships between sessions. It does not guarantee a better performance because item embeddings learned by solely utilizing a single session (inter-session) have less item transition information than utilizing both intra- and inter-session ones. Some existing methods tried to enhance recommendation performance by adopting memory modules and global transition graphs; however, those need more computation cost and time. We propose a novel algorithm called Logit Averaging (LA), utilizing both (i) local-level logits, which come from intra-sessions item transitions and (ii) global-level logits, which come from gathered logits of related sessions. The proposed method shows an improvement in recommendation performance in respect of accuracy and diversity through extensive experiments.

1. Introduction

With the rapid growth of Internet commerce platforms, recommender systems now play an imperative role in improving user experience and satisfaction on the platforms [1]. Predicting users’ next behavior from current and previous actions not only alleviates information overload problems but can also offer a personalized service to targeted users in various applications [2].
A conventional recommender system [3,4,5,6,7,8] utilizes a history of user-item interactions and side information to learn each user’s long-term and static preferences on items. Interest-aware Message-Passing GCN (IMP-GCN) [8] was a model that plays to the strength of the Graph Convolution Network in recommender systems by alleviating the over-smoothing problem. Liu et al. pinpointed the risk of involving higher-order neighboring users with no common interests with a user and learning their embeddings in the graph convolution operation. To overcome the addressed problem, IMP-GCN performed high-order graph convolution inside subgraphs that consist of users with similar features. We found this model very effective in its ability to avoid the propagation of non-related information from high-order neighbors into embedding learning. IMP-GCN was designed based on the user-item bipartite graph datasets, which consist of user and item features. However, this form of dataset is often invalid in many cases such as numerous anonymous users interacting with items in a very short period of time or the same account can be shared with multiple people [9]. Thus, a new system to deal with the problem has emerged, which is a session-based recommendation [10]. Unlike conventional ones, session-based recommender systems only utilize the ongoing movement of an anonymous user in one session to predict the next item, where a session is composed of items that users interacted with within a continuous period of time. Owing to its practicality in various e-commerce platform scenarios, the session-based recommender system has gained much attention.
To predict the users’ next click items, the session-based recommender system utilizes a deep neural network and has achieved outstanding performances. Recurrent Neural Networks (RNNs) or attention mechanism models were first applied to a session-based recommendation task for extracting sequential patterns, which is a straightforward approach to handling a session as it is the sequence of items user clicked sorted by time [11,12]. NARM [11] utilized gated recurrent unit (GRU) [13], which is the variation of RNN to extract sequential patterns in sessions. SRSAN [12] took dependencies among items in sessions by using self-attention networks.
Furthermore, recent studies incorporate Graph Neural Networks (GNNs) [14] to construct a session-based recommender system due to its effectiveness in representing session consistency along with item dependencies and shows a competitive performance comparable to RNNs-based approaches [15,16,17,18,19]. SRGNN [15] was the first GNN-based framework and has been used as a basic structure in many studies. A session graph was sent into gated-GNN [20] to obtain the latent representation of nodes, and the session representation was generated by a soft-attention mechanism based on the node representations. Gupta et al. [16] normalized representations of items and sessions generated from GNNs to increase the overall recommendation performances. Chen et al. [17] introduced a novel GNN model that focuses on aggregating the information passed from neighboring nodes using a GRU while preserving the item orders in sessions. This layer is called the Edge-Order Preserving Aggregation (EOPA) layer and shows a promising performance while minimizing the information loss problems. TAGNN++ [18] adaptively activated the attention model with respect to varied target items and improved the expressiveness of the model. Many researchers have tried to capture the accurate preference of an anonymous user from a short session, but it is still hard. To overcome this, Li et al. [19] applied disentangle representation learning to make better item representation. They assumed that the chunks of item embeddings represent factors such as categories and colors, and learned item embeddings with a distance correlation function aiming for the factors to disentangle each other. It is useful for the session-based recommendation task because of the absence of additional information in a session dataset.
However, existing GNN-based models only make predictions with items in a single session and do not consider the global-level contexts that exist between inter-sessions. They ignore the useful item transition information from other sessions. In [9], the session-based recommendation has several unique characteristics: Session consistency, Repeated item consumption, Timeliness of sessions, and Sequential dependency. Here, we mainly take sequential dependency into consideration. Some items can be explored in a certain order from the majority of users. For example, we can easily find the item-transition information, [‘phone’, ‘phone case’] or [‘phone’, ‘wireless earphone’]. This pattern will appear in multiple sessions. Thus, taking the global-level context of items into consideration is necessary to improve the prediction accuracy and diversify the recommendation lists. In addition, the number of items in a session is lower than 10 generally. The short length of sessions may lead to incorrect predictions when we make a prediction based on a single session. Therefore, it will be helpful to recommend items with global relations from multiple sessions which share the same items.
Some studies are trying to model the global relations of sessions. Wang et al. [21] applied collaborative neighborhood information to recommend items. It consists of an inner memory encoder and an outer memory encoder that extract the current session’s context and the collaborative information from other sessions which are similar to the current session, respectively. Wang et al. [22] conducted a similar attempt to infer global context through the GCE-GNN (Global Context Enhanced Graph Neural Networks) model. It constructed a global graph using all the unique items that exist in inter-sessions and extracted the global-level item embedding. Fei et al. [23] proposed MAN (Memory Augmented Model) for a session-based recommendation, which augments a base recommender model using updated memory, and recommends items with memory components. Zhou et al. [24] utilized a graph including item relations of each day to capture timely user interests. By doing this, an item can have different neighborhoods on different days. Huang et al. [25] tried to model inter- and intra- item transition through multi-task learning. Each session went into the self-attention layer to capture intra-sessions item transition, and item representations which are constructed at this step were set as initialized values of item representations for learning global item relations. They could improve the recommendation accuracy by designing different layer architectures depending on whether the item is from global or local relations. Xia et al. [26] also tried to reflect the global relation among sessions. They proposed two different graph view augmentations with respect to a session dataset to adopt the self-supervised co-training method on a session-based recommendation. One is the item view graph, generally utilized in session-based recommendation, and the other is the session view graph which represents the inter-session relationship with Jaccard similarity. By learning from these two views, items and sessions, session representations became richer. The Heterogeneous Global Graph Neural Network (HG-GNN) [27] was another model that attempts to extract global item co-occurrence by constructing user-item interactions and global co-occurrence item graphs. They used heterogeneous global graphs to learn long-term user preferences and item representations. Even though their work was to implement the truly personalized session-based recommendation, we found this work has a limitation on data availability. Most of the existing session datasets do not contain user information including their user IDs. Since session-based recommender systems are proposed to make recommendations solely based on the sequence of items anonymous users interacted with within a limited period of time, HG-GNN can be utilized in very few circumstances where the user’s ID and historical sessions are available.
However, those attempts rather increase the computation costs in storing memory and extracting global-level item representation with over a million item nodes and fails to show drastic performance improvement. We propose a simple and novel algorithm called Logit Averaging (LA) that reflects global relations among sessions by averaging short-head item’s logits which are the last representation of the model. There were several preceding studies using logit to improve representation learning in classification tasks such as logit adjustment. Adjusting logits encourages a large relative margin between logits of rare labels versus dominant ones. Menon et al. [28] have conducted extensive studies on classical logit adjustment techniques and proposed a novel logit adjusted loss function. Even though the effectiveness of the logit adjustment loss function in balancing the class label during training, we found this technique hard to apply in session-based recommendation tasks. Menon et al. utilized image datasets, such as CIFAR-10 and iNaturalist, to examine their proposed methods, which have only nine to ten classes. However, in session datasets, the class of each session is thousands of different last-clicked items. If we apply the logit adjustment loss technique to a session-based recommender system, the training of the model will not be performed properly.
Thus, we propose a suitable logit training technique for a session-based recommender system called Logit Averaging. We perform experiments on multiple session-based recommendation models and verify that our algorithm outperforms in prediction accuracy and diversity. Oour accomplishment in increasing the item diversity is especially important in recommender systems because increasing the probability of retrieving unusual or novel items relevant to users highly improves the users’ satisfaction with the system [29,30,31,32]. We summarize our contributions as follows:
  • By interpolating logits of the sessions with the same target value, we propose a simple yet effective method to incorporate global contexts for session-based recommendation and achieve high accuracy and diversity;
  • We conduct extensive experiments on four real-world datasets in six different deep-learning-based session recommendation models and successfully demonstrate the effectiveness of our method;
  • We can easily “plug and play” Logit Averaging (LA) in various neural network-based session recommendation models.
The structure of this paper is as follows. In Section 2, we define the problem and the corresponding assumption we made. Furthermore, an in-detail description of our proposed algorithm, Logit Averaging (LA), is mentioned. In Section 3, we describe experimental settings, including the datasets, baseline models, and evaluation metrics. Then, we make a detailed analysis of the experimental results. Finally, in Section 4, we summarize our proposed method and discuss the future works.

2. Proposed Method

2.1. Problem Definition and Notation

The final goal of the session-based recommendations task is the prediction of a last clicked item based on an anonymous user’s past clicked items. A set I = { i 1 , , i N } denotes the set of all unique items in all sessions. An anonymous user’s session can be represented by s = [ i 1 , i 2 , i 3 , , i t ] , ordered by time. The recommendation model predicts the next clicked item, the ( t + 1 )th item given a session s, then we get probabilities y ^ for all candidate items in I. We split items into short-head and long-tail according to the Pareto Principle [33]. The items which account for 75% of all item appearances are short-head items, and the rest are long-tail items, respectively— I H and I T .

2.2. Logit Averaging

We propose a simple module that averages logits from a prediction layer for the better recommendation, so the other parts of a recommender system, such as an item encoder, a session encoder and a prediction layer, can be any models. In other words, our module can easily be plug-and-play for any other existing neural network-based session recommendation models. As shown in Figure 1, our proposed technique “Logit Averaging (LA)” is located between the session encoder and the final prediction layer.
Most neural network-based recommendation models are structured as the white box in Figure 1. Sessions in a batch go into the model as inputs and an item encoder learns item embeddings. Session embeddings are generated from a session encoder using item embeddings. The final prediction is commonly calculated by the cosine similarity between all unique item embeddings and session embeddings of the batch.
Logit averaging is a module that performs as augmentation, which can inject global context at the same time. As mentioned above, recent deep learning-based session recommendation models recommend items in respect of a single session. It is not enough to make better embeddings when solely using intra-session transition and not considering inter-sessions. Sessions have item consequences where several sessions have common items. This means that the sessions that share the same items have similar item transitions and can help to supplement the information with each other. Then, it is easier to predict items than use a single session alone.
GCE-GNN [22] tried to solve the session recommendation task with a global transition using a global item graph. The graph merges all the items in the entire sessions into a single global graph. This shadows the hidden global contexts in sessions and cannot distinguish the predominant or minor trends.
To overcome the problem, we propose a method to average logits of sessions that share the same target items instead of constructing a large graph. In detail, logits of each sample are obtained by calculating cosine similarity between session embeddings and item embeddings, and the logits are replaced by the mean value of logits that have the same target items. The Logit Averaging and replacing are only performed by sessions that have short-head items as target items. As shown in Figure 2, all the four public session datasets have long-tail item distribution, meaning that the majority of sessions share a few numbers of target items [34]. We assume that sessions with the same target items will have shared a global context. The short-head items take up seventy-five percent of sessions, so it can inject global relations into the majority of sessions without adopting all items. In other words, averaging logits of short-head items easily supplements the global relations information and increases the recommendation performance.
It is not enough to utilize averaged logits only for making better predictions because averaging can ignore the original item transition characteristics of sessions. Therefore, we adopt two loss functions, one is for the original logits and the other for the averaged logits. We get the final loss by the sum of two losses with λ .
L o r g ( y , Z o r g ) = i = 1 m y i log ( Z i o r g )
L o r g ( y , Z L A ) = i = 1 m y i log ( Z i L A )
L = L o r g + λ · L L A .
The detailed process is shown in Algorithm 1.
Algorithm 1 Logit Averaging.
Input: sessions S, Short-head Items I H , Item Encoder I E ( ) , Session Encoder S E ( )
    Initialize IE and SE
    for each iteration do
            V I E ( S )
            S e m b S E ( V , S )
            Z o r g = S e m b · V
           for all  I H  do
                 if  I i H in target items of sessions then
                       Z k L A = 1 | n | k = 0 n Z k ( where the target item of Z k are same as I i H )
                 end if
           end for
           L is calculated by using Equation (3)
end for

3. Results

We conducted extensive experiments to evaluate the performance of our proposed method. This section discusses the performance comparison between the proposed method and baseline methods. Before that, we describe the datasets, the baselines, and the evaluation metrics utilized in our experiments.

3.1. Setup

3.1.1. Datasets

We employed four real-world datasets: Yoochoose (https://www.kaggle.com/chadgostopp/recsys-challenge-2015, accessed on 22 April 2022), Diginetica (https://competitions.codalab.org/competitions/11161#learn_the_details-data2, accessed on 22 April 2022), RetailRocket (https://www.kaggle.com/retailrocket/ecommerce-dataset, accessed on 22 April 2022) and Tmall (https://ijcai-15.org/repeat-buyers-prediction-competition/, accessed on 22 April 2022), which are commonly used in session-based recommendation methods and the statistics of each dataset are shown in Table 1. Due to the size of the dataset, we used the recent fraction of 1/64 for the Yoochoose datasets. We keep only items appearing over five times and sessions whose length is more than one. The latest data are set to be a test dataset and the previous are set to be a training dataset. The label of each session is the last clicked item of it. The proposed method, Logit Averaging, works when the target of the sample is in the short-head items set and we split the items into short-head items and long-tail items. We sorted items by the number of item appearances in descending order. Then we chose the items as short-head, which are in the top 75% cumulative sum.

3.1.2. Baseline Methods

To compare the recommendation performance of the baseline methods and the proposed method, we employ six different baselines. Those can be categorized into two groups—attention-based methods and GNN-based methods. We utilize NARM [11] and SR-SAN [12] for the attention-based methods and SR-GNN [15], NISER [16], EOPA [17] and TAGNN++ [18] for the GNN-based methods. We set the hyperparameters of each baseline method as the baseline did. The specific hyperparameters used in the experiments are described in Table A2. The experimental results are shown in Table 2.

3.1.3. Evaluation Metrics

To evaluate the performance of all algorithms, we use three metrics—Recall@K, MRR (Mean Reciprocal Rank)@K, and Coverage@K. We set K to 10 and 20 for all the metrics.
  • Recall is the metric to measure recommendation accuracy. It is the proportion of correct prediction amongst top-k recommendation lists.
    R e c a l l = # h i t s | S | × 100 ,
    where # h i t s is the number of hits that the generated the top-K recommend list containing the ground truth and | S | is the number of test sessions.
  • Mean Reciprocal Rank (MRR) also measures recommendation accuracy taking ranks into consideration. MRR is the mean of ground truths’ reciprocal rank in the lists.
    M R R = 1 | S | i = 1 | S | 1 r a n k i × 100 ,
    where r a n k i is the reciprocal position of the ground truth in the top-K recommendation list.
  • Coverage is the ratio of the unique items that appears in recommendation lists.
    C o v e r a g e = | I r e c | | I | × 100 ,
    where | I r e c | is the number of unique items included in the top-k recommendation list, and | I | is the total number of unique items in the test dataset.

3.2. Experiment Result

Table 2 shows the performance of the baseline models with and without Logit Averaging (LA). The rows are the baselines and their combination with LA, and columns are the evaluation metrics. We evaluate the performance in both top-10 and top-20 recommendation lists. For each dataset, we calculate the average percentage of the performance gain in performance in all six baseline models.
At the bottom of Table 2, there is the Total Gain (%) that calculates the average percent of the gain in all four datasets. For all six metrics, HR@10, MRR@10, Coverage@10, HR@20, MRR@20, and Coverage@20, applying LA to the baselines improves the average performance by 0.48%, 1.86%, 2.78%, 0.26%, 1.78%, and 2.90%, respectively. This result proves the effectiveness of our proposed method in improving prediction accuracy and item diversity.
For RetailRocket dataset, HR@10, MRR@10, and Coverage@10 have increased by 0.86%, 3.23%, and 1.97%, respectively. HR@20, MRR@20, and Coverage@20 have increased by 0.54%, 3.14%, and 1.84%, respectively. Not only does prediction accuracy improve but also diversify the recommendation list.
In the case of the Tmall dataset, MRR@10 and Coverage@10 have increased by 0.86% and 2.02% while HR@10 decreases by −0.98%. Even though there has been a slight decrease in Recall, applying LA helps to make an accurate reciprocal ranking prediction and increases the item diversity.
For the Yoochoose 1 / 64 dataset, the performance of all three metrics, HR@10, MRR@10, and Coverage@10, improves by 0.15%, 1.18%, and 4.58%, respectively. It particularly increases item coverage more than other metrics. As Isufi et al. [35] pointed out, usually, recommendation accuracy is tied with diversity in a delicate trade-off. Thus, a great increase in item coverage along with recommendation accuracy when LA is applied is a remarkable achievement.
Additionally, for the Diginetica dataset, HR@10, MRR@10, and Coverage@10 increase by 0.48%, 2.18%, and 2.53% respectively when LA is applied to the original baselines. This trend can also be observed when K = 20 . The performance of HR@20, MRR@20, and Coverage@20 improves by 1.29%, 2.13%, and 2.47%.
The six baseline models have shown a different tendency to improve recommendation results depending on whether the model is based on an attention mechanism or GNNs. Firstly, the attention-based models, NARM and SR-SAN, show a large improvement in Hit Ratio and MRR after training with LA methods. SR-SAN gains 13.99% on MRR@10, but drops 3.2% on Coverage@10 on average. The GNN-based models, EOPA, NISER, SR-GNN, and TAGNN++, show the opposite tendency of attention-based models. The models applying LA show the steepest rise in coverage. SR-GNN and TAGNN++ gain 49.48% and 22.15% on Coverage@10, respectively.
We compare the performance of HR@20, MRR@20 and Coverage@20 in two different datasets in the same baseline model, TAGNN++ [18]. The result is shown in Figure 3. The salmon color bar is the performance of baseline models, and the yellow bar indicates the performance of baselines with our proposed method LA. Lastly, we applied a data augmentation technique, GraphMix [36], to further improve the performance, and the result is shown in the skyblue bar. For the two datasets, LA improves the result of HR@20, MRR@20, and Coverage@20. When GraphMix is used in addition to LA, it helps to increase prediction accuracy and item coverage in both datasets.

3.3. Ablation Study

For the ablation study, we conducted two different experiments. First, we applied various data augmentation techniques to our proposed method. Second, we compared the performance of our method and GCE-GNN. The experimental results are shown in Table 3 and Table 4, respectively.

3.3.1. Effectiveness of Augmentation

We utilized augmentation methods with four baselines in order to overcome the data sparsity problem in session datasets. We conduct four different data augmentation methods, which are GraphMix [36], FLAG [37], Random deletion, and Random insertion.
GraphMix [36] is proposed as a regularization method for GNN based semi-supervised object classification. It interpolates the hidden states and the corresponding labels to make a better feature representation. Even though it targets object classification tasks, we find this technique also applicable to session-based recommendation tasks as well. We interpolate the feature representations of two random sessions within a batch and let there be two target values, making it similar to a multi-label classification task. The equation is shown below.
F m i x = λ · F a + ( 1 λ ) · F b
L m = λ · C r o s s E n t r o p y L o s s ( p r e d , y a ) + ( 1 λ ) · C r o s s E n t r o p y L o s s ( p r e d , y b )
L = L o + L m ,
First, we interpolate two random feature representations of sessions a and b, F A and F b with respect to λ value. We set λ value as 0.6 in all experiments. Then the p r e d is calculated using the session model, and the loss value is calculated using cross entropy as in Equation (8). The loss value calculated from mixed logits, L m , is interpolated with respect to λ value as well. The final loss value is obtained in addition to the original loss L o and the modified one, L m .
FLAG [37] is introduced as a data augmentation technique, specifically for graph datasets to enhance the performance of the graph neural network through iteratively augmented node features with gradient-based adversarial perturbations during training.
Lastly, we execute a random deletion and random selection method, which is a straightforward data augmentation method widely used in the natural language processing (NLP) domain. For each session within a batch, we randomly delete or insert the items to sessions. For random insertion, we use a set of unique items within a batch to randomly select the one item to insert.
Table 3 shows the average percentage of the performance gain of different data augmentation methods in four baseline models, which are both attention-based (NARM, SRSAN) and GNN-based (NISER, TAGNN++). The raw result value of this experiment is shown in Table A1 in Appendix A.
Even though the results differ depending on the datasets, GraphMix and FLAG show an impressive performance, improving Recall, MRR and item coverage. Especially for the Tmall datset, GraphMix increases HR@20, MRR@20, and Coverage@20 by 2.15%, 0.97%, and 2.72%, respectively. However, applying random deletion and random insertion rather degrades the performance in almost all datasets. Thus, we could infer that GraphMix and FLAG are effective data augmentation techniques that could alleviate the data sparsity problem in session datasets, and further diversify the item recommendation lists.

3.3.2. Comparison on GCE-GNN

We compare the performance of our methods with GCE-GNN [22], which injects the global relation information by employing a global item transition graph. We conducted experiments using two datasets: Tmall and Yoochoose. GCE-GNN is the performance we obtain with the official code (https://github.com/CCIIPLab/GCE-GNN, accessed on 22 April 2022). As shown in Table 4, our method outperformed the existing method in overall performance. Although the coverage of our method was lower than that of GCE-GNN with Yoochoose datasets, TAGNN++ with LA and GraphMix augmentation improves all metrics by a large margin. Therefore, we demonstrate that our proposed methods, LA and the augmentation technique, are effective for improving recommendation performance.

4. Discussion

The recommender system is an essential part of an e-commerce platform to improve user experience and satisfaction. Most conventional recommendation models utilized user and item interaction information and side-information such as users’ explicit feedback and user profiles [3,4,5,6,7,8]. However, in many cases, users look around internet shopping websites without logging in and some users share the same account even if they logged in. Thus, the session-based recommender system, which utilizes anonymous session sequences to recommend items, has gained much attention.
Session-based recommendation aims to predict an anonymous user’s next action, whether she or he is likely to purchase based on the user’s clicked-item in the current session, as sequences. Most recent research on session-based recommendations makes predictions solely based on a single session without incorporating global relationships between sessions [15,16,17,18,19,22]. We argue that this method does not guarantee better performance because item embeddings learned by utilizing a single session (inter-session) have less item transition information than utilizing both intra- and inter-session item transitions simultaneously. Some existing methods [21,22,23,24,25,26,27] tried to enhance recommendation performance by adopting memory modules and global transition graphs; however, they still carry limitations in terms of computation cost. We proposed a novel yet simple algorithm called Logit Averaging (LA), helping to learn both local-level logits which come from intra-sessions’ item transitions and global-level logits, which come from gathered logits of related sessions. By considering the long-tail item distribution of session datasets, we stored and aggregated the logits of sessions with target items that the majority of other sessions shared. Then, we applied the mean value of logits as a global context that a single session might miss. Extensive experiments verified that the baseline models with our proposed method, LA (Logit Averaging), outperformed the baseline models. The total gains of LA with baselines on the baselines are all larger than zero, so we demonstrated that the LA module helps to increase both recommendation accuracy and diversity. Addtionally, we conducted additional experiments to confirm the improvement of the proposed methods with existing augmentation methods. The baseline methods with LA and GraphMix [36] showed the largest increment for all metrics.
Although we proposed the novel and simple method to overcome the challenges where session-based recommendation employs global item relations among the sessions, there are still a few limitations. In [9,24], sessions have temporal features, and user and items interactions and sessions represent consumption trends. Therefore, considering the temporal feature of sessions while collecting the global item transitions remains for our future work. Besides, the complex inter-relationships between sessions are still not entirely understood, and we think this is a fertile space for future exploration.

Author Contributions

Conceptualization, H.Y. and G.K.; methodology, H.Y. and G.K.; software, H.Y. and G.K.; formal analysis, H.Y. and G.K. and J.-H.L.; investigation, H.Y. and G.K.; data curation, H.Y. and G.K.; writing—original draft preparation, H.Y. and G.K.; writing—review and editing, H.Y. and G.K.; visualization, H.Y.; supervision, J.-H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by MSIT (Ministry of Science and ICT), Korea, under the ICT Creative Consilience program IITP-2021-2020-0-01821 and Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2019-0-00421, Artificial Intelligence Graduate School Program (Sungkyunkwan University).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. RetailRocket dataset can be found here: [https://www.kaggle.com/retailrocket/ecommerce-dataset, accessed on 22 April 2022], Tmall dataset can be found here: [https://ijcai-15.org/repeat-buyers-prediction-competition/, accessed on 22 April 2022], Yoochoose dataset can be found here: [https://www.kaggle.com/chadgostopp/recsys-challenge-2015, accessed on 22 April 2022], Diginetica dataset can be found here: [https://competitions.codalab.org/competitions/11161#learn_the_details-data2, accessed on 22 April 2022].

Acknowledgments

This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ICT Creative Consilience program (IITP-2021-2020-0-01821) supervised by the IITP (Institute for Information & communications Technology Planning & Evaluation). This work was supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2019-0-00421, Artificial Intelligence Graduate School Program (Sungkyunkwan University).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1

Table A1 shows the raw result value of different data augmentation methods in four different datasets and four different baseline models. For data augmentation experiment, we use NARM [11], SRSAN [12], NISER [16], and TAGNN++ [18] to evaluate the performance.
Table A1. The result of different data augmentation methods in four different models. In column LA, we mark ’X’ to indicate the baseline without using LA and mark ’O’ to indicate the baseline with using LA.
Table A1. The result of different data augmentation methods in four different models. In column LA, we mark ’X’ to indicate the baseline without using LA and mark ’O’ to indicate the baseline with using LA.
DatasetModelAugLAHR@10MRR@10Coverage@10HR@20MRR@20Coverage@20
RetailRocketNARMXX39.2420.7695.5245.9221.2497.63
XO39.0221.0795.8645.6421.5397.80
GraphMixO37.8919.4695.9944.7619.9498.14
FLAGO39.9421.1695.5446.2121.6497.75
Random deletionO39.8319.4895.1846.6819.9597.50
Random InsertionO38.9719.4294.8846.4119.9497.17
NISERXX33.2422.2488.6637.5422.5393.42
XO33.3322.0988.5937.6622.3993.39
GraphMixO33.1922.3785.6537.4922.6690.79
FLAGO33.4022.4589.5637.7622.7594.40
Random deletionO34.7117.5196.4439.4517.8398.57
Random InsertionO32.2716.3596.8036.4016.6498.83
SR-SANXX37.9617.2593.9744.9117.7496.53
XO37.5017.2692.4444.3317.7595.45
GraphMixO33.8920.1380.2239.1420.4984.66
FLAGO37.7617.1793.3244.5617.6596.26
Random deletionO37.4217.0191.9045.5717.5894.86
Random InsertionO35.0615.9790.5342.6116.4894.10
TAGNN++XX34.3021.5481.5939.8321.9285.47
XO34.1021.7182.2339.6122.0986.21
GraphMixO38.0023.3690.9643.5823.7593.47
FLAGO38.3523.1889.4744.4723.6192.62
Random deletionO36.0719.5182.4042.9319.9886.46
Random InsertionO31.5916.8077.4238.2917.2782.76
TmallNARMXX17.1412.5217.6819.1812.7128.68
XO16.4612.2117.2818.5012.3528.17
GraphMixO19.1813.7418.5621.6213.8830.85
FLAGO16.6512.4117.8118.7012.5128.84
Random deletionO18.3113.2518.2319.8613.3829.60
Random InsertionO18.7012.9118.5020.7413.0530.28
NISERXX18.8914.0418.9321.0314.1831.83
XO18.1113.5219.2120.3513.6032.43
GraphMixO19.0813.7618.6520.4513.8531.08
FLAGO19.1813.8019.0321.1313.9332.01
Random deletionO18.3113.5818.8219.7713.6831.97
Random InsertionO16.4612.6618.8618.2112.7831.83
SR-SANXX18.9912.7412.6721.6212.9219.65
XO20.5514.2512.6323.1714.4319.54
GraphMixO22.4014.6516.2724.8314.8326.07
FLAGO19.2812.6212.4922.0112.8119.30
Random deletionO19.2811.8413.1222.5912.0320.45
Random InsertionO17.9210.5513.0020.8410.7620.12
TAGNN++XX33.1116.1715.8839.8216.6425.01
XO30.7714.5516.3636.4214.9625.42
GraphMixO34.2717.1117.5043.3317.7628.06
FLAGO34.8615.3816.5645.4716.1426.01
Random deletionO30.4814.1716.7536.5114.6026.20
Random InsertionO21.8110.5915.4426.5810.9423.52
Yoochoose 1 / 64 NARMXX59.9529.4629.2470.6730.2035.73
XO59.9629.6528.8870.6130.4035.37
GraphMixO55.6928.0836.9166.6628.8545.52
FLAGO60.5029.5729.8371.0730.3335.89
Random deletionO59.9929.3829.1770.5430.1335.75
Random InsertionO60.4929.6830.5071.0030.4237.09
NISERXX60.2530.9129.4670.5531.6434.61
XO60.2330.8829.4070.4831.6034.73
GraphMixO60.5831.0828.6570.5431.7833.29
FLAGO60.4530.7729.9170.7131.4935.34
Random deletionO60.3629.0431.8270.4129.7438.69
Random InsertionO59.8829.1834.1769.8829.8842.22
Yoochoose 1 / 64 SR-SANXX53.9025.9628.2864.8926.7332.62
XO53.7526.0028.3164.9126.7732.63
GraphMixO54.8726.4830.1966.1427.2734.88
FLAGO54.3126.1028.4065.2426.8632.86
Random deletionO55.5425.8528.5367.2126.6532.61
Random InsertionO52.2823.3628.9965.5424.2733.08
TAGNN++XX60.7030.6326.4870.9331.3530.72
XO60.8431.1128.7971.1631.8333.46
GraphMixO60.7530.8627.6471.1531.5932.27
FLAGO60.6730.9527.8271.0831.6831.72
Random deletionO58.3229.4316.7968.7929.1718.96
Random InsertionO57.3227.8817.0368.4428.6519.12
DigineticaNARMXX35.8618.6068.7844.8319.2282.60
XO35.7618.5169.0544.4919.1283.07
GraphMixO34.7618.3869.0243.3918.9783.16
FLAGO36.3018.4969.0645.4019.1282.91
Random deletionO35.9618.6168.9545.3219.2582.66
Random InsertionO36.5918.7168.3145.5919.3682.13
NISERXX34.1217.9865.0241.8918.5277.34
XO33.7918.0265.2841.6818.5778.25
GraphMixO34.2718.0659.2342.0418.5970.59
FLAGO34.0418.1065.4141.6918.6378.41
Random deletionO33.5518.3764.1839.9518.8177.65
Random InsertionO33.1118.2864.0439.6818.7477.69
SR-SANXX32.5416.6167.6041.5017.2278.56
XO32.9316.9366.6842.0217.5677.34
GraphMixO32.9416.9570.6641.7217.5681.55
FLAGO32.6216.6267.5741.2817.2278.39
Random deletionO32.0714.8567.2142.4415.5879.28
Random InsertionO31.5014.0466.7841.8714.7679.36
TAGNN++XX33.9017.3462.8142.6417.9574.00
XO36.4218.3168.8545.4118.9379.83
GraphMixO36.3618.4370.8945.0219.0380.91
FLAGO33.7717.3662.5442.2917.9673.66
Random deletionO34.1715.3766.9343.6116.0378.87
Random InsertionO34.2018.0263.9941.9718.5676.07

Appendix A.2

Table A2 shows the hyperparameter settings that we executed. All the setting is same from the original implementation codes that authors of the paper published. For Yoochoose 1 / 64 and Diginetica which have much larger train sessions, some modification on batch size in TAGNN++ were necessary. Otherwise, the memory error occurred. Especially when experimenting Random Deletion and Random Insertion augmentation techniques, batch size of Yoochoose 1 / 64 was 8 and 16 for diginetica.
Table A2. The hyperparameter settings of six baseline models.
Table A2. The hyperparameter settings of six baseline models.
HyperparameterModel
NARMEOPANISERSRGNNSRSANTAGNN++
Batch Size12812812812812864
Hidden Size100-10010096100
Epoch1003030303030
L2 Penalty-0.00010.000010.000010.000010.00001
Learning rate0.0010.0010.0010.0010.0010.001
Learning rate decay rate0.1-0.10.10.10.1
Number of steps after which the learning rate decay80-3333
The dropout ratio for features-0.2----
GNN Propagation--1111
Number of SAN layer----1-
Number of heads of multi-head attention----2-
Multipler of hidden size----1-
Embedding Dimension5032-- -
Number of layers13----

Appendix A.3

Figure A1 shows the performance differences depending on the various head-tail ratio setting. We examine the performance using SRGNN model with Diginetica dataset. The legend in each plot represents the recommendation performance of metrics. We found that the performance on both HR and MRR depending on the head-tail ratio do not differ very much. The performance differences were within ± 1 , meaning that head-tail ratio are not significant hyperparmeter in Logit Averaging method.
Figure A1. We compare the performance of LA in various head ratio from 5% to 30%. The legend shows the metrics. We examine the performance in SRGNN model using Digineitca dataset.
Figure A1. We compare the performance of LA in various head ratio from 5% to 30%. The legend shows the metrics. We examine the performance in SRGNN model using Digineitca dataset.
Applsci 12 04256 g0a1

References

  1. Zhang, S.; Yao, L.; Sun, A.; Tay, Y. Deep learning based recommender system: A survey and new perspectives. arXiv 2019, arXiv:1707.07435. [Google Scholar] [CrossRef] [Green Version]
  2. Covington, P.; Adams, J.; Sargin, E. Deep neural networks for youtube recommendations. In Proceedings of the 10th ACM Conference on Recommender Systems, Boston, MA, USA, 15–19 September 2016; pp. 191–198. [Google Scholar]
  3. Sarwar, B.; Karypis, G.; Konstan, J.; Riedl, J. Item-Based Collaborative Filtering Recommendation Algorithms. In Proceedings of the 10th International Conference on World Wide Web, Hong Kong, China, 1–5 May 2001; pp. 285–295. [Google Scholar]
  4. Rendle, S.; Freudenthaler, C.; Schmidt-Thieme, L. Factorizing Personalized Markov Chains for Next-Basket Recommendation. In Proceedings of the 19th International Conference on World Wide Web, Raleigh, NC, USA, 26–30 April 2010; pp. 811–820. [Google Scholar]
  5. He, X.; Liao, L.; Zhang, H.; Nie, L.; Hu, X.; Chua, T. Neural Collaborative Filtering. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 173–182. [Google Scholar]
  6. Chen, Y.; de Rijke, M. A collective variational autoencoder for top-n recommendation with side information. In Proceedings of the 3rd Workshop on Deep Learning for Recommender Systems, Vancouver, BC, Canada, 6 October 2018; pp. 3–9. [Google Scholar]
  7. Dong, X.; Yu, L.; Wu, Z.; Sun, Y.; Yuan, L.; Zhang, F. A hybrid collaborative filtering model with deep structure for recommender systems. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  8. Liu, F.; Cheng, Z.; Zhu, L.; Gao, Z.; Nie, L. Interest-aware message-passing gcn for recommendation. In Proceedings of the Web Conference 2021, Ljubljana, Solvenia, 19–23 April 2021; pp. 1296–1305. [Google Scholar]
  9. Choi, M.; Kim, J.; Lee, J.; Shim, H.; Lee, J. Session-Aware Linear Item-Item Models for Session-Based Recommendation. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 2186–2197. [Google Scholar]
  10. Wang, S.; Cao, L.; Wang, Y.; Sheng, Q.Z.; Orgun, M.A.; Lian, D. A Survey on Session-Based Recommender Systems. arXiv 2021, arXiv:1902.04864. [Google Scholar] [CrossRef]
  11. Li, J.; Ren, P.; Chen, Z.; Ren, Z.; Lian, T.; Ma, J. Neural attentive session-based recommendation. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 1419–1428. [Google Scholar]
  12. Fang, J. Session-based Recommendation with Self-Attention Networks. arXiv 2021, arXiv:2102.01922. [Google Scholar]
  13. Cho, K.; Merrienboer, V.B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar]
  14. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the 5th International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  15. Wu, S.; Tang, Y.; Zhu, Y.; Wang, L.; Xie, X.; Tan, T. Session-based recommendation with graph neural networks. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019. [Google Scholar]
  16. Gupta, P.; Garg, D.; Malhotra, P.; Vig, L.; Shroff, G. NISER: Normalized item and session representations to handle popularity bias. In Proceedings of the 1st International Workshop on Graph Representation Learning on Its Applications (CIKM ’19), Beijing, China, 3–7 November 2019. [Google Scholar]
  17. Chen, T.; Wong, R.C.W. Handling information loss of graph neural networks for session-based recommendation. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual Event, 23–27 August 2020; pp. 1172–1180. [Google Scholar]
  18. Mitheran, S.; Java, A.; Sahu, S.K.; Shaikh, A. Introducing Self-Attention to Target Attentive Graph Neural Networks. arXiv 2022, arXiv:2107.01516. [Google Scholar]
  19. Li, A.; Cheng, Z.; Liu, F.; Gao, Z.; Guan, W.; Peng, Y. Disentangled Graph Neural Networks for Session-based Recommendation. arXiv 2022, arXiv:2201.03482. [Google Scholar]
  20. Li, Y.; Tarlow, D.; Brockschmidt, M.; Zemel, R.S. Gated Graph Sequence Neural Networks. In Proceedings of the 4th International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  21. Wang, M.; Ren, P.; Mei, L.; Chen, Z.; Ma, J.; de Rijke, M. A Collaborative Session-Based Recommendation Approach with Parallel Memory Modules. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 345–354.
  22. Wang, Z.; Wei, W.; Cong, G.; Li, X.L.; Mao, X.L.; Qiu, M. Global context enhanced graph neural networks for session-based recommendation. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 25–30 July 2020; pp. 169–178. [Google Scholar]
  23. Mi, F.; Faltings, B. Memory Augmented Neural Model for Incremental Session-Based Recommendation. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, Yokohama, Japan, 7–15 January 2020; pp. 2169–2176. [Google Scholar]
  24. Zhou, H.; Tan, Q.; Huang, X.; Zhou, K.; Wang, X. Temporal Augmented Graph Neural Networks for Session-Based Recommendations. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 11–15 July 2021; pp. 1798–1802. [Google Scholar]
  25. Huang, C.; Xia, X.; Dai, C.; Bo, Z. Graph-Enhanced Multi-Task Learning of Multi-Level Transition Dynamics for Session-based Recommendation. In Proceedings of the 35th AAAI Conference on Artificial Intelligence, Virtual Event, 2–9 February 2021; pp. 4123–4130. [Google Scholar]
  26. Xia, X.; Yin, H.; Yu, J.; Shao, Y.; Cui, L. Self-Supervised Graph Co-Training for Session-based Recommendation. Assoc. Comput. Mach. 2021, 11, 2180–2190. [Google Scholar]
  27. Pang, Y.; Wu, L.; Shen, Q.; Zhang, Y.; Wei, Z.; Xu, F.; Chang, E.; Long, B.; Pei, J. Heterogeneous global graph neural networks for personalized session-based recommendation. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, Virtual Event, 21–25 February 2022; pp. 775–783. [Google Scholar]
  28. Menon, A.; Jayasumana, S.; Rawat, A.; Jain, H.; Veit, A.; Kumar, S. Long-tail learning via logit adjustment. In Proceedings of the 2021 International Conference on Learning Representation, Virtual Conference, 3–7 May 2021. [Google Scholar]
  29. Hurley, N.; Zhang, M. Novelty and Diversity in Top-N Recommendation—Analysis and Evaluation. ACM Trans. Internet Technol. 2011, 10, 14. [Google Scholar] [CrossRef]
  30. Bradley, K.; Smyth, B. Improving recommendation diversity. In Proceedings of the Twelfth Irish Conference on Artificial Intelligence and Cognitive Science, Maynooth, Ireland, 1 September 2001; pp. 141–152. [Google Scholar]
  31. Miyamoto, S.; Zamami, T.; Yamana, H. Appearance frequency-based ranking method for improving recommendation diversity. In Proceedings of the 2019 IEEE 4th International Conference on Big Data Analytics (ICBDA), Suzhou, China, 15–18 March 2019; pp. 420–425. [Google Scholar]
  32. Wang, L.; Zhang, X.; Wang, T.; Wan, S.; Srivastava, G.; Pang, S.; Qi, L. Diversified and scalable service recommendation with accuracy guarantee. IEEE Trans. Comput. Soc. Syst. 2020, 8, 1182–1193. [Google Scholar] [CrossRef]
  33. Anderson, C. The Long Tail: Why the Future of Business is Selling Less of More; Hachette Books: London, UK, 2006. [Google Scholar]
  34. Park, Y.J.; Tuzhilin, A. The long tail of recommender systems and how to leverage it. In Proceedings of the 2008 ACM Conference on Recommender Systems, Lausanne, Switzerland, 23–25 October 2008; pp. 11–18. [Google Scholar]
  35. Isufi, E.; Pocchiari, M.; Hanjalic, A.T. Accuracy-diversity trade-off in recommender systems via graph convolutions. Inf. Process. Manag. 2021, 58, 102459. [Google Scholar] [CrossRef]
  36. Wang, Y.; Wang, W.; Liang, Y.; Cai, Y.; Hooi, B. Mixup for Node and Graph Classification. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 12–16 April 2021; pp. 3663–3674. [Google Scholar]
  37. Kong, K.; Li, G.; Ding, M.; Wu, Z.; Zhu, C.; Ghanem, B.; Taylor, G.; Goldstein, T. Flag: Adversarial data augmentation for graph neural networks. arXiv 2020, arXiv:2010.09891. [Google Scholar]
Figure 1. General deep-learning-based session-based recommendation models framework with the proposed methods.
Figure 1. General deep-learning-based session-based recommendation models framework with the proposed methods.
Applsci 12 04256 g001
Figure 2. Long-tail distribution of real-world session datasets. From the left, the datasets are Diginetica, Yoochoose, Retailrocket, Tmall. The red part represents short-head and the blue part represents long-tail.
Figure 2. Long-tail distribution of real-world session datasets. From the left, the datasets are Diginetica, Yoochoose, Retailrocket, Tmall. The red part represents short-head and the blue part represents long-tail.
Applsci 12 04256 g002
Figure 3. Comparing performance in two different datasets in the TAGNN++ model. Compare to baseline model performance, applying LA (Logit Averaging) improves performances in every metrics. In addition to LA, applying GraphMix, a data augmentation technique, improves performance of HR@20 and MRR@20.
Figure 3. Comparing performance in two different datasets in the TAGNN++ model. Compare to baseline model performance, applying LA (Logit Averaging) improves performances in every metrics. In addition to LA, applying GraphMix, a data augmentation technique, improves performance of HR@20 and MRR@20.
Applsci 12 04256 g003
Table 1. Data Statistics.
Table 1. Data Statistics.
StatisticsDigineticaRetailrocketYoochoose 1 / 64 Tmall
# Train Sessions186,670113,287124,47265,286
# Test Sessions15,93671,23515,3241027
# Items43,09827,41337,48440,728
Avg. Length4.853.663.976.68
Table 2. The performance result with and without LA (Logit Averaging) method. In column LA, we mark ’X’ to indicate the baseline without using LA and mark ’O’ to indicate the baseline with using LA. Gain (%) is the average gain of the baselines with LA on the original baselines.
Table 2. The performance result with and without LA (Logit Averaging) method. In column LA, we mark ’X’ to indicate the baseline without using LA and mark ’O’ to indicate the baseline with using LA. Gain (%) is the average gain of the baselines with LA on the original baselines.
DatasetModelLAHR@10MRR@10Coverage@10HR@20MRR@20Coverage@20
RetailRocketNARMX39.2420.7695.5245.9221.2497.63
EOPA33.7417.1381.0742.5117.7286.91
NISER33.2422.2488.6637.5422.5393.42
SR-GNN31.9818.8176.0937.2419.1780.40
SR-SAN37.9617.2593.9744.9117.7496.53
TAGNN++34.3021.5481.5939.8321.9285.47
NARMO39.0221.0795.8645.6421.5397.80
EOPA33.6317.3381.2842.2817.9186.98
NISER33.3322.0988.5937.6622.3993.39
SR-GNN34.4021.9285.3439.4522.2789.29
SR-SAN37.5017.2692.4444.3317.7595.45
TAGNN++34.1021.7182.2339.6122.0986.21
Gain (%)0.863.231.970.543.141.84
TmallNARMX17.1412.5217.6819.1812.7128.68
EOPA19.589.7032.2724.6410.0244.43
NISER18.8914.0418.9321.0314.1831.83
SR-GNN16.5511.0013.7918.7911.1721.52
SR-SAN18.9912.7412.6721.6212.9219.65
TAGNN++33.1116.1715.8839.8216.6425.01
NARMO16.4612.2117.2818.512.3528.17
EOPA19.439.8131.6224.1410.1443.66
NISER18.1113.5219.2120.3513.6032.43
SR-GNN16.8511.9215.4719.6712.1124.81
SR-SAN20.5514.2512.6323.1714.4319.54
TAGNN++30.7714.5516.3636.4214.9625.42
Gain (%) −0.980.862.02−0.920.712.46
Yoochoose 1 / 64 NARMX59.9529.4629.2470.6730.2035.73
EOPA51.8324.9128.8162.8525.6831.74
NISER60.2530.9129.4670.5531.6434.61
SR-GNN59.5629.4723.1969.8730.1926.54
SR-SAN53.9025.9628.2864.8926.7332.62
TAGNN++60.7030.6326.4870.9331.3530.72
NARMO59.9629.6528.8870.6130.4035.37
EOPA51.5424.9029.0262.6425.6831.73
NISER60.2330.8829.4070.4831.6034.73
SR-GNN60.4630.9027.6870.5531.6132.06
SR-SAN53.7526.0028.3164.9126.7732.63
TAGNN++60.8431.1128.7971.1631.8333.46
Gain (%) 0.151.184.580.141.154.84
DigineticaNARMX35.8618.6068.7844.8319.2282.60
EOPA34.2514.5262.0947.0915.3775.91
NISER34.1217.9865.0241.8918.5277.34
SR-GNN32.8416.8360.9541.4617.4371.44
SR-SAN32.5416.6167.6041.5017.2278.56
TAGNN++33.9017.3462.8142.6417.9574.00
NARMO35.7618.5169.0544.4919.1283.07
EOPA34.5414.6062.3347.0315.4775.81
NISER33.7918.0265.2841.6818.5778.25
SR-GNN33.8517.7264.4742.0318.2976.34
SR-SAN32.9316.9366.6842.0217.5677.34
TAGNN++36.4218.3168.8545.4118.9379.83
Gain (%) 1.882.182.531.292.132.47
Total Gain (%) 0.481.862.780.261.782.90
Table 3. The average performance gains (%) of four baseline methods with LA and different data augmentation methods on the original baseline methods.
Table 3. The average performance gains (%) of four baseline methods with LA and different data augmentation methods on the original baseline methods.
DatasetAugmentationLAHR@10MRR@10Coverage@10HR@20MRR@20Coverage@20
RetailRocketXO−0.200.09−0.15−0.240.08−0.05
GraphMixO−0.440.88−1.73−0.810.85−1.50
FLAGO1.180.542.041.200.562.00
Random deletionO0.82−2.071.551.61−2.021.09
Random InsertionO−1.71−3.31−0.03−1.12−3.28−0.05
TmallXO−0.56−0.240.08−0.80−0.280.10
GraphMixO1.700.951.462.150.972.72
FLAGO0.46−0.320.181.42−0.270.25
Random deletionO−0.44−0.660.44−0.73−0.690.76
Random InsertionO−3.31−2.190.16−3.82−2.230.15
Yoochoose 1 / 64 XO−0.010.170.480.030.170.63
GraphMixO−0.73−0.122.48−0.64−0.113.07
FLAGO0.280.110.620.260.110.53
Random deletionO−0.15−0.82−1.79−0.02−1.06−1.92
Random InsertionO−1.21−1.72−0.69−0.55−1.68−0.54
DigineticaXO0.620.311.410.690.321.50
GraphMixO0.480.321.400.330.310.93
FLAGO0.080.010.09−0.050.010.22
Random deletionO−0.17−0.830.770.12−0.811.49
Random InsertionO−0.25−0.37−0.27−0.44−0.370.69
Table 4. The performance comparison on our proposed model with TAGNN++ and GCE-GNN using Tmall and Yoochoose 1 / 64 datasets.
Table 4. The performance comparison on our proposed model with TAGNN++ and GCE-GNN using Tmall and Yoochoose 1 / 64 datasets.
DatasetModelHR@10MRR@10Coverage@10HR@20MRR@20Coverage@20
TmallGCE-GNN23.7615.2016.1826.2915.3525.71
TAGNN++30.7714.5516.3636.4214.9625.42
TAGNN++ w. GraphMix34.2717.1117.5043.3317.7628.06
Yoochoose 1 / 64 GCE-GNN60.3229.6629.7971.0230.4034.98
TAGNN++60.8431.1128.7971.1631.8333.46
TAGNN++ w. GraphMix60.7530.8627.6471.1531.5932.27
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, H.; Kim, G.; Lee, J.-H. Logit Averaging: Capturing Global Relation for Session-Based Recommendation. Appl. Sci. 2022, 12, 4256. https://doi.org/10.3390/app12094256

AMA Style

Yang H, Kim G, Lee J-H. Logit Averaging: Capturing Global Relation for Session-Based Recommendation. Applied Sciences. 2022; 12(9):4256. https://doi.org/10.3390/app12094256

Chicago/Turabian Style

Yang, Heeyoon, Gahyung Kim, and Jee-Hyoung Lee. 2022. "Logit Averaging: Capturing Global Relation for Session-Based Recommendation" Applied Sciences 12, no. 9: 4256. https://doi.org/10.3390/app12094256

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop