Abstract
To predict what products customers will buy in next transaction is an important task. Existing work in next-basket prediction can be summarized into two paradigms. One is the item-centric paradigm, where sequential patterns are mined from customers’ transactional data and leveraged for prediction. However, these approaches usually suffer from the data sparseness problem. The other is the user-centric paradigm, where collaborative filtering techniques have been applied on customers’ historical data. However, these methods ignore the sequential behaviors of customers which is often crucial for next-basket prediction. In this paper, we introduce a hybrid method, namely the Co-Factorization model over Sequential and Historical purchase data (CFSH for short) for next-basket recommendation. Compared with existing methods, our approach conveys the following merits: 1) By mining global sequential patterns, we can avoid the sparseness problem in traditional item-centric methods; 2) By factorizing product-product and customer-product matrices simultaneously, we can fully exploit both sequential and historical behaviors to learn customer and product representations better; 3) By using a hybrid recommendation method, we can achieve better performance in next-basket prediction. Experimental results on three real-world purchase datasets demonstrated the effectiveness of our approach as compared with the state-of-the-art methods.
Introduction
Market basket analysis aims to discover meaningful patterns from massive customers’ purchase data [1]. It helps retailers to analyze the selling trends, to optimize the deployment of goods, and to understand customers’ preferences. With the prevalence of mobile applications and online e-commerce systems, market basket analysis becomes even more important in stimulating the consumptions and enlarging the selling profits, by providing the key technologies for personalized next-basket recommendation.
Generally, existing approaches to next-basket recommendation can be summarized into two paradigms. One is the item-centric paradigm. The key idea in the paradigm is “customers bought one product is also likely to buy some other products”. This recommendation paradigm has been applied to many e-commerce services such as Amazon. A number of approaches have been proposed to mine the meaningful sequential patterns from customers’ transactional data [2–4], which are similar to that of mining the association rules in data mining. However, the rule-based methods suffer from the data sparseness problem: the number of mined sequential patterns are always limited and it is hard to generalize the patterns to new products and users in real recommendation tasks. Although factorization methods have been proposed and applied on sequential patterns [5], these methods rely on individual sequential patterns and the data sparseness problem is not fully addressed.
Another approach is the user-centric paradigm. The key idea is “one is likely to buy the products favored by similar customers”. Collaborative filtering techniques have been applied [6–10]. A typical way is to represent customers’ historical purchase behaviors as a customer-product matrix where each entry represents the co-occurrence of the corresponding of customer and product in historical transaction. Matrix factorization is applied to learn the low-dimensional representations for both customers and products. In the method, the transactional sequence information is lost in the customer-product matrix. Therefore, the recommendation is made based on customers’ general interests. It is hard to capture the sequential purchase behaviors of customers, which is often crucial for next-basket prediction.
In this paper, we propose our hybrid method for the next-basket recommendation, namely Co-Factorization model over Sequential and Historical purchase data (CFSH for short). Specifically, on one hand, sequential pattern mining methods are applied to the massive transactional data to obtain the purchase sequential patterns. The mined patterns are represented as a product-product matrix. On the other hand, a customer-product matrix is constructed for representing the customers’ historical purchase behaviors, as that of in the conventional user-centric methods. These two matrices are then simultaneously factorized to learn the low-dimensional representations of both customers and products. With the learned representations, the next-basket prediction overcomes the problems that the previous approaches suffer from.
Compared with existing next-basket recommendation methods, our approach has the following advantages:
By mining and factorizing global sequential patterns, it avoid the sparseness problem in traditional item-centric methods, which rely on the limited number of individual sequential patterns;
By factorizing the product-product and customer-product matrices simultaneously, our approach fully exploits both sequential and historical behaviors to learn better representations for both customers and products;
By adopting a hybrid recommendation method, our approach enjoys the advantages from both the user-centric paradigm and the item-centric paradigm, and thus achieved better performances in real recommendation tasks.
We conducted experiments over three real world purchase data sets: two from retailers and one from the e-commerce. Compared with the state-of-the-art baseline methods including the methods based on sequential pattern mining and the methods based on collaborative filtering, our approach performed significantly better. The results demonstrated the effectiveness of our approach in real world next-basket recommendation task.
Related work
In this section, we provide background for basket recommendation. Two widely used recommendation models in market basket analysis, namely item-centric model and user-centric model, are introduced.
Item-centric model
Sequential patterns have been widely observed in customers’ purchase behaviors and they are greatly useful in basket analysis. For example, when customers bought cameras, they probably bought SD-cards for the cameras in the next transaction. As an item-centric model, sequence mining is originally introduced for market basket analysis where the temporal relations between retail transactions are mined. Sequence mining is then extended to many other complex domains such as telecommunication, network detection, etc [11]. It is a topic of data mining concerned with finding statistically relevant patterns between data where the values are delivered in a sequence. Therefore, given the criteria support and confidence [12], early work on sequence mining algorithms like ApriorALL, GSP and SPADE are designed for mining frequent sequence of products [13]. These works focus on involving temporal dynamics into recommendations and lots of interesting patterns have been discovered [2]. In the past decades, many shopping malls have adopted sequential pattern analysis to discover temporal associations across transactions [10]. All these models focus on the observed sequential patterns, thus face the problem of data sparseness.
Recently, another simple but popular way used to model sequential patterns is concerning Markov assumption on customers’ sequential behaviors. These works factorize product-product matrix based on the mined sequential patterns to describe product transition in customers’ purchase behaviors [14]. Much of work focuses on learning representations for customers and products. Wang et al. [15] concern both intra- and inter-association patterns, and design a generative topic model to describe patterns’ distribution on a n-dimensional shopping interest space. Christidis et al. [7] explore the use of probabilistic topic models on transaction product sets to learn representations of both customers and products. Chen et al. [16] uses an n-gram model to predict which music customer will listen next, and to construct a playlist automatically given a seed music. Koenigstein et al. [3] describes product-product co-occurrences based on Markov chains. In another recent paper, Rendle et al. [5] use customer-specific Markov chains to model customers’ selection of products (FPMC for short). This model faces the data sparsity problem when factorizing the personal sequential patterns. The model assumes that all sequential patterns have the same weight for prediction. No statistics is conducted to validate whether the patterns are from the frequent product set. It is hard to ensure that intrinsic patterns are discovered.
In summary, most of the research regarding to sequential pattern mining focus on customers’ temporal behaviors. The approach leads to the following two difficulties: 1) Sequential pattern mining methods face the data sparse problem, and can only make recommendations base on observed rules. The ability of making personalized recommendation is limited. 2) Models based on sequential patterns did not consider historical purchase data, which makes the recommendation results often inaccurate and biased. Previous work uses a fixed size of sliding window to preserve historical products and generate recommendations [17]. However, it still fails to capture customer‘s general interest to products well. How to model customers’ historical data is still a big challenge in the approach.
User-centric model
User-centric model is another widely used technique in market-basket analysis. Collaborative filtering is often used in the method to analyze customers’ historical purchase data. It can be further categorized into the memory-based approach and the model-based approach [18]. The memory-based approach provides the recommendation by studying the similarities among customers or products [4]. As an example, in e-commerce, retailers recommend one product to a customer because similar customers also purchases the product. The model-based approach tries to find a low-rank approximation of the customer-product rating matrix, and uses the values in the approximated matrix to recommend products [19]. As an example of model-based approach, matrix factorization is gaining rising attention in both explicit and implicit feedback applications such as Netflix [20, 21]. Many studies have been conducted. Lee [9] designed a binary customer-product matrix, and viewed the prediction problem as a two-class classification problem. Rendle et al. [22] assume that customers prefer the bought products than the ignored ones. They constructed product pairs as the training data and optimize for getting a correctly ranked product list rather than for assigning ranking scores for single products.
In many cases, however, the customers’ purchasing behaviors evolve over time. It is not reasonable to simply utilize the collaborative filtering methods for personalized recommendation, as the products purchased at different time periods might be significantly different. For example, a customer is likely to purchase candle and cake around her birthday, while she may be interested in electronic products in other days. Traditional collaborative filtering algorithms which describing customers’ preference based on customers’ historical purchase data fail to capture the evolution of customers’ purchasing interests effectively.
In conclusion, collaborative filtering seldom concerns the influence of customer‘s temporal interests to customers’ next purchase behaviors, and thus may limit the accuracy of prediction.
Our framework
In this paper, we propose to factorize both customers’ sequential and historical purchase data for next-basket recommendation. By adopting such a hybrid recommendation method, our approach enjoys the advantages of both item-centric and user-centric paradigms. In the same time, these two paradigms complement each other and can achieve better performances. In this section, we will present the proposed method, namely the Co-Factorization model over Sequential and Historical purchase data (CFSH for short) in detail.
Specifically, we will first describe how to factorize the two matrices which are constructed based on the sequential and historical data for next-basket recommendation, respectively. The hybrid model CFSH is then presented based on the above two recommendation paradigms. Finally, we present the optimization algorithm for the proposed hybrid recommendation model.
Notations
Let I = {i1, i2, …, i|I|} denotes the set of products, where |I| denotes the total number of unique products. Let U = {u1, u2, …, u|U|} denotes the set of customers, where |U| denotes the number of unique users. In transactional data, we use to denote the transactions corresponding to customer um, ordered by the transaction-time, where |Tm| denotes the number of transactions associated with user um. Moreover, denotes the number of products involved in the k − th transaction of customer um. We use rm,n ∈ R to denote the times the product in ∈ I was purchased by the customer um ∈ U in the customer’s transactional data.
Factorizing sequential purchase data
The item-centric paradigm has been widely adopted in next-basket prediction. The key idea is “who bought one product is likely to buy another”. Thus, it is critical to find the correlated products in this paradigm. To achieve the purpose, many approaches have been proposed to mine meaningful sequential patterns from customers’ transactional data [2–4], in a way similar to association rule mining in data mining.
However, an obvious drawback of traditional sequential modeling methods is the data sparseness, i.e., we statistic personal sequential patterns on a retail dataset with the weight larger than 1, we found that every customer obtained only 9.4 patterns. In our work, we propose to mine the sequential patterns in a global way, and factorize the mined patterns to obtain low dimensional production representations for better personalized recommendation. Here we first give the definition of sequential patterns in transactional data.
Definition 1 Sequential Pattern Given the transaction set of customer um, Sequential Pattern is defined as a weighted pair of products <ia, ib, wab>, where , , m < n, and wab denotes the support of sequential pattern ia ⇒ ib.
Existing work on sequential patterns focuses on the contiguous sequential pattern (CSP for short), by restricting the patterns mined from consecutive transactions of each individual customer [5, 12, 23]. Fig 1 shows an example. In this figure, a customer has three transactional records. Six CSPs are generated based on the records. It is obvious that the CSPs capture the local dependency of customer’s purchase behaviors. For example, a customer would probably buy a sim card in the next transaction if she bought a phone in the previous transaction. The mined CSPs can be represented with a matrix W, where element wa,b corresponds to the pattern <ia, ib, wab>.
Fig 1. Contiguous patterns mined from a single customer, with quite a few patterns mined.
The CSPs mined from individual purchase data, however, are extremely sparse. Therefore, simply applying these patterns in recommendation systems would result in poor generalization performance. Researchers also proposed to collect the sequential patterns from all the users and further factorize these patterns to obtain low dimensional representations, as shown in Fig 2. In recent, Rendle et al. [5] proposed to assemble all the customers’ sequential pattern matrices into a tensor and then apply factorization techniques over the tensor. However, since the objective of the factorization is to recover the sparse local patterns in essence, it would be very sensitive to the noise in individual data.
Fig 2. Global transition matrix gathered all customers’ sequential patterns.
Element with? are missing values.
To address the problem, we propose to assemble all the customers’ sequential pattern matrices into one matrix and apply factorization over this matrix. In this way, we can make the factorization focus on those globally salient patterns and more robust to the noise in the individual purchase data.
Specifically, we mine the CSPs from each customer’s transactional data, gather all the CSPs, and represent them into one global sequential pattern matrix W, with the element wa, b corresponding to the pattern <ia, ib, wab> where wab denote the overall weight of the pattern ia ⇒ ib in the transaction data. We then factorize W to learn the low dimensional representation of the products using the following objective function.
| (1) |
where denotes the k-dimension product representation matrix, and λ is the regularization coefficient.
Based on the learned low dimensional product representations, we can then provide personalized next-basket recommendation based on the customer’s latest transaction information. Specifically, we can inference the preference of customer um to product in in the next k-th transaction based on the products he bought in the latest transaction as follows:
| (2) |
where ql ∈ Q and qn ∈ Q indicate representations of product il and in respectively,operator “*” means dot product of two vectors. With the customer’s score on each product, we can sort the products and obtain the top-K products to be recommended to customers.
In next-basket prediction, factoring contiguous sequential patterns makes the assumption of “customers’ next purchase behavior is only related to what he has bought in the most recent transaction”. That is to say, the model only takes the local dependency between adjacent transactions into consideration and ignores the long dependency ignored. However, we can imagine there actually are some long dependency among customers’ transactions. For example, when a customer purchased a phone, he may probably buy a phone case after using one month and thus these two purchase behaviors are no longer in two consecutive transactions. To further capture the long dependency between transactions, we can involve the non-contiguous sequential pattern (NCSP for short) in our sequential pattern mining procedure. Here non-contiguous sequential pattern refers to the sequential pattern mined from non-consecutive transactions, as shown in Fig 3.
Fig 3. CSPs and NCSPs mined from transactions of a single user.
Solid lines stand for CSPs, while dotted lines represent NCSPs with the skip of 1 mined from transactions of user ui.
There are two ways to leverage the NCSPs and CSPs in learning of the low dimensional product representations for recommendation. One simple way is merging these two sets of the patterns into one set. That is to combine all the NCSPs and CSPs to form the new sequential pattern matrix W, and apply the matrix factorization to learn the product representations. With the learned representations, we then obtain personalized recommendation for each customer based all his historical transactions.
The other way is to treat different sequential patterns differently. Specifically, we propose a line combination model over the NCSPs and CSPs.
where Wi denotes the matrix of NCSPs with the skip of i transactions, and αi represents the influence of each NCSPs to users’ purchasing behaviors. Obviously, W0 corresponds to the matrix of CSPs.
With the learned the low dimensional product representations, we can also use the linear combination model to provide personalized recommendation based on each customer’s historical transactions.
Factorizing historical purchase data
As aforementioned, item-centric paradigm mainly leverages the correlation between products for recommendation. Another way is to exploit the correlation between users, so called user-centric paradigm. The idea behind is that “one is likely to buy the products favored by similar customers”, where collaborative filtering techniques have been widely applied. A typical way is to represent customers’ historical purchase behaviors as a customer-product matrix, and apply matrix factorization to learn the low-dimensional representations of both customers and products for recommendation.
Specifically, we first construct the customer-product matrix R based on customers’ historical transaction data, as shown in Fig 4:
Fig 4. Customer-product matrix mined from transactions.
Users’ purchase count to a certain product indicates customers’ general interest to it. The customer is more likely to buy a product in the next transaction the one he bought frequently before. Elements with? indicate unobserved purchase count.
We then factorize the customer-product matrix R to learn both the low dimensional representations of customers and products with the following objective function.
| (3) |
where um is represented as a vector pm in P, . We assume P indicating customers’ static preference to products. The task of this model is to learn customers’ general tastes to products.
Based on the learned representations of customers and products, we can provide personalized next-basket recommendation by calculating the customer’s preference on each product as follows:
| (4) |
where pm and qn are representations of user um and product in respectively. We then sort all products according Eq (4) and recommend top K products to each user.
Hybrid method
The two recommendation paradigms described above provide personalized next-basket prediction from different ways, i.e., one leverages the correlation between products and the other relies on the correlation between customers. The previous one can well capture the sequential behaviors of customers and the latter can better model customer’s general interests by ignoring transactions and assuming the exchangeability of purchased products. Therefore, a natural idea is to combine the two recommendation paradigms so that we can enjoy the powers of both paradigms and meanwhile complement each other to achieve better performance.
In our hybrid recommendation method, we propose to simultaneously factorize the product-product matrix and user-product matrix. In order the make the alignment between the learned representations for products, we require that they share the same low dimensional representations of products. Therefore, the objective function is described as follows:
| (5) |
Where α, β denotes the influence of users’ general tastes and temporal tastes to the next purchase behaviors. When β = 1, this model reduces to factoring customers’ sequential patterns, when we set α = 1, our model concerns only customers’ historical purchase behaviors.
With the learned low dimensional representations of customers and products, we can provide personalized next-basket recommendation with the same linear model using customers’ general interests as well as the latest transaction. Specifically, the preference of customer um to the product in can be inferenced as follows:
| (6) |
Where represents the purchase count of tn−1-th transaction for user m. After sorting pref for all products, we recommend top-K products to customers.
Optimization of the hybrid objective
When optimizing the hybrid objective, we find there is no closed-form solution for Eq (5). Here we introduce an alternative minimization algorithm to approximate the optimal results [24, 25]. The basic idea of this algorithm is to optimize the loss function with respect to one parameter, with all the other parameters fixed. The algorithm keeps iterating until convergence or the maximum of iterations. Specifically, we use P: |U| * k, Q: |I| * k stand for two low-rank matrix of customer set U and product set I.
Algorithm 1 hybrid factoring sequential pattern and historical data
Input: :Input R, W, m, n, k, α, β, λ, num
Output: P,Q
i = 0
P ← Rm*k, Q ← Rn*k
repeat
i ← i + 1;
;
;
until converge or t≥ num
return P, Q;
First we fix parameter Q, and calculate value of P to minimize Eq (5). Let Λ be the Lagrange multiplier, then we get Lagrange function as follows:
| (7) |
the gradient is:
| (8) |
With the given KKT complementary condition, we have the following equation:
| (9) |
Then we can obtain the gradient rule of P:
| (10) |
The gradient rule of Q can be obtained in the similar way.
| (11) |
The complete algorithm is shown in Algorithm 1.
Complexity analysis
The learning algorithm has two parts: (1) Complexity of updating matrix P according Eq (10),which is O(|U| ⋅ |I| ⋅ k). (2) Complexity of learning matrix Q according Eq (11), which is O((|U|+ |I|) ⋅ |I| ⋅ k). Thus when using Algorithm 1 to learn our model, the total complexity is O((|U|+ |I|) ⋅ |I| ⋅ k). In the proposed approach, k is very small, so it has a very good practice performance, and we have found that it converges fast after only a few iterations.
Experiment
Data description
We conducted empirical experiments to evaluate the effectiveness of our proposed method on next-basket recommendation. The experiments were conducted on three real-world transactional datasets, including two retailer datasets: BeiRen and Tafeng, and one e-commerce dataset TaoBao.
The BeiRen dataset is collected by a large retail department store in China, recording purchase of products during the period from 2012 to 2013. The Tafeng datasetis released by RecSys, which covers products from food, office supplies to furniture. The Taobao dataset is an online e-commerce dataset released by Taobao, which records the online transactions in terms of brands.
First, we conduct pre-processes on these transactional datasets. For BeiRen dataset and Tafeng dataset, we filtered the products that were bought less than 10 times. For the Taobao dataset, which is quite small, we filtered the products that were bought less than 3 times, to obtain sufficient data for training and prediction. Details of the three datasets are shown on Table (1).
Table 1. Data set statistics.
| id | name | # customers | # products | # transactions |
|---|---|---|---|---|
| 1 | BeiRen | 13736 | 5920 | 242894 |
| 2 | Tafeng | 9238 | 7973 | 37269 |
| 3 | Taobao | 191 | 292 | 1805 |
We split all the datasets into training and testing sets, where the last transaction of each customer is taken as the test set, and all the previous transactions are taken as the training set.
Baseline methods
To evaluate the recommendation performance of our model, we compare our model to several state-of-the-art next-basket recommendation methods, including the methods of item-centric and user-centric paradigms. We list all the baseline methods as follows:
TOP: The most popular Top-K products are recommended.
NMF: A state-of-the-art user-centric recommendation method based on collaborative filtering technique. Nonnegative Matrix Factorization is applied on customers’ historical purchase data. It can also be viewed as a sub-model of our method in which only the user-centric paradigm is adopted.
BPR: A generic method for learning models of personalized ranking based on pairs of orders (i.e. the user-specific order of items) [22].
FPMC: In FPMC, costumers’ sequential purchase behaviors are represented as a tensor and factorization is conducted to learn both low dimensional representations of customers and products [5].
FCSP: A sub-model of our method in which only the item-centric recommendation paradigm is adopted. FCSP factorizes contiguous sequential patterns to learn low dimensional representations of products for next-basket prediction.
FSP1: A sub-model of our method, which is an extension of FCSP by taking into account all the contiguous and non-contiguous sequential patterns. These two kinds of patterns are represented in a single matrix for learning and prediction.
FSP2: A sub-model of our method, which is an extension of the FCSP method by taking into account all the contiguous and non-contiguous sequential patterns. The two kinds of patterns are combined linearly for learning and prediction.
For each method, we run 10 times, and take the average as the final result. Both CFSH and datasets are available at https://github.com/sgc1993/cfsh.git.
Evaluation metric
We adopt F-measure as the evaluation measure for the Top-K Recommendation. F-measure is a weighted combination of precision and recall that produces scores of ranging from 0 to 1 and is accepted by many researchers as the metric for recommendation [5, 26–28]:
| (12) |
Comparison on different sequential models
First, we evaluated the effectiveness of different item-centric models over sequential patterns, including FCSP, FSP1, and FSP2. The purpose is to test whether it is beneficial to involve long dependency between transactions. For the two extension models which mine patterns from non-consecutive transactions, we only consider the skip-1 transaction patterns. The results of the item-centric models over the three datasets are shown in Fig 5. From the results we can see that all the models perform similarly, which indicates that non-consecutive sequential patterns bring limited help for recommendation.
Fig 5. Comparison of factoring sequential patterns.
We analyzed the reasons. On one hand, the number of unique non-consecutive sequential patterns are much less than that of consecutive sequential patterns. If we took the weight into consideration, the difference could be even larger. Therefore, the learned representations of products will not be affected when the non-consecutive patterns were taken into account, because the optimization may still concentrate on the consecutive patterns. On the other hand, the non-consecutive sequential patterns contain more noise than consecutive ones. If we treat the non-consecutive patterns equivalently with the consecutive patterns, the noise may hurt the performance. By using the linear combination models with decayed weights on long dependent patterns, we can obtain similar or slightly better results.
Based on the above results, we use the item-centric model with only consecutive sequential patterns as our sub-model in the following comparison.
Performance on next-basket prediction
In this section we compare our hybrid model to state-of-the-art methods in next-basket recommendation.
Fig 6 shows the results on Tafeng, Taobao and BeiRen respectively. We can see the Top popular method perform worst on all three datasets. It indicates that the next-basket recommendation problem is not trivial. By only using the popularity of products, we cannot generate good performance on next-basket prediction. NMF perform better than FCSP, we assume that users’ long-term interests may be more important in predicting users’ purchase behaviours than users’ sequential patterns. Meanwhile, BPR can produce much better results than NMF and FCSP methods. This result show that in item recommendation domain directly optimizing for the task of personally ranking can perform better than traditional methods of recovering customer-product matrix. Moreover, by considering both item-centric and user-centric information, FPMC can obtain slightly better results over BPR. It demonstrates the benefit of combing the two kinds of information in recommendation. However, due to the sparseness of the tensor in FPMC and the difficulty in factorizing the tensor, the improvement is very small.
Fig 6. Comparison of our model CFSH to TOP,FCSP,NMF,BPR, and FPMC methods on three datasets.
Comparing to other methods, our model outperforms all the state-of-the-art methods, with F1-score promoted at least 1.2% on Tafeng dataset, 1.1% on BeiRen dataset, 1.3% on TaoBao dataset respectively. The results demonstrate the effectiveness of combining two recommendation paradigms for next-basket prediction.
Influenced of α on hybrid model
In this section we study the impact of parameter alpha in our hybrid model. Parameter alpha is a co-efficient which tunes the balance between item-centric and user-centric recommendation paradigms in our work.
When alpha approximates 1, the model turns into a pure user-centric model, which means that the customer is more likely to purchase products those similar customers have bought. When alpha approximates 0, the model becomes an item-centric model, and predict customers’ next purchase only relying on what he/she has bought in the latest transaction.
Here we show the results over different α on Tafeng dataset. Similar results can also be obtained from the other two datasets. We vary the value of α from 0.1 to 0.9 with the dimension of 50. Fig 7 show the performance results, where the F1-score is influenced by α especially on the two ends. When α takes the medial values, i.e. from 0.3 to 0.8, the performance is quite stable for our model.
Fig 7. Performance of our model when adjusting alpha from 0.1 to 0.9 on Tafeng dataset with the dimension 50.
Conclusions
In this paper, we proposed a new hybrid recommendation method, namely CFSH, for next-basket prediction based on massive transactional data. The major purpose is to leverage the power of both item-centric and user-centric recommendation paradigms in capturing correlations between products and customers for better recommendation. This is achieved by factorizing customers’ sequential and historical purchase matrices simultaneously to learn customer and product representations better. Moreover, for the item-centric model, we propose to mine the sequential patterns from transactions in a global way to overcome the sparseness problem in sequential modeling. By conducting experiments on three real world purchase datasets, we demonstrated that our approach can produce significantly better prediction results than the state-of-the-art baseline methods.
In the future work, we would like to further analyze the correlations between products and costumers, so that we can better exploit this information and understand how these correlations affect each other. Moreover, we would like to integrate more context information into our model, e.g. time and location. Obviously, people’s shopping behavior may largely be affected by these factors. To present the next-basket recommendation at right time and right place would be very critical to task.
Acknowledgments
This research work was supported by the fundamental Research for the Central Universities, the National Natural Science Foundation of China (No.61802029,No.U1536121,No.41401376).
Data Availability
All files are available from the Beiren database on Github (https://github.com/sgc1993/cfsh.git).
Funding Statement
This research work was supported by the fundamental Research for the Central Universities, the National Natural Science Foundation of China (No. 61802029, No. U1536121, No. 41401376).
References
- 1. Cavique L. A scalable algorithm for the market basket analysis. Journal of Retailing and Consumer Services. 2007;. 10.1016/j.jretconser.2007.02.003 [DOI] [Google Scholar]
- 2.Keunho Choia GK Donghee Yoob. A hybrid online-product recommendation system: Combining implicit rating-based collaborative filtering and sequential pattern analysis. Electronic Commerce Research and Applications. 2012;.
- 3.Koenigstein N, Koren Y. Towards Scalable and Accurate Item-oriented Recommendations. In: Proceedings of the 7th ACM Conference on Recommender Systems. RecSys’13. New York, NY, USA: ACM; 2013. p. 419–422. Available from: http://doi.acm.org/10.1145/2507157.2507208.
- 4. Linden G, Smith B, York J. Amazon.Com Recommendations: Item-to-Item Collaborative Filtering. IEEE Internet Computing. 2003;7(1):76–80. 10.1109/MIC.2003.1167344 [DOI] [Google Scholar]
- 5.Rendle S, Freudenthaler C, Schmidt-Thieme L. Factorizing Personalized Markov Chains for Next-basket Recommendation. In: Proceedings of the 19th International Conference on World Wide Web. WWW’10. New York, NY, USA: ACM; 2010. p. 811–820. Available from: http://doi.acm.org/10.1145/1772690.1772773.
- 6. Mild Andreas R T. An improved collaborative filtering approach for predicting cross-category purchases based on binary market basket data. Journal of Retailing and Consumer Services. 2003;. [Google Scholar]
- 7.Christidis A K. Exploring Customer Preferences with Probabilistic Topics Models. In: European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases; 2010. p. 12–24.
- 8.Gatzioura M A; Sanchez Marre. A Case-based Recommendation Approach for Market Basket Data. Intelligent Systems, IEEE. 2014;.
- 9.Jong-Seok Lee e. Classification-based collaborative filtering using market basket data. Expert Systems with Applications. 2005;.
- 10.R Forsatia c M R Meybodib. Effective page recommendation algorithms based on distributed learning automata and weighted association rules. Expert Systems with Applications. 2010;.
- 11.Chand Chetna G A Thakkar Amit. Sequential Pattern Mining: Survey and Current Research Challenges. International Journal of Soft Computing and Engineering. 2012;. [Google Scholar]
- 12. Agrawal R, Imieliński T, Swami A. Mining Association Rules Between Sets of Items in Large Databases. SIGMOD Rec. 1993;22(2):207–216. 10.1145/170036.170072 [DOI] [Google Scholar]
- 13.Srikant R, Agrawal R. Mining Sequential Patterns: Generalizations and Performance Improvements. In: Proceedings of the 5th International Conference on Extending Database Technology: Advances in Database Technology. EDBT’96. London, UK, UK: Springer-Verlag; 1996. p. 3–17. Available from: http://dl.acm.org/citation.cfm?id=645337.650382.
- 14.Shani G, Brafman RI, Heckerman D. An MDP-based Recommender System. In: Proceedings of the Eighteenth Conference on Uncertainty in Artificial Intelligence. UAI’02. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc.; 2002. p. 453–460. Available from: http://dl.acm.org/citation.cfm?id=2073876.2073930.
- 15.Pengfei Wang YL Jiafeng Guo. Modeling Retail Transaction Data for Personalized Shopping Recommendation. In: 23rd International Conference on Information and Knowledge Management; 2014.
- 16.Chen S, Moore JL, Turnbull D, Joachims T. Playlist Prediction via Metric Embedding. In: Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. KDD’12. New York, NY, USA: ACM; 2012. p. 714–722. Available from: http://doi.acm.org/10.1145/2339530.2339643.
- 17.Mobasher B, Dai H, Luo T, Nakagawa M. Effective Personalization Based on Association Rule Discovery from Web Usage Data. In: Proceedings of the 3rd International Workshop on Web Information and Data Management. WIDM’01. New York, NY, USA: ACM; 2001. p. 9–15. Available from: http://doi.acm.org/10.1145/502932.502935.
- 18. Adomavicius G, Tuzhilin A. Toward the Next Generation of Recommender Systems: A Survey of the State-of-the-Art and Possible Extensions. IEEE Trans on Knowl and Data Eng. 2005;17(6):734–749. 10.1109/TKDE.2005.99 [DOI] [Google Scholar]
- 19. Su X, Khoshgoftaar TM. A Survey of Collaborative Filtering Techniques. Adv in Artif Intell. 2009;2009:4:2–4:2. [Google Scholar]
- 20.Koren Y. Factorization Meets the Neighborhood: A Multifaceted Collaborative Filtering Model. In: Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. KDD’08. New York, NY, USA: ACM; 2008. p. 426–434. Available from: http://doi.acm.org/10.1145/1401890.1401944.
- 21. Koren Y. Collaborative Filtering with Temporal Dynamics. Commun ACM. 2010;53(4):89–97. 10.1145/1721654.1721677 [DOI] [Google Scholar]
- 22.Rendle S, Freudenthaler C, Gantner Z, Schmidt-Thieme L. BPR: Bayesian Personalized Ranking from Implicit Feedback. In: Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence. UAI’09. Arlington, Virginia, United States: AUAI Press; 2009. p. 452–461. Available from: http://dl.acm.org/citation.cfm?id=1795114.1795167.
- 23.Chen J, Shankar S, Kelly A, Gningue S, Rajaravivarma R. A Two Stage Approach for Contiguous Sequential Pattern Mining. In: Proceedings of the 10th IEEE International Conference on Information Reuse & Integration. IRI’09. Piscataway, NJ, USA: IEEE Press; 2009. p. 382–387. Available from: http://dl.acm.org/citation.cfm?id=1689250.1689319.
- 24.Ding C, Li T, Peng W, Park H. Orthogonal Nonnegative Matrix T-factorizations for Clustering. In: Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. KDD’06. New York, NY, USA: ACM; 2006. p. 126–135. Available from: http://doi.acm.org/10.1145/1150402.1150420.
- 25.Lee DD, Seung HS. Algorithms for Non-negative Matrix Factorization. In: Leen TK, Dietterich TG, Tresp V, editors. Advances in Neural Information Processing Systems 13. MIT Press; 2001. p. 556–562. Available from: http://papers.nips.cc/paper/1861-algorithms-for-non-negative-matrix-factorization.pdf.
- 26. Godoy D, Amandi A. User Profiling in Personal Information Agents: A Survey. Knowl Eng Rev. 2005;20(4):329–361. 10.1017/S0269888906000397 [DOI] [Google Scholar]
- 27.Lacerda A, Ziviani N. Building User Profiles to Improve User Experience in Recommender Systems. In: Proceedings of the Sixth ACM International Conference on Web Search and Data Mining. WSDM’13. New York, NY, USA: ACM; 2013. p. 759–764. Available from: http://doi.acm.org/10.1145/2433396.2433492.
- 28. Lin W, Alvarez SA, Ruiz C. Efficient Adaptive-Support Association Rule Mining for Recommender Systems. Data Min Knowl Discov. 2002;6(1):83–105. 10.1023/A:1013284820704 [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
All files are available from the Beiren database on Github (https://github.com/sgc1993/cfsh.git).







