Forecasting the Effects of In-Store Marketing on Conversion Rates for Online Shops

As webstores usually face the issue of low conversion rates, finding ways to effectively increase them is of special interest to researchers and practitioners alike. However, to the best of our knowledge, no one has yet empirically investigated the usefulness of various in-webstore marketing tools like coupons or different types of product recommendations. By analysing clickstream data for a shoe and a bed online store, we are contributing to closing this gap. In particular, we use our present data to build more general hypotheses on how such purchasing incentives might function and on how they could be used in practice.


Introduction
Throughout the last several years, internet advertising revenue has grown constantly to about $59.6bn in the U.S. alone (see Figure 1) with the pace notably picking up since 2010. Parallel to this development, according to a recent release from the U.S. Department of Commerce, the retail e-commerce sales share of total retail sales has nearly doubled from about 4% in 2010 to 8.1% in 2016 (cf. [1]).  As such trends are usually accompanied by increasing competition, two main challenges arise for online retailers which we will discuss in more detail below. Due to the general direction of the present paper, we shall especially focus on the second one. For a more general overview of marketing in computer-mediated environments (which is beyond the scope of our paper), we refer to [2].
Firstly, customers need to find their way to a shop's website which can be facilitated by, e.g., search engine optimization (SEO), contextual search engine advertising (SEA) or graphical banner advertising (see, e.g., [3][4][5] or the very recent studies of [6,7]).
Secondly, once potential customers have reached a webstore, they need to be "encouraged" to actually convert from visitor to buyer. In the literature, there can be found various discussions on how to increase this conversion rate. For once, especially in the earlier days of e-commerce, trust played an important role and could be increased by investing in the general design of an online shop to project the ability of a certain quality in terms of products and services (cf. [8][9][10][11][12][13]). However, the detailed shape, style and content of a website or a shop can also boost the conversion rate if it fits to the characteristics of the targeted customer group (cf. [14]). This can go as far as having a webstore that automatically "morphs" to match the suspected needs of each specific visitor (cf. [15]). In this context, based on their generated clickstream data (i.e., how they move through a webpage), customers can be characterized either by statistical methods or, more dynamically, by certain learning models (cf. [16]) in order to specifically target them with in-store marketing.
However, due to the rare academic availability of such clickstream data containing in-webstore marketing campaigns, the relevant literature here is still scarce. However, by partnering up with a digital consulting start-up that specializes in increasing the effectiveness of webshops by boosting their conversion rates, we were able to obtain such a unique data set from two stores in the online retail space-a shoe and a bed shop. This provides us with the possibility to investigate two very different product types as shoes and beds differ significantly in various characteristics (price, duration of use, etc.).
Therefore, in the present paper, we are aiming to further close the above-mentioned research gap by examining the impact of different types of (in-store) purchasing incentives on each shop's conversion rate. Additionally, we cast a further glance on the forecasting power of general clickstream data extending earlier in-store studies (cf. [17][18][19][20][21]). In more detail, we measure the conversion rate effects of advertising in the form of coupons, product sliders (i.e., "Other products that might interest you:") or information about the shopping behaviour of peers (i.e., "Other customers bought these products:"). Additionally, we account for individual-specific exogenous effects captured by clickstream data like how the customer reached the online shop or how many pages he viewed.
Our work adds to current research in two ways: to the best of our knowledge, this is the first time that the effects of in-webstore marketing approaches like coupons have been empirically measured via clickstream data. Additionally, we suggest certain data-driven hypotheses built from our data on how such purchasing incentives could be used best in practice.
The structure of our article is as follows: after a thorough literature review carried out in the next section ("Relevant Literature"), we discuss the present data set and investigate the potential effects of purchasing incentives from a more descriptive standpoint in "Data". To address any bias in statistical tests due to potentially non-random selection of control and target groups in our data, we invoke classical generalized models of the logit-type ("Statistical Model Setup"). We present our results in "Empirical Findings" and extensively discuss the impact of certain purchasing incentives allowing possible interactions with other exogenous variables. Finally, in "Conclusions and Implications", we summarize our approach and give recommendations on how to generally identify customer groups which can be purposely targeted to boost an online shop's conversion rate.
All statistical analyses in the present paper have been carried out with the software package R-3.5.1.

Relevant Literature
A first step towards selling products and services online is obviously taken by attracting users to a webshop which can, e.g., either be done via SEO, contextual SEA (cf., e.g., [22][23][24]) or (graphical) banner ads. With the availability of various types of clickstream data, i.e., information about how users move through the web, a large part of online advertising research has focussed on such contextual and graphical ads, although it might be difficult to deduce direct effects: in this context, Xu et al. [25] find that banner ads have a rather low impact on direct conversion rates but might encourage future site visits via different advertising formats. Their study is, however, limited in the sense that their data does not contain visits resulting from non-paid search engine results or direct URL type-ins.
An important insight was that, due to some kind of information overflow, the use of display ads is accompanied by the risk of avoidance (cf. [3]). In order to counteract this effect, such ads can be customized or personalized to catch the attention of users (cf. [4,6,7,[26][27][28] or the recent study of [14]). In addition, one can increase the obtrusiveness of ads which has been shown to positively impact visitors' purchase intentions even though a combination of both approaches might be harmful due to privacy concerns (cf. [5,28,29]).
In general, behavioural targeting, personalized advertising or so-called website morphing (cf. [15]) has become a deeply investigated research stream.
However, all of these studies focus somewhat on increasing conversion rates by attracting users to a webstore. Therefore, in a second step, one might also ask how in-webstore advertising might be helpful to convert users which have already found their way to a shop by, e.g., offering time-limited price discounts or luring visitors to view certain products in more detail. The effect of such coupons and general in-store marketing campaigns has already been extensively studied in real-world offline stores over the last decades: As it is well-known, one advantage of coupons lies in their simplicity as a marketing tool (cf. [30]). At first, to asses their effectiveness, several authors considered just the pure redemption rate, i.e., the number of coupons which are being used compared to the amount which has been distributed (see [31]). In addition, Reibstein and Traver [31] developed a model to determine various factors which influence the redemption rate, namely (amongst other things) the method of distribution, the face value of coupons and the discount offered by the coupon.
However, later on, Bawa and Shoemaker [32] pointed out that redemption rates as a single measure of success are misleading as high rates are not equal to profitability (cf. [33]). They suggested to rather focus on incremental sales and tried to identify target households which are most lucrative.
Dhar and Hoch [34] concluded that the effect of in-store coupons for grocery chains might be larger than the one for off-the-shelf price discounts while [35] pointed out that the mere exposure to customized coupons might have a positive effect on sales and profitability while such coupons work better when they are unexpected.
In a more recent study, Reichhart et al. [36] compared digital e-mail and mobile text message coupons by conducting a field experiment with exposed users from an opt-in database. They could find that text message delivery caused better conversion rates even though the response rates proofed to be lower. In a similar context, Danaher et al. [37] conducted a two-year trial in a shopping mall with similar text message coupons and found that traditional offline coupon characteristics, e.g., face value, are again the main influencing factor for coupon effectiveness. In addition, the authors found that location and time of delivery influence redemption, e.g., customers tend to redeem a coupon more often if the corresponding store is closer. Additionally, they suggest that a shorter expiration length has a positive effect as they signal urgency to customers.
Next to coupons, more general in-store marketing campaigns are in the focus of current research, as well. For example, Chandon et al. [38] conducted an eye-tracking experiment and concluded that, even though the number of product facings has a positive effect on conversion rate for certain types of shop visitors, gaining attention is itself not always sufficient to increase sales. Inman et al. [39] carried out several intercept interviews and identified characteristics of shop visitors that tend to make unplanned purchases.
More recently, Hui et al. [40,41] investigated visitors of a grocery store by video-tracking and identified product categories that are more likely to elicit unplanned purchases than others. Additionally, they suggest that coupons which tempt customers to deviate from their determined in-store travel path are nearly twice as effective as those who do not. Furthermore, Zhang et al. [42] concluded that interactive social influences can positively affect sales as well, especially when shoppers show certain behavioral cues.
Having all the above mentioned in-store marketing research available, it might be an obvious step to straightforwardly transfer these results to webstores. However, as Bucklin et al. [43] pointed out, there are major differences between the choice making processes of online customers and classical real-life shop visitors that can influence the impact of customized promotions (see, e.g., [44]). As a consequence, it is not clear if obtained results on coupons and in-store marketing do hold in a digital world. In fact, the relevant literature does not yet provide a final answer.
As the conversion rate of a typical online shop rarely exceeds 5% (cf. [17]), a large part of early studies focussed on identifying visitors who are more likely to buy than others. In this context, based on clickstream data for an online-shop selling cars, Sismeiro and Bucklin [19] decomposed the purchasing process into several steps describing each one with a probit model. In particular, they find that repeated visits do per se not go hand in hand with a higher conversion rate-even though an earlier study suggested that browsing behaviour might change in such situations (see [45]). In more detail, Moe and Fader [17,18] found evidence that even though customers who visit an retail online-store more frequently tend to have a higher conversion rate, this is mainly due to the subgroup of shoppers whose frequency behaviour actually increased over time.
In terms of directly increasing conversion rates, Wilson [46] suggested potential positive impacts of easier and more transparent purchase/checkout processes as well as free-shipping promotions in a B2B context even though their study is only based on several hundred shop visitors.
More generally speaking, an understanding of customer decision rules in online stores (cf. [20]) could be used for individual promotional targeting. In this context, Schellong et al. [21] (extending an earlier study of [47]) try to classify the in-store shopping behaviour of a fashion retailer's visitors by clustering browsing and search patterns.
Still, while several of the above studies already suggested that targeted in-store marketing based on clickstream data might be a fruitful approach, detailed empirical evidence on various such techniques has been scare. One of the few exceptions is the study of [48] who investigate the effect of showing visitors personalized product recommendations. In the following, by partnering up with a digital consulting start-up, we aim to further close this research gap by studying the effects of various in-store marketing campaigns, especially the impact of time-limited discounts in the form of coupons and certain types of product recommendations.

Data
In this section, we want to briefly describe our data set and provide a first descriptive overview. In particular, we received a month of customer data from two different mid-sized online shops from our research partner, a start-up in the e-commerce space. These shops shall henceforth be denoted by their specific business sector, i.e., "shoe shop" and "bed shop". The chosen month is February 2015 for which we have data from 1 February 2015 to 28 February 2015 for the shoe and from 6 February to 15 February for the bed shop. The reasoning behind this time period is that no particularly large public holiday (which might bias our results) falls within this month. Additionally, the data is outdated enough such that one should not be able to draw any conclusion on the identity of the individual shops. This is especially important from a compliance perspective as, due to our agreement with our research partner, we are not allowed to share precise information about the specific online stores.
In particular, we were not provided with details about the stores' websites (like their design, style of product presentation, etc.).
Similarly to, e.g., the study of [7], our data is available on a cookie level which can have several drawbacks in terms of user identifiability (cf. [7,49] for a more detailed discussion) but is still industry standard.
For each user/cookie, we have information about the particular day and time of his visit, how he reached the website and for how long he stayed, how many pages he viewed, if and when he put any items in the shopping cart, the aggregated value in EUR of these, if he actually converted and whether he was shown any purchasing incentives like a 'Coupon' (cash value between 5-10 EUR), a 'Slider' ("Other products that might interest you:") or a '(Group) Motivation' ad ("Other customers bought these products:"). Table 1 summarizes these available variables and their scales. After removing incomplete data points and transmission errors, we are left with a total of 266,773 visits for the shoe and 98,266 for the bed shop. Of these numbers, a total of 188,430 (70.63%) and 76,728 (78.08%) visitors are unique from a cookie level standpoint, i.e., the rest arises due to various users visiting the stores on multiple occasions during February 2015. From a statistical point of view, as a consequence, our individual data points are probably not perfectly independent but might (depending on the user behaviour) correlate. Given the comparably small amount of data available for our study, it is hard to adequately account for such dependence effects: to visualize this issue, let us assume that the impression of an initial shop visit stays with the average user for about seven days. Then, even when we fully accounted for multiple visits, about 25% of our data might be diluted by such effects without us even knowing. Therefore, for the sake of simplicity, we shall from now on model each cookie as being a unique visitor. However, we want to stress that this is probably the biggest drawback of our study and needs to be addressed in future research.
As a consequence of the above and in line with comparable studies (e.g., [7]), we will not aim to generalize the results of our study to different shops and time periods but rather use our present data for a more exploratory approach trying to understand how purchase incentives might affect users and build data-driven hypotheses for a more general framework. Table 2 provides a global summary of our data situation and shows that about 30% of all visits for the shoe and about 50% for the bed shop were shown some purchasing incentives. Regarding the nature of such influences, Table 3 shows that around 73% for the first and 96% for the second shop were of a unique type while the rest is mostly a combination of two. Given the small amount of observations for multiple incentives, we shall neglect any potential interactions effects between these for the analysis to come. Considering the sales of the individual shops and splitting all users into a non-influenced and an incentivised group, we can clearly see that the conversion rate roughly doubles for both stores once at least one purchasing incentive has been shown. A simple two-sided binomial test confirms that these results are significant on all sensible levels, see Table 4, even though the actual effects differ substantially as can be seen by Figure 2. However, it is hard to draw robust conclusions based on these tests as one needs to ensure that both groups, non-influenced and incentivised visitors, consist of users having the same overall characteristics. Otherwise, it would be quite easy to manipulate the effects of purchasing incentives by showing these mostly to such visitor groups which already have an ex-ante higher conversion rate. Even though our research partner ensured us that the selection of incentivised customers was "mostly random", we want to address this issue in more detail.
As a first step, for both shops, we split all users into non-influenced and incentivised groups. Now, Figures 3 and 4 present the marginal distributions of all collected variables for each store in total and broken down into the "control" and "test" customer categories. Even though these plots look mostly fine, a few differences spring into view: for example, for the shoe shop the frequencies of at least one item in the shopping basket are higher for influenced users. However, this makes sense as purchasing incentives are usually not directly shown from the start but only after the visitor stays some time on the shop's webpage. As a larger group of customers leaves our online stores directly after a few seconds, the difference in the frequencies can (partly) be explained. However, more worrisome is, e.g., the mismatch for the bed shop and the variable 'Channel'. Apparently, the share of visitors which typed the shops address directly into the browser is underrepresented in the incentivised group. From our perspective, this is a clear violation of having a completely random allocation of incentives. Therefore, we do not feel comfortable relying on the results of binomial tests as carried out in Table 4. However, as we will discuss in the next section, our chosen statistical approach of applying generalized linear models allows us to tackle and mitigate such "non-randomness". Additionally, we will provide a minimal working example to show that these setups can efficiently detect if a conversion rate increase for incentivised users is only due to a unclean or tactical selection of the target group or due to the purchasing incentive itself.

Statistical Model Setup
This section shall provide an overview of our chosen statistical approach. As we are aiming to econometrically model the impact of several exogenous (mostly categorical) variables on each shop's endogenous conversion rate, we chose a classical logistic regression setup, i.e., a generalized linear model with logit link. In the following, we shall briefly review the necessary properties of such setups. For a more detailed background, we refer an interested reader to, e.g., [50]. Now, in particular, for the ith visitor, the (conditional) probability of a purchase is assumed to be determined via where . . , X m i ) contains the realizations of the exogenous m variables (and 1 to include an intercept) for the ith visitor and β = (β 0 , β 1 , β 2 , . . . , β m ) represents the model parameters.
After obtaining the ML estimates, we can assess each model's fit by its deviance given by with C i := 1, ith visitor converts, 0, else, which basically measures the difference between the likelihood of the chosen setup and a fully saturated model. Therefore, the smaller the deviance, the better the in-sample fit of the present model. The deviance of a model with just a constant is called null deviance. Additionally, having two nested setups, the difference between both deviances is χ 2 -distributed under the null hypothesis of the smaller model being the correct one. To further asses the validity of our models, we shall make use of the Akaike information criterion (AIC) as well.
In a first step, we want to test whether the larger models have an advantage over the smallest one implying that at least one influence-type has a significant impact on the conversion rate. Additionally, as from a practical perspective, a shop's owner would be interested in identifying certain visitors for which incentives are most effective, we shall furthermore consider the significance of each individual incentive including potential interaction effects. Regression coefficients thus can be interpreted via their implied ceteris paribus changes on the conversion odds. Generally, when it comes to interaction effects, additional care has to be taken (cf. [51]). However, as Model 3 only includes interactions between a dummy and one other variable, we are on the safe side here.
Finally, we can investigate each setup's in-sample forecasting capabilities by assuming that π i > 0.5 implies a predicted conversion for the ith visitor. Considering a 2 × 2-cross table, we can check how much the forecasts agree with the actual data. Similarly, we shall check the out-of-sample prediction power by randomly separating our data into a training and a test set. Then, in a first step, the parameters β are estimated on the training data and the actual model forecasts are then evaluated for the test set. As indicated in the previous section, it remains to illustrate how our chosen logistic regression approach mitigates and tackles the issue of potential "non-randomness" in our data. For this purpose, let us generate a test data set (called 'Test Data 1') with 300,000 observations and (only) two exogenous variables present: Channel (TypeIn or SEO) and Coupon (Yes or No) (see Table 5). Additionally, we assume that the coupons have no influence while SEO-visitors convert in 2.0% and TypeIn-visitors in 5.0% of all cases. Now, if we show a coupon to 10% of all SEO-and 60% of the TypeIn-visitors (distributed such that the independence assumption between coupon and sale holds) a simple binomial test would indicate a significant conversion rate difference between influenced and non-influenced customer groups, see Table 6. However, invoking all three introduced types of logistic regression setups (see Table 7), it becomes clear that this seemingly significant effect is just an artificial one due to the obviously inappropriate distribution of the coupon variable.  Similarly, we generate a second data set ('Test Data 2') equal to the first one. However, the effect of coupons within the channel-groups is made such that, for SEO-clients, the conversion rate is approximately 11.2% higher through the purchase incentive while, for TypeIn-visitors, the effect is even 29.4%. A simple binomial test as carried out in Table 6 can not differentiate both data sets while our logistic regressions neatly catches even the interaction effect between both exogenous variables as illustrated by Table 7: Model 2 implies that on average a coupon increases the purchasing odds by exp(+0.227) − 1 = 25.5% while Model 3 indicates exp(+0.271) − 1 = 31.1% higher odds for TypeIn-visitors and an exp(+0.271 − 0.162) − 1 = 11.5% increase for SEO-customers.

Empirical Findings
In this section, we shall discuss the obtained results from Model 1-3 for each shop which are individually presented in Tables 8 and 9.

Results on Shoes
Starting with our shoe shop data and the raw benchmark model, we see that users who actually type the store's web address manually into their browser are significantly more likely to make a purchase than visitors coming from any other source. In fact, all other things being equal, the conversion odds decrease between 66-68% for all other groups. Furthermore, these odds are reduced by around 2.5% for each additional page view while they decrease by around 36% for any additional minute on the shops website. Obviously, this relationship is truly rather nonlinear and we should not rely too much on the actual numbers even though they confirm the reasonable conjecture that with time, the conversion probability of users goes down. Additionally, as we can see by considering the coefficient of the variable 'BasketTime', this decay is countered if visitors actually put some items into their digital shopping cart.
Interestingly, for our shoe shop (and partly in contrast to the bed store), users with more items in the basket and a higher total basket value are also more likely to buy, which might, however, be explained by assumingly higher return rates. Sadly, these were not available to confirm our hypothesis.    Table 9. Logistic regression models for our bed shop data. Null deviance equals 28,096. p-values equal or below 0.05/0.01/0.001 are indicated by * / * * / * * * .

Model 1 Model 2 Model 3
Intercept  Considering Model 2 which now accounts for some users being shown certain purchasing incentives, we can see that the effect of 'Slider' and 'Motivation' are both highly significant but surprisingly negative. In fact, these advertising tools seem to decrease the conversion odds of an individual user by around 50-53%. Coupons on the other side seem to have no statistically significant effect. While the former observation could be at least heuristically explained by some kind of information overload (cf. [52,53] for a good overview and discussion on this general issue), the later seems to be counterintuitive for us.
Finally, Model 3 takes potential interaction effects into account as well. It turns out that it performs not only best in terms of deviance (as χ 2 tests can reject the smaller models on all sensible significance levels) and AIC but also offers an explanation to the counterintuitive results on purchasing incentives of Model 2. In fact, coupons per se seem to have a positive effect of about 62% on the conversion odds but significantly alter the model's parameters for some of the other variables. For example, the negative impact of an increase in 'PageDuration' is not as pronounced compared to the non-coupon group which could be neatly explained by intuition again: if at some point a user is presented with a coupon, his initial decision process is altered and he needs additional time to process the new information which is not bad in that case.
However, the coefficient of 'PageViews' changes rather dramatically: while for Model 1 the conversion odds decreased by around 2.5% for each page view, the addition of a coupon tripled this effect to about 7.2%. This finding stands in opposition to results on offline coupons in physical stores which have shown to positively influence sales when they encourage customers to deviate from their planned walking trajectory (cf. [40,41]).
Regarding the other purchasing incentives, the picture is not as bright. Interestingly, a slider seems to have a significant advantage if shown to visitors not coming via 'TypeIn' and seems to negatively influence users who already have some items in their basket. We explain these observations as follows: users who access our store via 'TypeIn' probably already have some products in mind and might get distracted by seeing other items. The same argument could be used to explain the interaction effect with 'BasketNumber'. For 'Motivation', similar things can be observed, which is not massively surprising as these two advertising tools are somewhat akin to each other.
Finally, as an additional measure to assess the goodness-of-fit for our setups, we compare their in-and out-of-sample forecasting power. The in-sample results are depicted in Table 10 and further confirm that Model 3 seems to perform best. Cohen's kappa increases significantly to about 0.834 and the specificity is 0.816 compared to 0.758 and 0.775 for the smaller setups while the false-positive rate stays almost constant. We want to stress that, due to the generally low conversion rates in our data, these measures need to be interpreted with care. Probably most important for practical perspectives is to identify buyers and therefore an increasing specificity is of special concern to us.
To assess the out-of-sample forecasting capability of our models, we randomly split our data into a 60% training and a 40% testing set. All parameters are estimated on the training data while the forecasting performance is judged using the testing set. We repeated this procedure 1000 times and the cross tables in Table 11 present the means for each model. On these, we again calculated Cohen's kappa, the specificity and the false-positive rate. In fact, there seems to be nearly no difference compared to Table 10, which we take as further confirmation of Model 3 and its implications for the impact of the considered purchasing incentives.

Results on Beds
Again, we shall start with the raw benchmark Model 1. In particular, we notice that most coefficients look quite similar to our shoe shop. However, a few differences spring into view. For once, 'AggBasketValue' shows a highly significant negative sign of −0.001 which can be explained by the fact that beds are usually more expensive than shoes. As a consequence, there are generally higher outliers for 'AggBasketValue' in this data set, which skews the linear regression coefficient into the negative territory (as a visitor cannot convert if he has no items in his basket, there is trivially some initial positive effect as well). More interesting from our perspective, however, is the highly significant negative effect of −0.929 of 'Hour (0-7)', which says that the odds of seeing a purchase between these hours is about 60.5% lower than seeing one in the afternoon hours. Such a time-dependence (which we were not able to detect for the shoe shop) makes sense somewhat as purchasing a bed is probably mostly done after a more complex decision process which usually does not take place in the night or in the very early morning. A similar feature can be found for 'Hour (19-23)' as well, even though it is with 12.5% lower odds not as pronounced.
The same argument could also explain why the coefficient of 'PageDuration' is not as negative as for our shoe shop-again, users seem to need more time to think more about a bed purchase. The fact that more 'PageViews' have a stronger negative effect on the purchasing odds does not necessarily contradict these considerations as visitors who consider more products can be thought of being just at the beginning of their (longer) decision making process.
Additionally, shoppers coming via 'SEA' show 104% times higher purchasing odds than 'TypeIn' visitors. We try to explain this feature by the following logic: when it comes to buying shoes, more determined users tend to directly go to "their" online shop of choice to search for suitable products-while, when looking for beds, shoppers either do not know of suitable online stores or rather search for a certain brand already.
When it comes to Model 2, most relations from the raw benchmark model seem to be still there. In contrast to the shoe store, showing users a coupon seems to increase the purchasing odds by about 159% while 'Motivation' seems to decrease these by around 45%. However, 'Slider' seems to have no significant impact at all.
Finally, when considering the interaction setup in Model 3, we find that 'Weekend' increases the purchasing odds significantly by around 20%, probably with the same explanation why people do usually not buy beds in the night or early morning. For 'Coupon' (which now increases the purchasing odds by approximately 126%), we can (similar to the shoe store) find a positive interaction with 'PageDuration', while 'Slider' only seems to have significant positive effects on users who visit the shop via a specific link. In this context, it might be surprising that the coefficient of 'Channel(SEA)' is globally negative now. However, this is counteracted by the interaction effect with 'Motivation', i.e., SEA-shoppers are still rather receptive when it comes to recommendations. A similar feature, although not as strong, can be found for 'Channel(SEO)' as well.
Generally, χ 2 tests based on the deviances and AIC again strongly favour Model 3 over the smaller ones. However, when we assess the models' forecasting performances, the results are rather disappointing showing no significant increase in Cohen's kappa for neither the in-sample nor the out-of-sample forecasts (cf . Tables 12 and 13). Additionally, the actual levels of Cohen's kappa and the specificity are discouraging in their own right. As a consequence, although we are able to find certain significant effects of customer specific variables and purchasing incentives, there seem to be other important factors missing to successfully identify future buyers. However, as buying a bed (in comparison to shopping for shoes) probably entails a more complex decision-making process in the background, this is not very surprising.

Conclusions
We analyzed clickstream data containing various in-webstore marketing tools (i.e., purchasing incentives) for a shoe and a bed online shop. In particular, we find that, while identifying future buyers via a logistic regression setup works quite well for the shoe shop, the forecasting performance for the bed store is comparably weak.
When it comes to quantifying the impact of purchasing incentives on the conversion odds, we hypothesize the following general relationships and implications for in-webstore marketing tools: There seems to be a general positive and significant impact of coupons. Considering interaction effects with various clickstream variables, it seems to make sense to especially target users who have watched a lower number of items for a longer time (i.e., whose 'PageViews' is smaller and whose 'PageDuration' is larger compared to others).
However, for 'Slider' and 'Motivation', the situation is more complex: these tools seem to work only for certain user groups and have a globally negative impact. In fact, it seems to be best to target only visitors not coming via typing the shops' web address directly into their browser and who do not yet have a larger amount of items in their basket.
From a research standpoint, we want to stress again that the main purpose of this study was to derive data-driven hypotheses for a more general framework. Therefore, the above needs to be tested and verified in future research.
However, based on the above, we can already state the following brief managerial implications: if a deeper analysis of your visitors' behaviour is not (yet) available, concentrate on showing coupons to customers whose attention is focused on a small number of products for a longer time. Purchasing incentives like 'Slider' and 'Motivation' should only be shown to visitors who reached the shop's website not via 'TypeIn' and who do not yet have a product in their shopping basket. In the meantime, it might make sense to carry out a thorough analysis of your customers by either charging your own statisticians with this task or by hiring external help.