Notes
Notes
- of how to construct a customer satisfaction survey and of how to plan and execute an experiment
(off-line or on-line)
- of statistical methods to answer questions related to acquisition, retention and development of
customers
Skills
Competences
- To select, apply and reflect on a variety of statistical approaches to customer related business
problems
- To communicate the results of findings to management in an appropriate way.
1. Course Organization
Customer analytics – a use of a customer databases to enhance marketing productivity through more
effective acquisition, retention and development of relationships with customers.
Marketing productivity - Evaluate decisions made on economic terms (i.e. does it have an influence
on profits for a company?). does the plan you have for implementation of marketing activity, does it
have a better outcome in economic terms than an alternative.
Fundamental ways to improve profits when we talk about customers is (1) to be better at acquiring
customers, retaining them and/or developing relationship than the competition. (2) sell more to
existing customers, sell to new customers.
Reading material
https://blackboard.au.dk/bbcswebdav/pid-2955791-dt-content-rid-10792015_1/courses/BB-Cou-Hold-
41544/Customer_Analytics_460202U040_update%2003.03.pdf
Survey Data
In-store and online sales data (transactions)
Loyalty program data
Web browsing data
Experiments
Social Networks data
Text (Word of Mouth)
Survey data
The information that we can collect from the customers (categories of databases in the company):
Inactive customers -
An important distinction in transaction data (that will be used in future analysis) is the distinction
between contractual and non-contractual settings. Example:
Non contractual
- What is the probability that a customer is still alive?
- How many transactions can we expect in the future from a customer
Contractual
Retention:
- What is the expected duration of a relationship?
Churn:
- How can we identify customers who are most likely to churn?
Overall
- What is the value of the whole customer base?
How can we sell more to the existing customer and hence increase the value of our
customers?
Market basket data can be represented in a binary format, where each row corresponds to a transaction
and each column corresponds to an item
This simplistic representation of real market basket data ignores certain important aspects of the data
such as the quantity of items sold or the price paid
User based collaborative filtering. Bayesian networks might also be used for this.
Web browsing data – clickstream data
customer activity
Sequential data:
Markov assumption: each state is dependent only on the previous one. Dependency given by a
conditional probability
Concept is trivial
RFM analysis
This analysis is a way to obtain different segments. It is made of Recency, Frequency and Monetary
value, that comes from transactional data. It represents the classical way to segment customers. It is
the second most often way to analyse customer data by database marketers. It has a link to the value
of the customer.
How does it come about?
Analyzing a sample
Rank all customers according to recency and divide into 5 quintiles. For customers in recency quintile
1: rank by frequency and divide into 5 quintiles. Repeat step 2 for the other recency quintiles.
For each of the 25 recency-frequency groups rank by monetary value and divide into 5 quintiles.
e.g. we have 40 000 customers. Dividing them into groups multiple times, leaves us with 125 groups.
This is called the sequential approach (the traditional way) and is used most often.
The advantage is that you get the same amount
of customers in each group.
The interpretation of the frequencies will not be
the same across the different groups. That is the
disadvantage of this approach. Same with
monetary, the interpretation will be different for
different groups.
Concept of LIFT
If we look at a random 10% of the potential customers, we expect to get 10% likely responders. Can
we select 10% of the potential customers and get more than 10% of likely responders? If so, we
realize “lift”. This is a key goal in data mining.
The cumulative lift chart
1. x – axis gives population percentile
2. y – axis gives lift
3. the top 10% of the scorers are 3 times
more likely to respond than a random
10% would be
RFM characteristics
Simple and transparent, but
- “The net present value of the profits linked to a specific customer once the customer has been
acquired after subtracting incremental costs associated with marketing, selling, production and
servicing over the customer’s lifetime” (HANS JØRN JUHL, lecture 2)
Individual CLV
Contractual and non-contractual setting
Simplifying assumptions:
Margins are constant over time and all cash flows take place at the beginning (or end) of the
period
Estimated by (total gross margin in a period)/(no. of customers)
Retention rate is constant over time and estimated based upon a cohorte
Discount rate is constant
Length of projection is infinite
Only customers who are doing business with the firm
Lost for good model
An important Distinction
Margin multiple:
2.5/1.15=1.17
You can calculate how much the customer value increases based on these tables and these formulas.
3. CLV continuing
If we can:
It is assumed that number of customers are taking S curve from the graphs. The s curve can be
expressed:
The challenge is that we cannot estimate any model of S(t) directly from the data. Migration models
are used to calculate non-contractual settings.
Calculation of P:
Contractual setting
Continuous time
Discrete time
Survival rate and retention rate
“Future retention rates for a specific cohort of customers in a discrete time contractual setting should
take customer heterogeneity into account”
There are no variables that varies through time. any analyst wishing to do so would need to forecast.
We assume that this basic model refers also to future periods (therefore this model can be used for
forecasting). Also, it is assumed that the consumer will follow geometric distribution and
heterogeneity will follow the basic distribution.
(for visualization check R code)
You can evaluate through graphs (r code) or check results:
6) Estimate the parameters (of mixing distribution) by fitting the aggregate distribution to the
observed data
7) Use the model to solve the marketing decision problem/provide the required information
Individual level – Timing
A Bernoulli trial is a probabilistic experiment in which there are two possible outcomes, “success” (or
1) and “failure” (or 0), where ϴ is the probability of success.
Repeated Bernoulli trials lead to the geometric and binomial distributions.
The (shifted) geometric distribution can be used to model discrete-time timing data.
B(α,β) is the beta function, which can be expressed in terms of gamma functions:
Continuous – many transactions during a period. The most typical situation in practice.
Stylized pattern of purchasing over time by a cohort of customers in a non-contractual setting and
latent attrition
Customers disappear over time. it happens at a different point in time. they might leave forever. We
assume that customers buy at a steady rate (same expected number of transactions per period) and
then they will leave under unobserved time. these are the two assumptions that we make.
The biggest problem is that the Poisson distribution is memoryless. In both models for the continuous
non-contractual settings. Regardless of this, the model has been proven to work.
- Γ(·) is the gamma function – because of mathematical convenience we use this distribution.
(flexible)
- r is the “shape” parameter and α is the “scale” parameter
- The gamma distribution is a flexible (unimodal) distribution … and is mathematically
convenient.
If we combine the two mixing distribution and the assumption of the individual level then we get the
aggregate distribution.
Counting negative binomial distribution
[The first milestone paper in the study of buyer behavior was “The Pattern of Consumer Purchases”
(Ehrenberg 1959) which showed that the Negative Binomial Distribution (NBD) provided a very good
description of the numbers of purchases of individual brands of consumer goods in a single time
period. From ”Ehrenberg’s contribution to social and marketing research” ]
Calculating the probability and the expected number of customers having zero purchases
Summary of probability models and integrated models
BG/NBD
Necessary data:
We need to know
a) Recency (time of last transaction (t(x)) )
b) Frequency ( (x) number od repeat transactions in a time interval (0,T) )
c) Total time between first purchase and the end of the observation window (T)
We don’t need to know
a) Information on when each of the x transactions occurred
Basic input and output from the integrated probabilistic models for a continuous non-contractual
setting:
PARETO/NBD
Application area
Non-contractual setting where the opportunities for transactions are continuous as customers
can make as many transactions as they want and any time in each period
Attrition is unobserved. The company does not know if a silent period means that the
customer has left the company
Main questions to answer with the model
As long as a user is alive, the number of transactions follows a Poisson process with
transaction rate λ. Thus, if we left X(t) be a random variable which denotes the number the
probability of observing x repeat transactions in time interval (0,t) is:
The assumption also implies that the time between repeat transactions follows an exponential
distribution with transaction rate λ:
Although the number of repeat transaction for each user follow a Poisson process, the
heterogeneity in the users is captured in the heterogeneity in the transaction rate λ across
users, which follows a gamma distribution with the probability density function
Where r is the shape parameter and α is the scale parameter. From the gamma distribution we can find
the mean transaction rate as E(λ)= r / α and the variance of the transaction rate as V(λ)= r / α2
About timing. For modeling lifetime, the model assumes that each user has an unobserved
“lifetime” of length τ, after which the user is inactive. This unobserved lifetime has an
exponential distribution with a dropout rate μ:
Where s is the shape parameter and β is the scale parameter. From the gamma distribution we can find
the mean dropout rate as E(μ)= s / β and the variance of V(μ)= s / β2
Finally, we assume that the transaction rate λ and the dropout rate μ vary independently
across users. This assumption is debatable. We would e.g. expect that a user with a high mean
number of repeat transaction is a week, a large transaction rate λ, also has a longer expected
lifetime, a lower dropout rate μ, as the user is most likely an active and engaged customer.
Summary
RFM CLV
CLV contractual
Naïve approach (average). The expression depends upon when the company receives their
payments
Heterogeneity. The retention rate and hence the residual lifetime value depends upon the
number of renewals a customer has made. S-BG model
CLV non-contractual
Discrete Continuous
Migration models: Pareto/NBD or BG/NBD
Contagion perspectives
Satisfied customers buying adjacent offerings and further motivating others to purchase
through word of-mouth
Affective-state perspectives
Satisfied customers developing positive affinities and enhanced product and brand loyalties
leading to future purchases
Market-free perspectives
Earned reputation for satisfying customers creating barriers to entry for non-incumbents
leading to market share gains and additional scale effects
we can see that the affect is exponential, if you can increase the loyalty expressed by retention rate,
then value of the customer will increase in an exponential way.
The link between satisfaction and loyalty
It is not a linear
relationship.
More realistic is
the s shaped
curved.
The link from satisfaction to the economic effects 0 different performance measure
Definition
Satisfaction –
Is the customer’s fulfillment response. It is a judgement that a product or service feature, or the
product or service itself, provided (or is providing) a pleasurable level of consumption related to
fulfillment, including levels of under- or over fulfillment (Oliver, 1996)
Satisfaction as the post consumption consumer judgement of whether the good or service provided a
pleasurable level of overall usage-related fulfillment (Oliver, 2020)
Satisfaction is the post-consumption evaluation of a product and service (Mittal)
Micro or Macro:
The expectancy-disconfirmation model of satisfaction
There are two ways that we can measure expectations and performance and compare them two each
other. The two ways:
Direct and inferred approach
Can be
Will be
Must be
Should or ought to be
It is very important that when you ask your questions you make a
standard expectation definition otherwise the answers will be a mess
Servqual
- Model that is developed to compare expectations and perceptions. The idea is to subtract
expectation from perception across all levels (gaps).
Assurance Empathy
Ability to convey trust and confidence Ability to be approachable, caring,
Give a feeling that customers’ best interest understanding and relating with customer
is in your heart needs
Example: being polite and showing respect Example: being a good listener
for customer
Tangibles
Physical facilities and facilitating goods
Example: cleanliness
Purpose
To develop a scale for measuring perceived service quality based on a disconfirmation approach
P-E (P – perception (what a specific service provider offers), E – expectation (what the customer
would expect from ‘excellent’ service) )
Example – context effect
A telephone company
In general P-E lead to negative scores. And that’s why inferred approach is not preferable.
There are a lot of companies that occupy with this. The main purpose of these national index
satisfaction models is to measure customer satisfaction at (1)company level (2) section level (3)
national level
Demands of these models:
Swedish CS model
This model is problematic because it assumes some causal relationships. (e.g. that CS leads to CL.
These should be measured at different points in time).
The next problem is with customer complains is that probably not a lot of customers complained and
therefore there are too few obs. For complains.
A third problem is with customer expectations. All other things are determined by the model (there
are no arrows to it) meaning if you want to improve your model you need to improve customer
expectations outside the model.
This model is called structural equation model.
ACSI model
EPSI model
Customer loyalty –
A deeply held commitment to re-bui or re-patronize a preferred product or service consistently in the
future.
In 2003, Reichheld published an article in HBR, in which he claims his Net Promoter Score, is the
only number you need to grow, and the only number you need to manage customer loyalty.
A lot of companies use this score (Microsoft and others)
Reichheld’s claim: By substituting a single question for the complex black box of typical customer
satisfaction survey, companies can actually put consumer survey results to use and focus employees
on the task of stimulating growth (2003)
The single question answered by all respondents is “How likely is that you would recommend
Company X to a colleague or a friend? 0 – 10”
The calculation of NPS
Calculate the percentage of customers who respond with nine or ten (promoters) and the percentage
who respond with zero through six (detractors).
Subtract the percentage of detractors from the percentage of promoters to arrive at your net-promoter
score.
(check NPS in R)
Evaluation criteria
The main argument for measuring satisfaction and loyalty is that they are leading indicators
for business results
It is important to distinguish between transaction specific and cumulative satisfaction
It is important to distinguish between attitudinal loyalty and behavioral loyalty (retention)
The national index models are often used models as a starting point for a customer
satisfaction survey
The main advantage of national index models is that they provide possibilities for
benchmarking and measuring of developments over time
The NPS is a very popular indicator in business because it is simple and easy to communicate
The academic literature does not include many studies supporting NPS as an indicator for
revenue growth.
8. Survey
- A tool for understanding the workings of the human mind and the dynamics of social
interaction. Survey is one of the sources or primary data.
Type of data we can analyse
1. Customer identification data
2. Demographic data
3. Psychographic or lifestyle data (survey data)
4. Transaction data
5. Marketing action data
6. Other type of data (e.g. financial, competitive data)
Representativeness
A sample will be nearly perfectly representative of a population if a probability sampling methods is
used and if the response rate is 100%
Representatives increases monotonically with increasing response rate. NOT TRUE. Surveys with
low response rates can be more accurate than surveys with much higher response rates (Visser et al.,
1996)
e.g. telephone surveys would forecast election outcomes more accurately by accepting lower response
rates, rather than aggressively pursuing high response rates (Greenwald et al.)
Who are the respondents?
People who agree to be interviewed are likely to believe that:
Pretesting
Conclusion 2
The three pretesting methods differ in terms of the kinds of problems they detect and in the reliability
with which they detect these problems:
Questionnaire design
The quality of data collected by each method seemed equivalent in some studies
Recent studies found that the reliability and validity of open-ended questions exceeded that of
closed ended questions
Closed questions (disadv.)
Some responds feel difficulty when expressing their thoughts and feelings
Respondent’s laziness and lack of time
Conclusion 4
Open-ended questions seem to be more viable research tools in some studies while closed-ended
questions are more viable in other studies. No clear black and white answer.
When there are only numeric labels for the categories between the verbally labeled endpoints,
respondents probably assume the categories are equidistant
Providing verbal labels for all the categories, both endpoints and intermediate ones, produce
more reliable measurement (not al all, slightly, somewhat, pretty, very, extremely)
“very” which commonly appears as an anchor is not intense enough fo most applications
Conclusion 5
Reliability and validity can be significantly improved if all points on the scale are labeled with words,
because they clarify the meanings of the scale points.
Numbers should reinforce the meanings of the words rather than communicate conflicting meanings
In words:
1. Respondents must interpret the question
1.1. bring the sounds of the words into their working memories
1.2. break the words down into groups, each one representing a single concept, settling on the
2. meaning of each
2.1. build the meaning of the entire question by establishing the relations among the concepts
2.2. discard the original words of the question and retain their interpretations as they begin to
3. formulate an answer.
4. They must search their memories for relevant information
5. Then integrate that information into a single judgment (if more than one consideration is
recalled).
6. Finally, they must translate the judgment into a response by selecting one of the alternative
offered.
Types of respondents:
Optimizing – great deal of cognitive work required to generate an optimal answer. All four steps are
executed DILLIGENTLY
Weak Satisficing – Instead of generating the most accurate answers, respondents settle for merely
satisfactory ones. All four steps executed, but each one less diligently
Strong Satisficing – respondents may interpret each question superficially and select what they believe
will be a reasonable answer to the interview and researcher.
3 Conditions that foster satisficing
a) Task difficulty
b) Respondent’s inability
c) Respondent’s demotivation to optimize
Response order effects
Primacy effects (response choices presented early were most likely to be selected) (Satisfiers
are heavily affected)
Recency effects (response choices presented last were more likely to be selected)
OR
No order effects
Facts
Weak satisficing seems likely to produce primacy effects under conditions of visual
presentation
When response alternatives are presented orally, memory factors promote recency effects
more than primacy effects
Items presented at the beginning and end of a list are more likely to be recalled after the
question is read, particularly if the list is long
A primacy effect was stronger for people with more limited cognitive skills
More response order effects in questions involving more words and words that were difficult
to comprehend
Conclusion 6
Fact
Researchers have typically preferred to use rating questions rather than ranking questions.
But, a number of studies indicate that rankings yield higher-quality data than ratings.
Rankings are more reliable and manifest greater validity.
But what are the disadvantages of raking data? how close is 1st from 2nd in the ranking?
Example
- For example, when asked to rate the importance of a series of values (e.g. equality, freedom,
and happiness) on a scale from extremely important to not at all important, a satisficing
respondent can easily say they are all very important
- = Non-differentiation
- Nondifferentiation is more common among less educated respondents and is more prevalent
toward the end of a questionnaire
- Nondifferentiation is particularly pronounced among respondents low in verbal ability, for
whom fatigue is presumably most taxing
DO NOT KNOW option should increase the quality of data obtained by a questionnaire.
Respondents would be discouraged from offering meaningless opinions.
People sometimes offer such responses because
- they feel ambivalent about the issue
- do not understand the meaning of a question or the answer choices;
- think that they must know a lot about a topic to legitimately express an opinion
- reduced motivation at the end of survey
Krosnick(1999) found no difference in data quality between filtered and standard versions in
any of their experiments.
Number of categories
The standard advice has been to use 5 to 9 categories (Miller 1956, Cox 1980),
Magnitude scaling and feeling thermometers, offer a very large number of numerical options,
but respondents usually choose answers that are multiples of 5, 10, or 25(at the low end of the
continuum) and 50, 100, or 1000 (at the higher end), so that the number of categories used is
less than one might expect
Conclusion 8
Ratings are mostly used in practice because they are easier to analyze. Rankings have some
advantages, by forcing the respondents to think deeper and decide.
Regarding using “Do not know” option, some studies found no difference in data quality between
filtered and standard versions. But no black and white answer until more studies available.
Question type
1. Questions an attitudes and values
2. Questions on events and behaviors
Questions on events and behaviors
Do members of the target population have encoded the information?
e.g. Researchers may want to ask parents about their children’s immunization history, but some
parents will be unaware of what, if any, immunizations their children have received (Lee et al. 1999).
Major errors: omissions and intrusions when memory is altered
Overreporting errors
They can be reduced by using a panel survey with bounded interviews
In an initial bounding interview, respondents are asked to report events; the list of events reported at
time 1 is then consulted during the time 2 interview. If an item is reported at time 2, the interviewer
verifies that it is a new item and not a duplicate of the item reported at time 1
Response dimensions
Occurrence (simple yes-no items, closed frequency items, and open frequency items)
Absolute frequency (recommended open questions to avoid the answer is affected by the
response categories)
Relative frequency (“very often”, “seldom” …vague quantifiers)
Regularity or periodicity
Date of occurrence
Issues of relevance
Respondents may be annoyed at being asked a question that does not apply to them
Use filter questions to avoid this
Choose between:
Threatening behaviors
Strategies to reduce voting claims:
1. Ask the respondent to remember details about the vote
2. Presented multiple response categories for “didn’t vote” instead of the usual single one, and
phrased the “did vote” category in definite terms
e.g. “I thought about voting this time but didn’t”
3. Self-administered modes of administration produce both higher reports of socially
undesirable behaviors and lower reports of socially desirable ones
Questions about subjective phenomena
Focus groups are often used during questionnaire development to identify appropriate names and
relevant dimensions.
The probability that a confidence interval will include the population parameter
If the sample represents 10% of the population or more, apply the finite population correction to n
Example:
Conduct a survey to estimate the monthly amount invested in savings schemes so that the estimate
will be within ± 5.00€. What should be the sample size (n)?
The std. dev. Of the population (sigma) may be known from secondary studies, if not, it can be
estimated by conducting a pilot study; or researcher’s judgement.
If the population standard deviation is unknown, then, after the sample is drawn, we can estimate the
confidence interval for mean by employing sample standard deviation.
Multiple parameters
In most business research projects, however, several characteristics are of interest. If all variables are
equally important, the most conservative approach is either to take a sufficiently big sample (> 1000-
2000) or to evaluate the required sample statistically for each parameter and to select the largest value
for the required sample size.
Fact
n practice, in order to achieve the calculated sample size, a much greater number of potential
respondents have to be contacted. Why?
- Incidence rate of the phenomenon investigated
- Response rate
Conclusion 9