Quantity
Conjoint Analysis
Attribute levels (often) are assumed to only have additive contributions (main effects only, no interactions, cf. analysis of variance) L4-6
Fit and (internal) predictive validity of partworth estimates
Individuallevel partworth estimates: goodness of fit
- Pearson correlations r(stated preferences, predicted preferences) (= multiple correlation coefficient, R) will always be high
- – adjusted R’s (how well will partworth estimates predict preferences for full, non calibration profiles) won’t be high
- – partworth estimates have large confidence intervals (i.e. low reliability)
Solution
– Also elicit preferences for a number of full holdout profiles (not used for estimation)
– Compare predicted with stated preferences for holdouts
– Various holdout tasks with different measures of goodnessoffit
• may differ from calibration task
• rating Pearson correlation
• ranking rankorder correlation (e.g., Kendall’s tau)
• choices firstchoice hit rate
Note
- The importance of one attribute depends on the range of the attribute used to structure the stimuli, eg. price levels of 28,38,48 and levels of 28,32,36 (paper 4A)
- a relationship between the number of levels and the inferred importance of the attribute to the respondent: more levels, more importance
Determine Criteria
Select Attributes
- actionable
- important to individuals; actually affect consumer choice
Attribute Levels
- relatively large number of stimuli versus number of parameters
- a relationship between the number of levels and the inferred importance of the attribute to the respondent: more levels, more importance
Attribute Combination
- maximum 32 profiles
- In terms of predictive validity: is it really necessary/useful to estimate (two and higherorder) interaction effects, or is estimation of only main effects (additive model) good enough?
- interaction effects tend to have low predictive validity
- Fractional factorial designs also exist for unconfounded estimation of all main effects and some interaction effects
Aggregation: Choiceshare predictions for simulation profiles
Transform predicted preferences into predicted choice probabilities
- deterministic choice rule: winner gets all
• max utility (also firstchoice) rule probabilistic choice rules: winner gets most, but not all
• BTL rule; logit rule
when to use deterministic and when to use probabilistic? Involvment dependent
Transform predicted choice probabilities into predicted choice shares: sum across respondents
Aggregation of part worths: comparability across respondents
- One way to ensure comparability of results across respondents: L4-22
How many profiles?
- Lower bound 1: (total number of attribute levels) - (total number of attributes) + 1
– rationale: we need more data than to-be-estimated parameters
– lower bound for fractional factorial (42x23) (maineffects) design? - Lower bound 2: (number of levels of attribute with most levels) x (number of levels of attribute with one butmost levels)
Scale development: Reliability and validity
Procedures
Step 1: Definition of the construct
• Domain: what is and what is not included?
Step 2: Generate item pool
• Theory and good thinking
• Literature research
• Expert interviews
• Qualitative research in target group (e.g., focus-group discussions)
• Rules
– Redundancy
– Item editing
- S3 Select reduced set of items based on qualitative
judgements - S4 Pilot test: collect data from reasonable sample (e.g.,
n=200) - S5 Factor analysis: to investigate dimensionality
- S6 Reliability analysis
Three ways to estimate reliability
- Test-retest method
correlation between scores on same scale at two
different points in time
Problems: changes; remember; cannot measure first reaction;
- Alternative forms
correlation between scores on equivalent scales at two
different points in time
Problems:
-How to construct equivalent scales?
-problems of test-retest
3a. Internal consistency: Split-half method
correlation between scores on two subsets, each with
k/2 items (random split)
- Correlation between two halves gives reliability of scale of half the actual length
• Reliability (rk) of the whole scale can be obtained by Spearman-Brown formula (L8-20) - Problems
- Subsets need not be equivalent (will this give a too high, or a too low
estimate?) - Different splits may give different results
- Perhaps too easy to stay consistent
3b. Internal consistency: Cronbach’s a:
• Gives average of all possible split-half reliabilities
Problems
- Subsets need not be equivalent
- Perhaps too easy to stay consistent
- Considered to only give a lower bound for the actual reliability
– You can also get the standardized alpha (useful when items are measured on different, incommensurable scales)
Step 7. Purify scale
• If scale is not reliable enough
– compute a after deletion of item candidates for deletion
– carefully inspect content (!!!)
Step 8. New data collection
• Face/content validity
does the set of items look OK?
• Criterion validity
does the scale predict what it is supposed to predict?
– concurrent (predict something else now) versus predictive (predict something in the future) criterion validity
• Construct validity
does the scale really measure the construct?
Step 9
• Develop norms
Who belongs to the high-involvement group?
Who passes and who fails the exam?
Compositional Perceptual Mapping
CMP
CPM within process of product positioning
- identify (primary and secondary) competitors
- determine how competitors and your own product are perceived and evaluated
- determine competitors' and own position
- analyze consumer preferences (and underlying needs, motivations, habits)(segments)
- select attractive position, considering a particular segment
- try to achieve the position with all elements of the marketing mix
- monitor position over time
Steps in exploratory factor analysis
2) Number of factors to be retained?
- A priori ideas/theory
- Eigenvalue-larger-than-1
- Elbow in scree plot (eigenvalue)
- Cumulative % of variance accounted for (VAF) more than 60
- Not more than 50% of (absolute) residual correlations larger than 0.05
- Interpretability
3) Rotation
- orthogonal rotation (VARIMAX, QUARTIMAX) or
nonorthogonal rotation (OBLIMIN)
4) Interpretation
- label/interpret components/factors ... by looking at (pattern) loadings (can be interpreted as contributions of factors to attributes)
• statistically significant loadings
• practically significant loadings (0.30, 0.40, 0.50) - Compute mean factor score for each product, on each
dimension - Make a perceptual map with the products in it, on the basis
of their mean factor scores
Rationale behind CPM
In daily life, consumers distinguish between products on
more primary needs/strategic benefits (halo effect)
Ratings on more specific attributes/benefits are related
to positions on strategic-benefit dimensions (these dimensions are eliciting)
- Therefore, exploratory factor analysis will reveal these
strategic-benefit dimensions
Use for marketing management
- Compare product perceptions in terms of
general dimensions - Predict effect of changes in product’s attribute
scores on strategic dimensions
– each dimension is weighted sum of attributes - Identify potential opportunities
Issues in CPM
Data collection
- different scales can be used
- ask respondents to only evaluate those products that
they are familiar with - warning: consumers tend to give inflated ratings to
product concepts (NPD) - Ynte’s lecture: difference between unrestricted versus
restricted (context) evaluations
Extended data matrix
- (L10 22-24)Correlation between attribute determined by two
sources of covariation: consumers and products
-the dominance of consumer heterogeneity can be prevented by preprocessing (centering or standardizing)
-Perceptions are generally assumed to be homogenous across
consumers (in contrast to preferences) No modeling of
(inter)individual differences.
Covariances vs correlations
Correlation matrix: each attribute has equal impact on the results
Covariance matrix: Attributes with more variance have a larger impact on the results
- when attributes are not in the same scale, it's better to analyze with correlation matrix
- the correlation matrix is calculated based on standardized data
Compositional VS Decompositional
- Advantages of CPM
+strong face validity
+data easy to collect
+easy to interpret
+one can choose actionable attributes - Disadvantages of CPM
– depends on pre-specified set of attributes
– dimensions are linear combinations of attributes
– attributes that are important for preference/choice, but
independent of other attributes tend to be overlooked
– arbitrary scaling of dimensions
Cluster
Hierarchical clustering procedures
- agglomerative
- divisive
Cluster distance
- single linkage
- complete linkage
- average linkage
- ward's method
- centroid method
How many homogeneous groups?
- A priori idea
- Absolute value of agglomeration coefficient
- Screetest criterion based on agglomeration coefficient
- Measures of tradeoff between withincluster homogeneity and number of clusters
-Variance between clusters versus variance within clusters, choose the greatest one
– Pseudo F-value does not follow theoretical F-distribution - Interpretability/usefulness of clusters
Disadvantage of hierarchical cluster analysis
- joins and divisions are irreversible → successive solutions may depend on minor details the data
- with equal dissimilarities (ties) in the data, successive solutions even depend on (arbitrary) order of respondents in the dataset!
Nonhierarchical cluster analysis
- start out with K random/selected cluster means (centroids)
in each of a series of iterative steps (optimizing):
we stop with the iterative steps, when there is no improvement in the cluster solution anymore
- criterion: sum of squared deviances from cluster mean
Disadvantages of nonhierarchical methods
- different seeds tend to give different solutions
- random seeds tend to give inferior solutions
- computationally more intensive
Validity
Internal validity
- convergent validity: compare results from different (reasonable) dissimilarity measures
- convergent validity: compare results from different clustering procedures
- cross-validation: compare results from one half of the sample to results from another half of the sample)
- simulation using random data: do we get more homogenous clusters with our actual data?
External validity
- compare results with expected clustering (theoretical classification ⇒ taxonomy)
- check expected relations (= ?? validity)
- does the clustering yield useful market segments? (criteria?)
1) Is the data suited for exploratory factor analysis?
- Number of observations
- Interval-scaled scores (?); Independent observations
(?); Normally distributed scores (?) - Sufficient association between attributes/variables
(multicollinearity)
• Sufficient number of large correlations
• Low partial correlations
• Bartlett’s test of sphericity
• KMO and MSA