Data‑driven segmentation often starts with clustering algorithms. See if you can spot the key concepts behind modern cluster analysis.
The default distance measure for k‑means on standardised numeric data is ______ distance.
Cosine
Euclidean
Manhattan
Hamming
In hierarchical clustering, the method that merges clusters based on the smallest average pairwise distance is called ______.
complete linkage
average linkage
single linkage
Ward’s method
The elbow method helps analysts decide the optimal ______.
distance metric
number of clusters (k)
scaling factor
principal components
Standardising variables before clustering prevents ______ variables from dominating the distance calculation.
categorical
irrelevant
binary
high‑scale
K‑prototypes handles mixed data by combining numeric distance with ______ distance for categorical variables.
Hamming
Mahalanobis
Jaccard
Cosine
The silhouette coefficient jointly measures cluster cohesion and ______.
kurtosis
skewness
separation
sparsity
In agglomerative clustering, merging continues until ______ cluster(s) remain.
one
a predefined silhouette score
k+1
zero
Probabilistic centre initialisation that improves k‑means consistency is known as ______.
fuzzy c‑means
random swap
Hartigan optimisation
k‑means++
DBSCAN labels observations in sparse regions as ______.
linkages
centroids
medoids
noise
Ward’s method in hierarchical clustering aims to minimise the total ______ within clusters at each step.
dimensionality
entropy
distance metric
variance
Starter
Keep exploring the fundamentals.
Solid
Great job—refine the finer points and you’ll lead the pack.
Expert!
You’ve mastered this segmentation concept—apply it boldly.