top of page
Search

How to Use AI (Machine Learning) in UX Studies Reliably

  • Writer: Mohsen Rafiei
    Mohsen Rafiei
  • 3 days ago
  • 4 min read

ree

Artificial Intelligence (specifically Machine Learning) has opened new possibilities for understanding user behavior. Yet, its value in UX research depends entirely on how and when it is applied. As researchers seeking to uncover why people behave the way they do, we often rely on summarizing performance through simple aggregates: average task completion time, average satisfaction scores, or average conversion rates. However, the so-called “average user” is largely a convenient fiction. Behind each mean value lies a spectrum of distinct behaviors and experiences that are too easily collapsed into a single, misleading number.


An average task time of 90 seconds, for instance, conceals a critical reality: half of users may complete the task in 30 seconds, while others struggle for 150. Averages compress meaningful variability into a single number, obscuring the moments where the user experience breaks down, or excels. Those extreme cases, or outliers, often contain the most valuable insights about usability failures, cognitive overload, or emergent user strategies.

However, detecting meaningful outliers in large-scale datasets is far from trivial. With thousands of user sessions, the challenge is distinguishing signal (meaningful deviations that reveal a UX issue or opportunity) from noise (data errors or random anomalies).


Outliers as Noise and Signal

Outliers in UX data arise for two main reasons. Some reflect measurement noise: system errors, bots, or invalid responses that distort distributions. Others represent genuine behavioral signals. A user who repeatedly clicks “Submit” on a broken form is not noise, they are an implicit usability test case. Similarly, an exceptionally fast participant may reveal an unrecognized efficiency in task flow.

The researcher’s responsibility is not merely to remove anomalies but to classify them appropriately. Automated outlier detection can assist this process, yet interpretation remains a human task grounded in contextual understanding of user behavior, system design, and study protocols.


From Descriptive Statistics to Machine Learning

Traditional statistical rules, such as z-scores or interquartile ranges, identify extreme values relative to a single variable. These methods are valuable for initial screening but inherently limited: they assume linear relationships, treat each dimension independently, and rely on arbitrary thresholds.

Machine Learning methods, in contrast, detect complex, multi-dimensional deviations. They can account for how multiple behavioral features, time, click count, error rate, eye-tracking measures, interact to define “normal” versus “unexpected” behavior. This allows researchers to surface nuanced patterns that conventional descriptive approaches would overlook.


Common ML Techniques for Outlier Detection in UX Data

Isolation Forest. This algorithm identifies outliers by recursively partitioning the dataset. Data points that are isolated after few partitions are likely anomalous. It is computationally efficient and well-suited to large telemetry datasets or A/B test logs.

k-Nearest Neighbors (k-NN). Here, each observation is compared to its nearest neighbors. Data points distant from their peers are flagged as anomalies. It is intuitive and particularly useful for detecting survey “speeders” or out-of-pattern responses.

Local Outlier Factor (LOF). This model evaluates not only distance but also local density. It recognizes that what is unusual in one context may be typical in another, ideal for identifying subpopulations within broader patterns (for example, a dissatisfied respondent within a generally positive cluster).

Autoencoders. These neural networks learn to reconstruct “normal” data and measure how poorly they reconstruct new inputs. High reconstruction errors reveal anomalies. Autoencoders excel with sequential or physiological data such as navigation traces, eye-tracking, or EEG streams, identifying moments where behavior diverges sharply from baseline.

Each technique has practical constraints. Isolation Forest and LOF require tuning hyperparameters that affect sensitivity. k-NN can become computationally expensive for very large samples. Autoencoders demand careful data normalization and sufficient “normal” data for training. Thus, model choice should reflect dataset size, dimensionality, and research goals.


Implementation Workflow

  1. Data Preparation. Machine Learning models assume structured, numerical input. Each row should represent a user or session, and each column a behavioral or perceptual metric. Outlier detection is only as reliable as the preprocessing, missing data, skewed distributions, or unstandardized scales can bias results.

  2. Model Selection and Execution. Tools like scikit-learn in Python make outlier detection accessible in minimal code. However, researchers must validate results through exploratory data visualization and, where possible, cross-checks with known benchmarks or manual review.

  3. Interpretation and Action. ML-based outlier scores identify candidates for further investigation, not definitive truths. Researchers should triangulate flagged cases with qualitative materials (e.g., session replays, user comments) to determine whether anomalies reflect noise, usability failures, or innovative user strategies.


Limitations and Considerations

While ML methods can efficiently surface non-obvious patterns, they are not substitutes for interpretive expertise. Models can overfit small datasets, misclassify valid behaviors as anomalies, or obscure transparency through algorithmic complexity. In UX research, false positives carry interpretive cost, investigating an algorithmic artifact wastes time and may misdirect design priorities. Moreover, ethical concerns arise when algorithmic profiling of user behavior is performed without clear consent or adequate anonymization. Therefore, ML-driven anomaly detection should complement (not replace) traditional analysis and researcher judgment. Combining quantitative screening with qualitative inquiry yields a richer, more reliable understanding of user experience variability.


The fixation on averages has long constrained UX insight. By integrating anomaly detection methods from machine learning, researchers can move beyond the “typical user” narrative and uncover the nuanced, sometimes messy stories that define real human interaction with technology. Yet, methodological rigor remains essential. Outlier detection must be guided by context, validated through triangulation, and interpreted within human-centered frameworks. The most valuable insights lie not in algorithmic novelty, but in connecting these advanced tools back to the fundamental question of UX research: why do users behave the way they do?

 
 
 
  • LinkedIn

©2020 by Mohsen Rafiei.

bottom of page