EMERGING DIALOGUES IN ASSESSMENT

Exploratory Analysis on Interactive Learning Analytics Dashboards: Pitfalls that Cause Unintentional Harm   

 

March 25, 2026

  • Priya Harindranathan, PhD
    Director Curriculum, Evaluation, and Accreditation, Office of Medical Education, Paul L. Foster School of Medicine, Texas Tech University Health Sciences Center El Paso

Abstract

Learning Analytics (LA) offers powerful tools for data-driven insights in education. The development of Learning Analytics Dashboards (LADs) provides intuitive interfaces to visualize, interpret, and act upon complex educational data. Despite their strengths and benefits in data exploration, exploratory drilldowns are associated with shortcomings including bias, fallacies, and statistical misinterpretations such as Simpson’s paradox and the multiple comparisons problem. This paper discusses how cognitive biases, choice overload, and manual drill-down errors can lead to flawed inferences and misguided educational interventions. It highlights how user-driven, curiosity-based exploration without statistical rigor increases the likelihood of false insights and resource misallocation. To mitigate these risks, the paper recommends integrating validation protocols, user training, interdisciplinary oversight, and evidence-based guidelines for dashboard use. Future directions include designing dashboards with intelligent “nudges” to guide users toward meaningful analyses and conducting multi-institutional studies to evaluate the impact of LADs on decision-making and institutional improvement.



Introduction

Learning Analytics Dashboards (LADs) have become important tools for visualizing and acting on institutional data. LADs that rely on predictive analytics are criticized for decision-making without human supervision. Exploratory data-analytics, an alternative to this approach, offers user engagement and cost-effectiveness by utilizing visualization and descriptive statistics to identify data patterns (Behrens, 1997).

As the volume, type, and frequency of data inflow on LADs increase, extracting meaningful information becomes challenging. Interactive views with multiple hierarchies to slice and dice data help visualize trends. Making the different attributes of data available to users via progressive addition of filters is known as ‘drill-down’ operation, which allows meaningful zoom-in to a granular level to perform comparative analyses on different subpopulations. In higher education institutes, both data enthusiasts and experts rely on visualizations on platforms such as Tableau to uncover complex data relationships (Zhao et al., 2017). Data enthusiasts who may not have formal backgrounds in data analysis could engage in curiosity-driven drill-down explorations that are not always data-driven to interpret results and make judgments (Shabaninejad et al., 2020). They could inadvertently switch between exploratory and confirmatory analysis, leading to systemic bias. Bias and fallacies in LADs arise when human interpretation of interactive visualizations leads to erroneous conclusions. Cognitive biases may cause users to overemphasize familiar patterns while overlooking contradictory evidence (Kahneman, 2011). Visualization choices, such as selective data filtering can lead to interpretive fallacies (Tversky & Kahneman, 1974).

Thus, while exploratory LADs are often framed as safer alternatives to predictive analytics, their interactive flexibility introduces statistical and cognitive vulnerabilities. The purpose of this work is to demonstrate how exploratory analysis of visualizations on LADs are associated with pitfalls that cause unintentional harm and can lead to systematically flawed institutional decisions about course design, teaching or learning.

Simpson’s paradox

Simpson’s paradox is a subset of a general class of phenomena known as “mix effects”, where the results of comparing the outcome between two groups can be confounded by a previously unexamined variable. Such misleading inferences fall under the broader category known as ecological fallacy in social sciences or omitted variable bias/confounding covariates in statistics.
A scenario where Simpson's paradox might occur on a LAD that tracks the performance of students on standardized tests across two departments over two years is examined below.

Aggregated Data

  • Year 1 Dept A: Overall pass rate = 80%
  • Year 1 Dept B: Overall pass rate = 76%
  • Year 2 Dept A: Overall pass rate = 76.6%
  • Year 2 Dept B: Overall pass rate = 71.6%

Based on aggregated metrics, administrators might conclude that Dept A is more effective than Dept B in preparing students. However, suppose the student populations differ substantially by gender distribution and performance patterns (see Table 1).

Table 1: Disaggregated Data by Gender
Year Department Male Pass Rate Female Pass Rate % Male % Female
 1  Dept A  85%  60%  80%  20%
 1  Dept B  90%  70%  30% 70% 
 2  Dept A  82%  55%  80%  20%
 2  Dept B  87%  65%  30%  70%


For year 1:

  • Among male students, Dept B (90%) outperforms Dept A (85%).
  • Among female students, Dept B (70%) outperforms Dept A (60%).

For year 2:

  • Among male students, Dept B (87%) outperforms Dept A (82%).
  • Among female students, Dept B (65%) outperforms Dept A (55%).

Thus, for both years, while aggregated data suggest Dept A performs better, disaggregated data reveal Dept B performs better within both subgroups (male and female students). Yet, since Dept A enrolls a larger proportion of higher-performing male students, its overall aggregated pass rate appears higher.

If administrators rely only on overall pass rates, they may conclude that since Dept A is performing satisfactorily no targeted intervention is necessary. However, disaggregated data reveal two critical issues:

  1. Systemic Decline Over Time: Both departments are experiencing declining performance across subgroups, possibly signaling a broader instructional concern.
  2. Persistent Gender Gap: Males consistently outperform females, suggesting an equity issue.

This highlights the importance of careful consideration of confounding variable (demographics) to avoid the risk of making high-stakes decisions based on aggregated metrics, resulting in misallocation of resources or faulty accountability decisions.

Multiple Comparisons Problem 

Users compare visualizations to mental images of what they are interested in, say, a trend or an unusual pattern. As more visualizations are examined (more comparisons made), the probability of discovering flawed insights increases merely by chance. This problem, well-known in statistics as the multiple comparisons problem (MCP), is overlooked in visual analysis.

Listed below is a scenario where MCP might occur on a LAD that tracks the average test scores of students in four departments.

  • Dept. B: Average test score = 82
  • Dept. C: Average test score = 88
  • Dept. D: Average test score = 80
  • Dept. E: Average test score = 86

If users conduct pairwise comparisons, they may find that Dept. C appears to outperform Dept. B and that Dept. E appears to outperform Dept. D, while other comparisons show no significant differences. However, when multiple comparisons are performed without adjusting for inflated Type I error rates, some statistically significant findings may simply reflect random variation. After applying appropriate corrections (e.g., Bonferroni adjustment or False Discovery Rate control), these apparent differences may no longer remain statistically significant.

This example highlights a critical risk on LADs that allow users to filter by department, demographic group, course level, semester or instructor. Each filter interaction implicitly creates a new hypothesis test, often without the user recognizing it. While the flexibility supports exploratory analysis, it dramatically increases the number of possible implicit comparisons and practitioners may inadvertently:

  • Initiate unnecessary interventions based on chance variation
  • Damage reputation of spuriously identified “underperforming” departments

Challenges with Manual Drill-down in Visual Analytics

Figure 1: Drill-down fallacy 

To summarize, the three challenges associated with manual drill-downs on LADs are:

  1. Choice overload: Choice overload refers to the cognitive burden users experience when presented with an excessive number of visualizations or interaction options. In manual drill-down, to uncover interesting insights about subpopulations, users have to determine how to filter the most effective attributes. As the number of attributes and values increases, the possible number of drill-down actions increases exponentially, leaving users disillusioned. When dashboards display too many indicators without clear prioritization, users struggle to interpret data effectively, leading to decision fatigue and reduced engagement (Iyengar & Lepper, 2000). Such overload can obscure actionable insights and diminish educational value of LADs (Verbert et al., 2014).
  2. Lack of insightful results: A large number of possible drill-down criteria can lead to uninformed choice of drill-down paths and results that are not insightful. Users who are diffident about the drill-down options could pick the filters by trial-and- error approach leaving several filter options unused or underutilized. Ignoring a true pattern that looks uninteresting can result in Type II errors, which may further lead to missed insights to inform continuous quality improvement. 
  3. Drill-down fallacies: Incomplete drill-down analysis can lead to omission of confounding factors and produce a class of errors known as drill-down fallacy (Lee et al., 2019). Here, a deviation found in the dataset is incorrectly attributed to a smaller subpopulation while the root cause at the parent level population is overseen. In Figure 1, if a user takes the drill-down path from ‘women’ to ‘Underrepresented minority (URM)-women’, they might infer that the perceptions of ‘URM women’ are very different from the reference group of ‘women’. While only 40% of ‘women’ said they felt isolated or judged as a member of their identity, 70% and 80% of ‘URM women’ respectively were in positive agreement. On the other hand, if the drill-down path is from the correct parent population (‘URM students’), the perceptions of ‘URM women’ will not be surprising. Although deviant from the reference group ‘women’, perceptions of ‘URM women’ are in accordance with ‘URM students.’

Discussion

When a large number of choices are presented to users, the resulting cognitive and information overload increases choice overload and effort in decision-making (Reutskaja et al., 2020). Resultingly, users end up performing only minimal drill-down operations guided by curiosity and not specific questions (Wise and Jung, 2019). 

To avoid drill-down fallacy in visualization, one approach is to explore all possible drill-down paths. However, since this approach is not scalable, there is no panacea for pitfalls of exploratory analysis on interactive LADs. When the insights are flawed due to fallacies, biases or omissions, there can be serious repercussions on the follow-up plans (Chan et. al. 2018). These challenges hamper the most effective use of data, especially by users without formal backgrounds in data analysis. In higher education institutes, considerable budget and human resource time is set aside in maintenance of interactive LADs and implementation of interventions resulting from rudimentary analytics. Absence of existing systems to evaluate the effectiveness of the action plans can lead to misallocation of resources based on flawed findings. It is important to include trained statisticians along with institutional leaders and assessment experts to develop consistent guidelines for the use of LADs.

For example, to mitigate MCP on interactive LADs, institutions could implement automatic statistical corrections for multiple testing, limit unstructured data fishing by prioritizing hypothesis-driven analysis over unrestricted exploration, and emphasize effect sizes and confidence intervals to prevent chance findings. Routine data disaggregation by relevant contextual variables, multivariate analysis, and statistical literacy training to ensure that aggregated performance metrics do not obscure subgroup trends can help overcome Simpson’s paradox. Decision thresholds can be embedded in policy by establishing predefined criteria for action (e.g., persistent differences observed across multiple semesters before intervention). Overall, thoughtful dashboard design processes would ensure that conclusions drawn from assessment data are both statistically sound and educationally responsible.

Future Directions and Recommendations

With the overwhelming permutations and combinations available for exploratory analysis on LADs, currently there is no proper documentation on whether analytics options are consumed in intended ways and instrumental in developing effective interventions and institutional strategic plans. Multi-institutional studies conducted on the effectiveness of the use of LADs will help determine if institutional resources are utilized in the best possible ways. Another important step would be to provide ‘intelligent nudges’ to users in the form of meaningful drill-down options and to conduct experimental studies that examine the differences between decision-making with and without nudges (Thaler, 2018).

 

References

Behrens, J. T. (1997). Principles and procedures of exploratory data analysis. Psychological Methods, 2(2), 131–160.

Chan, T., Sebok‐Syer, S., Thoma, B., Wise, A., Sherbino, J., & Pusic, M. (2018). Learning analytics in medical education assessment: the past, the present, and the future. AEM Education and Training2(2), 178-187.

Iyengar, S. S., & Lepper, M. R. (2000). When choice is demotivating: Can one desire too much of a good thing? Journal of Personality and Social Psychology, 79(6), 995–1006.

Kahneman, D. (2011). Fast and slow thinking. Allen Lane and Penguin Books, New York2.

Lee DJ-L, Dev H, Hu H, et al. (2019) Avoiding drill-down fallacies with vispilot: Assisted exploration of data subsets. Proceedings of the 24th International Conference on Intelligent User Interfaces, {IUI} 2019, Marina del Ray, CA, USA, March 17-20, 2019.

Reutskaja E, Iyengar S, Fasolo B, et al. (2020) Cognitive and affective consequences of information and choice overload. Routledge handbook of bounded rationality. Routledge, pp.625-636.

Shabaninejad S, Khosravi H, Indulska M, et al. (2020) Automated insightful drill-down recommendations for learning analytics dashboards. Proceedings of the Tenth International Conference on Learning Analytics & Knowledge. 41-46.

Thaler, R. H. (2018). Nudge, not sludge. Science361(6401), 431-431.

Tversky, A., & Kahneman, D. (1974). Judgment under Uncertainty: Heuristics and Biases: Biases in judgments reveal some heuristics of thinking under uncertainty. science185(4157), 1124-1131.

Verbert, K., Duval, E., Klerkx, J., Govaerts, S., & Santos, J. L. (2014). Learning analytics dashboard applications. American Behavioral Scientist, 57(10), 1500–1509.

Zhao Z, De Stefani L, Zgraggen E, et al. (2017) Controlling false discoveries during interactive data exploration. Proceedings of the 2017  Association for Computing Machinery acm international conference on management of data. 527-540.