Skip to main content
Top

Beyond Frequency: Using Epistemic Network Analysis and Multimodal Traces to Understand Temporal Dynamics of Self-Regulated Learning

  • Open Access
  • 19-10-2024
Published in:

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

The study investigates the temporal dynamics of self-regulated learning (SRL) in STEM environments using epistemic network analysis and multimodal traces. By collecting students' verbal and behavioral data, the research compares SRL processes between mastery and progressing students, revealing significant differences in the frequency, duration, and sequence of SRL events. The use of advanced analytical tools like ENA offers a nuanced perspective on the interplay between conscious and non-conscious thoughts and actions during learning, contributing to the development of targeted interventions to enhance student outcomes.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Introduction

Navigating the complex and often ill-structured nature of STEM technology-enhanced learning environments presents significant challenges for students, highlighting the critical importance of developing effective strategies to support their success (Kizilcec et al., 2017). Extensive research in the field of self-regulated learning (SRL) has consistently demonstrated a strong link between students’ SRL skills and their success in STEM contexts (Xu et al., 2023). SRL is the process by which individuals strategically manage their goal pursuit through metacognitive planning, monitoring, evaluating, and adapting their cognition, behavior, motivation, and affect (Greene et al., 2024). This enables them to achieve desired goals across various contexts and throughout their lifespan. However, understanding, evaluating, and providing effective support for SRL in STEM learning environments, as well as education in general, is a challenging task. One of the challenges arises from the inherently temporal and sequential nature of SRL (Azevedo, 2014).
SRL, like learning in broader contexts, is temporally bounded and unfolds over time (Azevedo, 2014; Reimann, 2009). Recognizing this temporal dimension is crucial for a nuanced understanding of how SRL-related events dynamically interact and influence each other over the course of the learning process (Ben-Eliyahu & Bernacki, 2015; Winne & Hadwin, 1998). Achieving insights into temporal dynamics in SRL processing, however, requires not only fine-grained data collection and methods (Bernacki, 2018) but also modeling approaches that accurately account for the sequential structure of the events during the process (Sung et al., 2022).
To address these concerns, our study employed multimodal approaches (Azevedo & Gašević, 2019) to assess SRL processes. We collected students’ multimodal traces, including both their think-aloud verbal trace data and their behavioral log digital trace data collected during technology-enhanced biology learning in a lab setting. Think-aloud verbal trace data allow for concurrent data collection of students’ reflections during SRL processes as they occur, providing relatively accurate insights into mental processes, whereas behavioral log trace data capture non-conscious or non-verbalized aspects of mental processes. By analyzing these multimodal traces, we investigated an aspect of SRL theory that has not been sufficiently tested: how temporal and sequential SRL processing differs between mastery and progressing students. By examining how SRL varies across different achievement levels, researchers can better understand the dynamic interplay between SRL processing and academic performance, thereby informing targeted interventions to support students’ learning processes, ultimately enhancing their learning outcomes. Using a multi-method approach, we explored count-based, duration-based, and sequence-based models of students’ SRL processing to uncover various aspects of the SRL process between different achievement levels. The count-based model served as a baseline, conventional approach that provides neither temporal nor sequential aspects of SRL events but rather a flattened representation of the events based on their frequency. The duration-based model focused on the distribution of time spent on SRL events, representing the temporal aspect of the SRL process. The sequence-based model, reflecting both temporal and sequential evolution of SRL events over time, employed an interleaving approach, which integrated both the think-aloud and digital trace data into a single, temporally accurate sequence. For the sequence-based model, we employed epistemic network analysis (ENA) as an analytic tool to operationalize and visualize the temporal structure of SRL processing.
Our research makes theoretical contributions by advancing understanding of the mechanisms underlying SRL processes during STEM learning. By adopting multimodal approaches that capture both conscious and non-conscious thoughts and actions during the process, we offer a nuanced perspective on students’ SRL processing within STEM contexts. Furthermore, our study makes methodological contributions by comparing different methods for modeling temporality and by utilizing a novel analytic approach (i.e., ENA) to model the temporal dynamics of SRL processes. While STEM education serves as a focal context due to the critical role of SRL in student success within this domain, our findings also contribute to an enhanced understanding of SRL dynamics that are broadly applicable across diverse educational contexts.

Theoretical Background

Approaches to Understand and Capture Students’ SRL Processing

During SRL processing, learners actively pursue desired learning goals through planful, evaluative, and reflective processes, dynamically shifting their focus between broader learning phases and more fine-grained moment-to-moment actions (Schunk & Greene, 2018). In SRL research, studies employing event-based methods (Winne & Hadwin, 1998) to examine SRL processes have established coded think-aloud protocol (TAP) data to identify specific, micro-level SRL processes (Bannert et al., 2014; Greene & Azevedo, 2007, 2009). In a think-aloud protocol (Ericsson & Simon, 1998), participants are asked to verbalize their thoughts, feelings, and actions, which researchers then code for micro-level SRL processing. Some of these micro-level codes have also been expanded to include aspects such as valence (Greene et al., 2018). These micro-level SRL processing codes, such as judgment of understanding, can then be aggregated into macro-level processes like monitoring (Greene & Azevedo, 2009). By collecting concurrent data of students’ reflections as they occur, think-aloud protocols offer fine-grained measurement and relatively accurate picture of SRL processing. However, this measurement method may not capture non-conscious or non-verbalized aspects of cognitive processes (Saint et al., 2020).
Given the limitation of any single method of capturing SRL, there is a growing interest in integrating multichannel and multimodal methods to investigate SRL processing. For example, digital trace data (Bernacki, 2018) can automatically capture and timestamp SRL behaviors that occur during the learning process (e.g., Azevedo et al., 2022). Researchers then can integrate data collected across other modalities such as eye tracking or affective data, to capture a more detailed, more varied, and accurate representation of SRL processing. However, using multichannel and multimodal data for understanding SRL processing presents challenges. These challenges include temporal alignment of multiple data streams, handling variations in data granularity, difficulty in interpretation, and other methodological and analytics challenges (Azevedo & Gašević, 2019). A primary challenge, however, stems from the inherently temporal and sequential nature of SRL (Azevedo, 2014).
Early SRL research commonly employed frequency-based modeling to investigate students’ SRL patterns (Saint et al., 2022). Conceptually, frequency serves as a proxy for the total occurrence of various self-regulatory processes, such as goal-setting, monitoring, and reflection. Methodologically, frequency is typically assessed by counting observable occurrences of SRL-related verbal traces or behaviors over a given period. Research has shown that frequencies of certain SRL behaviors correlate positively with better learning outcomes (e.g., Greene & Azevedo, 2009; Winters et al., 2008). However, frequency-based analyses do not capture the full complexity of the SRL process. Other factors such as the duration, sequence, and context of learning behaviors also shape SRL processing and subsequent learning outcomes (Maldonado-Mahauad et al., 2018). Recognizing this, time-based measurements in SRL research have emerged. One such approach is duration-based modeling, which assesses the amount of time spent on particular SRL processes, offering a proportional sense of their importance in SRL processing. Research has shown that the proportion of time allocated to specific SRL processes can predict learning outcomes (e.g., Taub et al., 2014), highlighting the importance of temporal considerations in understanding SRL. Another approach in time-based measurements is sequence-based modeling. This type of modeling focuses on particular common sequences of multiple SRL behaviors or events and how often they occur over time (Molenaar, 2014). Thus, frequency-based, duration-based, and sequence-based modeling of the SRL process are all being used to investigate the dynamic interplay of SRL processing (Saint et al., 2022).

Using ENA to Investigate Temporality-Focused SRL Processing

In the systematic review by Saint et al. (2022), which focused on temporally and sequentially oriented analyses of SRL, the most common approach identified was the use of frequency and time metrics to examine SRL processes. This demonstrates a missed opportunity where single-event-based methods are being used to study a temporal, multi-event process. Addressing this, Saint et al. (2022) indicated that among studies focused on temporality and sequence (n = 53), 13% employed ENA to examine co-temporality of the events in SRL processing and related strategies (e.g., Fan et al., 2021; Uzir et al., 2020). ENA is an analytical technique that identifies and quantifies connections among cognitive elements in discourse, both in individual and collaborative settings (Shaffer et al., 2016). It offers insights into temporal interconnections among learning behaviors by modeling co-occurrence of the events within recent temporal context (for more details, see Shaffer, 2017). ENA has been widely applied to activity-based learning analysis, including SRL research (Saint et al., 2022). In particular, multimodal traces, which encompass both conscious and non-conscious thoughts and actions during learning via TAPs and digital traces, are particularly suited for ENA modeling of the sequence and co-occurrence of these events compared to methods like process mining or ordered network analysis (ONA; Tan et al., 2022)), which typically assumes directional sequences. The temporal relationship among these multimodal traces in SRL processes can be less linear and more contextually intertwined. For instance, a student’s digital trace, such as submitting a quiz response, can occur before, during, or after think-aloud verbal utterances related to their SRL processes. In this context, ENA’s capability to account for the co-occurrences of multimodal traces within a recent temporal context using a time window is adept at capturing these complex interactions in a non-linear temporal framework. This approach enables us to explore how various SRL events may occur simultaneously or in close temporal proximity, providing insights into the dynamic and intertwined nature of SRL processes during learning activities.
However, when using ENA, researchers must consider several analytic choices to optimize its utility. These include data segmentation, event granularity (i.e., unified or varied analytic time unit), temporal window size, and guarding against model overfitting. As these analytic choices can significantly impact ENA results (e.g., Siebert-Evenstone et al., 2017), researchers must possess a thorough understanding of the data and its context. Additionally, ENA tends to emphasize stronger connections, potentially downplaying less frequent events or connections. Considering these factors throughout the process, researchers should employ an iterative and multi-method approach to data analysis and interpretation by looping back and forth between the data and the model.

Research Questions

In this study, we deployed frequency, duration, and sequential analyses of the same multimodal data from a learning task to investigate each method’s contribution to understanding the dynamics of SRL processing. Our investigation was guided by the following research questions:
  • RQ1: How does count-based SRL processing, as indicated by the frequency of multimodal traces, differ between students who mastered learning objectives and who were progressing towards mastery?
  • RQ2: How does duration-based SRL processing, as indicated by the proportions of time spent on multimodal traces, differ between students who mastered learning objectives and who were progressing towards mastery?
  • RQ3: How does sequence-based SRL processing, as indicated by the temporal interconnections among multimodal traces, differ between students who mastered learning objectives and who were progressing towards mastery?
Specifically, for RQ3, we used ENA of multimodal traces to explore how sequences of SRL processes might differ across groups. This comparison leveraged ENA’s utility in visually representing distinct cognitive structures within multimodal discourse between two groups through comparative analysis (e.g., Sung & Nathan, 2024). Also, we employed a qualitative approach to thicken our investigation into the theorized relations between SRL processes that were observed in ENA models, allowing for a multi-method, nuanced interpretation of micro-level SRL processes.

Methods

Participants

We recruited 49 undergraduate students from two parallel sections of the same introductory biology course offered in the Spring 2020 semester at a large public university in the southeastern United States. Participants were recruited through in-person announcements during lectures, weekly class emails, and the learning management system (LMS), Sakai, which was used in the course. Participants received a $75 gift card for their time upon completion, which typically lasted for an average of 75 min. One participant was excluded from the final dataset due to difficulties in verbalizing their thoughts aloud. Consequently, our analyses comprised data from 48 participants, aged between 18 and 26 years old. The majority of participants were first-year students (N = 40), and out of the total, 36 identified as female and 9 as biology majors (see Table 1 for detailed demographic characteristics of participants).
Table 1
Detailed demographic characteristics of participants (N = 48)
Variable
Category
Participants frequency (N)
Participants percent (%)
Gender
Male
12
25.00%
 
Female
36
75.00%
Year in school
First semester or first year
40
83.33%
 
Sophomore
3
6.25%
 
Junior
4
8.33%
 
Post baccalaureate
1
2.08%
Age
18
21
43.75%
 
19
24
50.00%
 
21
2
4.17%
 
26
1
2.08%
Major
Biology
9
18.75%
 
Pre-health
9
18.75%
 
Psychology
8
16.67%
 
Chemistry
5
10.42%
 
Undecided
4
8.33%
 
Others
13
27.08%
TOTAL
 
48
 

Procedures and Materials

This study closely followed ethical guidelines, obtaining written informed consent from participants prior to their involvement in any research activities. The research protocol was approved by the Institutional Review Board (IRB), ensuring participant confidentiality. Participants were informed of their right to withdraw at any time, and data were securely stored and accessible only to the research team.
Upon arrival at the lab, participants received an overview of the session and were asked to review and sign consent forms. Following this, they were instructed on the think-aloud protocol and practiced. The lab sessions were conducted individually, with each participant working one-on-one with one or two research assistants. Primarily, one researcher managed the session, ensuring that participants verbalized their thoughts effectively. In instances where two researchers were present, one served as the primary facilitator, while the second provided assistance as needed. During the lab session, participants were provided with learning materials from Lesson 23: How Populations Evolve II, part of an introductory biology course. The introductory biology course employed a high-structure active learning (Eddy & Hogan, 2014) course design across pre-class, in-class, and after-class phases, incorporating various learning activities. High-structure active learning is a student-centered educational practice demanding students’ intentional engagement in learning activities for knowledge construction (Bonwell & Eison, 1991; Lombardi et al., 2021). This course design was mirrored in the lab study. In the pre-class phase, participants were told to prepare for class as they typically would. During the in-class phase, participants watched a recorded video and answered practice questions on Learning Catalytics. Following this, in the after-class phase, participants completed a quiz on Mastering Biology. At the end of the session, participants filled out a demographics survey.

Pre-class, In-class, and After-class Learning Activities During the Lab Study

Guided Reading Questions
A set of six open-ended questions called guided reading questions (GRQs) were designed to assist students in comprehending the learning content by producing written responses to these prompts. Participants received the GRQs before the lab study to prepare and these GRQs could be downloaded and submitted to the LMS during the learning process. In the lab study, participants were required to complete both a designated reading passage from the course textbook and GRQs.
Homework
Another pre-class learning task given to participants was homework assignments, comprising of 5-multiple choice questions. Participants were allowed multiple attempts for each question and could request hints for up to three questions.
Learning Catalytics
Pearson’s Learning Catalytics, a web-based interactive classroom response system, facilitated real-time student engagement and assessment during in-class learning activities. It featured two questions designed to promote student interaction and participation while watching the lecture video.
Quiz
The quiz served as the learning task for the after-class phase and consisted of seven items, one of which had multiple parts. It was graded and used for identifying student groups with distinct learning performance. More details can be found in the “Identifying the Mastery And Progressing Groups Based on Learning Outcomes” section.

Measures: Multimodal Traces of SRL Processing

Think-Aloud Protocols

Verbal trace data were obtained using TAPs (Greene et al., 2018). Participants were instructed to verbalize their mental processes, including thoughts, emotions, and actions. It was emphasized that participants should refrain from elaborating or providing explanations for their mental processes, as studies have shown that doing so can alter the nature of their cognitive processing (Fox et al., 2011). While engaging with learning tasks, if a participant remained silent for more than three seconds, the researcher prompted them by saying, “Please keep talking.” TAP data were audio-recorded and transcribed.

Digital Traces

During the lab session, students interacted with multiple technology-enhanced learning environments, including the university’s LMS (i.e., Sakai). The LMS course site (see Fig. 1), designed to resemble the structure of the participants’ current semester’s biology course, served as the central platform for online activities, hosting course materials and providing links to other digital platforms used in the study. Participants completed homework and quizzes, while also submitting answers to formative items during lecture videos, akin to their regular coursework. Interactions with multiple platforms were automatically logged via server log files, capturing timestamped records of participants’ activities such as page loads, file downloads, and item submissions.
Fig. 1
Screenshot of the LMS course site, Sakai
Full size image
Figure 2 provides a visual timeline of the different learning activities and data collection during the lab session.
Fig. 2
Procedure of learning activities during the lab study
Full size image

Data Preparation and Analysis

Identifying the Mastery and Progressing Groups Based on Learning Outcomes

We used participants’ quiz scores as a measure of learning outcomes from the lab session. The average score across the seven quizzes was 79.75, ranging from 0.5 and 100, with a standard deviation of 23.47. Based on the histogram (see Fig. 3) displaying the distribution of quiz scores, we identified two distinct groups: (a) participants who achieved nearly full scores, demonstrating mastery-level learning achievement (21 students achieving ≥ 95% correct; henceforth, mastery group (blue)), and (b) participants whose scores reflected progress towards mastery-level achievement (27 students; henceforth, progressing group (red)).
Fig. 3
Histogram displaying the distribution of students’ quiz scores, with categorization into mastery and progressing groups (x-axis: percentage of correct quiz responses, y-axis: number of students)
Full size image

Think-Aloud Protocol Data

The audio-recorded TAP files were transcribed, and each researcher independently analyzed the transcripts. Verbal utterances were segmented into contextually meaningful units (1171 segments) and coded using a codebook adapted from Greene and Azevedo (2009). This codebook comprised over 50 TAP micro-level codes nested within six TAP macro-level codes, including task-specific processes (e.g., planning, restating a sub-task or sub-goal), monitoring (e.g., feeling of understanding, with valence), assessment strategies (e.g., guessing, ruling out answers), domain-general strategies (e.g., re-reading, summarizing), and domain-specific strategies (e.g., mathematical problem-solving). The coding scheme containing details of the TAP macro-level codes is shown in Table 2, whereas the list of TAP micro-level codes can be found in the Appendix. With the assistance of Nvivo, software supporting human annotation on audio or video recordings, the precise duration of each codable verbal utterance was tracked in seconds using the starting and ending timestamps of the event. Interrater reliability was assessed using percentage agreement instead of Kappa values due to sparse application of some micro-level codes, resulting in approximately 70% agreement between coders.
Table 2
Coding scheme for multimodal traces during SRL processing, including TAP macro-level codes and Trace codes
Code
Description
Example
TAP: task-specific
(Re)stating learning sub-goals or identifying the attributes of the task
“I’m going to click on the syllabus to see what I need to read.”
TAP: monitoring
Students’ metacognitive judgment of their learning or the relevance of contents to their learning, with valence ( ±)
 + : “That article was really helpful.”
-: “These pictures make it more confusing.”
TAP: assessment strategies
Students’ use of learning strategies while engaging in the learning assessment, including searching contents, ruling out answers, and reviewing feedback
“My guess is taxes would probably go up in order for the government to pay for all of the healthcare.”
TAP: domain-general strategies
Students’ use of general SRL strategies, including re-reading, summarizing, and drawing a new conclusion by combining information
“Reading the ones on the side to see if there’s any keywords.”
TAP: domain-specific strategies
Students’ verbalization while mathematical problem-solving, including writing out formula, using a calculator, and double-checking the calculation
“The total population is 240 plus 180 plus 80 so that’s 500 over 1,000.”
Trace: guided reading questions
Students’ digital log events related to the guided reading questions
The log events of downloading the guided reading questions, submitting their responses, and revising them
Trace: in-class tasks
Students’ digital log events related to the learning catalytics, which were in-class tasks
The log events of submitting responses to the in-class learning catalytics
Trace: pre-class tasks: incorrect
Students’ digital log events indicate that their pre-class homework responses are incorrect
The log events of submitting incorrect answer to pre-class homework
Trace: pre-class tasks: correct
Students’ digital log events indicate that their pre-class homework responses are correct
The log events of submitting correct answer to pre-class homework

Digital Trace Data

Digital trace data capturing participants’ interaction with digital tools during the lab session were collected from multiple digital platforms. In this study, we focused on participants’ digital traces related to pre-class and in-class learning tasks, such as GRQs, homework, learning catalytics, and quizzes, as these interactions provided insights into their engagement with these learning tasks. From this digital trace data, we generated four Trace codes (see Table 2 for details). Note that digital traces pertaining to quizzes were excluded from the Trace codes because these logs directly correspond to quiz scores, which were used to classify the mastery and progressing groups.

Data Preparation for RQ1 and 2

To address RQ1, which is count-based modeling of SRL processing, we first computed the aggregated counts of multimodal traces exhibited during the learning process in both the progressing and mastery groups. Subsequently, we normalized these raw event counts by the duration of individual task sessions (measured in seconds). This normalization step was essential due to the possibility of individual task session times being extended, particularly when students made errors in their homework or in-task quiz submissions, necessitating re-attempts. Such extensions could produce an increase in the number of additional SRL-related codes during the process by virtue of the additional time spent overall. The normalization process involved multiplying these counts by the average session time across all students (4464.42 s) to represent the average normalized frequency of multimodal traces in each group. Using such values, then we conducted a two-tailed independent t-test, comparing the progressing and mastery groups.
In addressing RQ2, which is duration-based modeling of SRL processing, we tracked the duration of each event in the multimodal traces throughout the learning process and quantified its time span. Subsequently, the duration of each code occurrence was divided by the total duration of individual task sessions (measured in seconds). For instance, if a student spent a total of 200 s verbalizing task-specific processes (task-specific in TAP macro-level), and their task session lasted 4000 s in total, the proportion of time spent on task codes for that student would be calculated as 5% of the total in-task time. Log events in the digital traces (i.e., Trace Macro-level codes), typically occurring in less than a second, were considered to be timed at 1 s. We then compared these proportions of time spent on multimodal traces between the mastery and progressing groups.
In this study, we did not exclude any outliers from our analysis. We carefully monitored and ensured that all participants diligently engaged in the experiment, thereby minimizing the possibility of excessively long session times or non-engagement. Consequently, our data analysis included all collected data points.

Data Preparation for RQ3: Temporal Alignment of Multimodal Traces Using Interleaving Approach for ENA Modeling

To model SRL in sequence-based fashion via ENA (RQ3), we conducted temporal alignment by organizing the coded multimodal trace data in a temporally interleaved manner (Sung et al., 2022). Initially, we sorted the coded multimodal trace data based on the starting time of each event to reflect the original temporal sequence of these interactions that occurred during the SRL process. Importantly, since verbal utterances in TAP data were more likely to have longer durations compared to click events in Trace data, students often made click events while verbalizing their SRL process. In such instances, to account for the temporally interleaved interactions among multimodal traces, when a digital trace occurred within the timespan of a TAP trace, we coded that timepoint as both a digital and a TAP trace. For example, as illustrated in the data structure example (see Fig. 4), while participant #8 made an utterance related to assessment strategies (17:43:18 – 17:43:26), they contemporaneously made a click event, submitting an incorrect response in homework submission (17:43:22). This digitally traced event occurred within the verbalization span. Here, the line of TRACE code was assigned the same TAP code that it bisected, producing an interleaved structure, one that more accurately reflected “temporally entangled” multimodal interactions (Sung et al., 2022) during the SRL process.
Fig. 4
Example of temporally interleaving coded multimodal trace data
Full size image
As depicted in Fig. 4, each row in the dataset represents segments, which are the smallest unit of meaning analyzed (Shaffer, 2017). Both TAP and TRACE data were segmented based on the occurrence of detectable micro-level codes associated with SRL processes (see the Appendix for details). These segments were then grouped into stanzas, which encompass collections of rows considered to occur within the same recent temporal context.
To investigate potential differences in sequence-based SRL processing between the mastery and progressing groups, and to account for temporal interconnections among multimodal traces, we constructed ENA models including ENA scatter plots and networks. For the ENA network modeling, we used a moving stanza window size of seven rows (the current row plus previous six rows). This approach enabled us to calculate the co-occurrences among codes within a sliding window size of seven rows over time. The selection of this window size was determined through a grounded analysis of the multimodal traces data from each group during SRL processing.
Following ENA modeling, we conducted qualitative analysis to reinforce and contextualize the findings from ENA by providing an in-depth examination of the emergence of micro-level SRL processes. This multi-method approach allows for a nuanced understanding of students’ SRL processes as captured by multimodal traces, providing a richer interpretation of the temporal and sequential dynamics underlying these processes.

Results

RQ1: Count-Based SRL Processing (Model 1)

Results (see Table 3) revealed that students in the progressing group (M = 152.55, SD = 42.58) exhibited statistically significantly higher overall SRL-related code occurrences than those in the mastery group (M = 129.92, SD = 33.89), with a moderate effect size (t = 2.05, p < 0.05, Cohen’s d = 0.58). Specifically, we found that students classified in the progressing group produced statistically significantly higher occurrences of task-specific (t = 2.50, p < 0.01, Cohen’s d = 0.73), assessment strategies (t = 2.03, p < 0.05, Cohen’s d = 0.58), and domain-general strategies (t = 2.53, p < 0.01, Cohen’s d = 0.71) in TAP macro-level codes, all with a moderate effect size. This shows that, on average, students in the progressing group, who exhibited relatively lower performance in in-task quizzes, tended to verbalize more about task-specific processes (e.g., restating a sub-task or sub-goal), assessment strategies (e.g., guessing), and domain-general strategies (e.g., re-reading, summarizing) compared to those in the mastery group (for t-test results on the raw occurrences of macro- and micro-level codes to see the results without normalization, see the Appendix). Furthermore, students in the progressing group exhibited statistically significantly higher occurrences of pre-class tasks: incorrect (t = 2.73, p < 0.01, Cohen’s d = 0.77) events but lower on guided reading questions (t =  − 2.11, p < 0.05, Cohen’s d =  − 0.62) events in Trace macro-level codes, both with a moderate effect size, indicating that they were more likely to submit more incorrect responses in their pre-class homework submissions, whereas also less inclined to revisit guided reading questions compared to those in the mastery group.
Table 3
Model 1: T-test results of the counts of TAP and Trace macro-level codes between the progressing and mastery groups (N = 48)
Codes
Progressing group (N = 27)
Mastery group (N = 21)
t
p-value
Cohen’s d
M
SD
M
SD
TAP Macro
   task-specific
26.44
8.75
19.93
9.17
2.50
0.02**
0.73
   monitoring
44.58
25.46
37.12
15.47
1.25
0.22
0.34
   assessment strategies
15.64
8.19
11.25
6.82
2.03
0.05*
0.58
   domain-general strategies
25.84
11.94
18.23
8.86
2.53
0.01**
0.71
   domain-specific strategies
26.67
8.02
29.46
7.06
 − 1.28
0.21
 − 0.37
Trace Macro
   guided reading questions
3.30
1.84
4.52
2.09
 − 2.11
0.04*
 − 0.62
   in-class tasks
2.83
0.48
2.93
0.90
 − 0.47
0.64
 − 0.15
   pre-class tasks: incorrect
3.04
2.33
1.44
1.74
2.73
0.01**
0.77
   pre-class tasks: correct
4.22
1.68
5.05
1.43
 − 1.85
0.07
 − 0.53
Total code occurrences
152.55
42.58
129.92
33.89
2.05
0.05*
0.58
* p < 0.05, ** p < 0.01; raw aggregated event counts were normalized by individual task session time (measured in seconds), then multiplied by the average session time (4464.42 seconds). A negative Cohen’s d indicates that the second mean (Mastery group) is higher

RQ2: Duration-Based SRL Processing (Model 2)

The t-test results (see Table 4) revealed that students in the progressing group (M = 30.93%, SD = 9.78%) spent statistically significantly more time producing TAP or digital traces during the task compared to those in the mastery group (M = 26.39%, SD = 5.02%), with a medium effect size (t = 2.09, p < 0.05, Cohen’s d = 0.56). Specifically, students in the progressing group allocated statistically significantly higher proportion of time spent on task-specific, with a moderate effect size (t = 2.27, p < 0.05, Cohen’s d = 0.66), indicating that they proportionally devoted more time to verbalizing task-specific processes (e.g., restating a sub-task or sub-goal) compared to those in the mastery group. Moreover, students in the progressing group exhibited statistically significantly higher proportion of time spent on pre-class tasks: incorrect (t = 2.73, p < 0.01, Cohen’s d = 0.77), but lower on guided reading questions (t =  − 2.11, p < 0.05, Cohen’s d =  − 0.62) in macro-level codes of Trace data compared to the mastery group. This shows that students in the progressing group spent proportionally more time on addressing errors in their pre-class homework submissions but less time on engaging guided reading questions for review, which aligns with the findings from RQ1.
Table 4
Model 2: T-test results of the temporal distributions of TAP and Trace macro-level codes between the progressing and mastery groups (N = 48)
Codes
Progressing group (N = 27)
Mastery group (N = 21)
t
p-value
Cohen’s d
M
SD
M
SD
TAP Macro
   task-specific
3.92%
1.57%
2.92%
1.47%
2.27
0.03*
0.66
   monitoring
3.51%
2.72%
2.75%
1.21%
1.29
0.21
0.34
   assessment strategies
4.56%
3.55%
3.51%
2.30%
1.25
0.22
0.34
   domain-general strategies
5.76%
7.94%
3.45%
2.43%
1.42
0.16
0.37
   domain-specific strategies
12.88%
2.80%
13.44%
2.54%
-0.72
0.47
-0.21
Trace Macro
   guided reading questions
0.07%
0.04%
0.10%
0.05%
-2.11
0.04*
-0.62
   in-class tasks
0.06%
0.01%
0.07%
0.02%
-0.47
0.64
-0.15
   pre-class tasks: incorrect
0.07%
0.05%
0.03%
0.04%
2.73
0.01**
0.77
   pre-class tasks: correct
0.09%
0.04%
0.11%
0.03%
-1.85
0.07
-0.53
Total duration
30.93%
9.78%
26.39%
5.02%
2.09
0.04*
0.56
* p < 0.05, ** p < 0.01; the aggregated duration of each code occurrence was divided by the total duration of individual task sessions (measured in seconds), resulting in the proportions of time spent on each code relative to the total in-task time. The remaining time proportions were attributed to other activities, such as reading content or watching lecture video, which were not considered in the SRL-related codes for this study. A negative Cohen’s d value indicates that the second mean (Mastery group) is higher

RQ3: Sequence-Based SRL Processing (Model 3)

Quantitative and Network Analyses Results Using ENA Models

We constructed ENA models using the coded multimodal traces of students’ SRL processing. First, we generated an ENA scatter plot, which serve as a graphical representation of plotted points, to examine whether statistically significant differences existed in multimodal traces of SRL processing between the progressing (depicted in red) and mastery (blue) groups (see Fig. 5). Each round plotted point on the scatter plot represents the network location of an individual student’s multimodal traces during SRL processing, determined by the average of the unit’s node weights in each network. To explore the difference between the two learner groups, we compared each group’s group means, depicted by the larger square points (1 red, 1 blue), along with 95% confidence intervals (t-distribution) represented as dashed boxes in red and blue, respectively. Non-overlapping boxes indicate statistically significant differences between the two means at the 5% level. The statistical analysis affirmed significant differences in the patterns of multimodal traces during SRL processing between the progressing and mastery groups, displaying a moderately large effect size (MProgressing = 0.06, MMastery =  − 0.08, t(45.99) = 2.66, p < 0.05, Cohen’s d = 0.75).
Fig. 5
ENA scatterplot: the progressing (red) and mastery (blue) groups, portraying individual student’s network locations of multimodal traces during SRL processing
Full size image
To examine which connections among the multimodal traces contributed to the differences between the progressing and mastery groups, we constructed mean epistemic networks for each group and conducted a comparative analysis of these networks (see Fig. 6). These ENA networks visualize the network graphs of each group, with nodes representing codes (listed in Table 2) and edge thicknesses indicating the relative frequency of co-occurrences among the codes. The mean subtracted network (Panel A of Fig. 6), obtained by subtracting one network from the other, distinctly illustrates the differences in students’ multimodal traces captured during SRL processing between the progressing (Panel B) and mastery (Panel C) groups. Panel A of Fig. 6 shows that during SRL processing, students classified as the mastery group (marked by blue connections) exhibited notably stronger connections between monitoring and domain-specific strategies of TAP macro-level codes. This indicates that in the mastery group, when students verbalized monitoring-related SRL processes (e.g., judgment of understanding, with valence), these utterances were more frequently accompanied by utterances related to domain-specific strategies (e.g., mathematical problem-solving), during SRL processing. On the other hand, students classified as the progressing group (marked by red connections in Panel A of Fig. 6) displayed relatively stronger connections between monitoring and domain-general strategies, as well as between assessment strategies and task-specific of TAP macro-level codes. This indicates that in the progressing group, students’ monitoring-related utterances were more frequently accompanied by utterances related to domain-general strategies (e.g., re-reading, summarizing), during SRL processing. Additionally, utterances related to the assessment strategies (e.g., guessing) were more frequently co-occurred with utterances pertaining to task-specific processes (e.g., restating a sub-task or sub-goal), compared to those in the master group.
Fig. 6
Mean epistemic network graphs of students’ multimodal traces during SRL processing: students in progressing group (red lines of Panel B), mastery group (blue lines of Panel C), and mean subtracted network (Panel A)
Full size image
Compared to TAP codes, the mastery and progressing groups did not show substantial differences in their TRACE codes, evidenced by relatively weak connections among the codes in the subtracted network, highlighting the difference between the two groups (see Panel A of Fig. 6). Still, students in the mastery group exhibited stronger connections related to guided reading questions and in-class tasks of TRACE macro-level codes, indicating more frequent co-occurrences of these digital trace events with other think-aloud verbalizations of SRL processes. In contrast, students in the progressing group demonstrated stronger connections with pre-class tasks: incorrect of TRACE macro-level codes, suggesting that their SRL-related verbalizations were more frequently associated with incorrect responses to pre-class learning tasks, which occurred more often than in the mastery group.
Interpreting ENA models involves analyzing the spatial positioning of codes within the ENA space (Shaffer, 2017). The y-axis reveals distinct patterns: students in the upper part, predominantly from the progressing group, prioritize domain-general strategies and assessment-related processes, indicating a broad approach to learning and testing strategies. In contrast, students in the lower part, primarily from the mastery group, concentrate on domain-specific strategies and intensive monitoring of their learning, reflecting a focus on subject-specific problem-solving and continuous self-assessment. This distribution underscores the diverse cognitive efforts and regulatory behaviors students employ, ranging from general learning techniques to targeted, reflective practices, which align with their learning outcomes.
In both groups, verbalizations of monitoring-related processes frequently co-occurred with other SRL codes. However, the specific codes temporally accompanying monitoring differed: in the mastery group, they were more frequently co-occurred with domain-specific strategies, whereas in the progressing group, they were more often linked with domain-general strategies. This discrepancy prompted further qualitative analyses to investigate the contexts of SRL processing in more depth.

Qualitative Results

To gain a deeper understanding of the nuanced differences in SRL processing between the two groups, we conducted qualitative analyses of students’ SRL processes in each group, with a particular focus on examining the connections between monitoring and domain-specific strategies (mastery group), and between domain-general strategies (progressing group), captured as TAP macro- and micro-level codes. These analyses were conducted on micro-level examination of the coded transcripts of SRL processing (see the Appendix for details of micro-level SRL codes).
Table 5 provides an excerpt from a student classified as the mastery group, as they verbalize their SRL process. In this excerpt, Student #41 is engaged in an in-task quiz involving mathematical problem-solving processes (mps, domain-specific strategies code; line 1). In line 2, the student engages in metacognitive monitoring by making a judgment that their answer to a question is incorrect (joc-), expressing “That’s not the answer we’re looking for.” Subsequently, the student assesses the accuracy of their mathematical computation (mma), stating “Why was I squaring it?” (monitoring code; line 3) This metacognitive assessment is linked to a more accurate mathematical problem-solving process (line 4), which is then followed by the student making another metacognitive judgment, affirming that their answer to a question is correct (joc + , monitoring code; line 5).
Table 5
An excerpt from a student in the mastery group during SRL processing
 
Speaker
Group
Coded transcript
TAP Macro
TAP Micro
1
Student #41
Mastery
Homozygous recessive, that's a genotype. So, it's asking about genotypes here, not allele frequencies
domain-specific strategies
mps
2
Student #41
Mastery
That’s not the answer we’re looking for
monitoring
joc-
3
Student #41
Mastery
Why was I squaring it?
monitoring
mma
4
Student #41
Mastery
It would be just that squared. Q equals 0.02. And then the genotype would be that squared
domain-specific strategies
mps
5
Student #41
Mastery
Feel good about that
monitoring
joc + 
MPS Mathematical Problem-Solving, MMA Monitoring Math Accuracy, JOC-/ +  Judgment of Correctness
As illustrated by this excerpt, students in the mastery group were observed to proactively engage in monitoring their learning during mathematical problem-solving processes. This proactive monitoring entailed cognitive processes during problem-solving such as identifying errors in their answers, ensuring mathematical accuracy, and making metacognitive judgments about their learning process.
On the other hand, students in the progressing group displayed distinct patterns in their SRL processing. In Table 6, an excerpt from a student in the progressing group is engaged with biology learning contents. Student #14 summarizes what they learned (sum) and draws a new conclusion by synthesizing information (fnc, domain-general strategies code; lines 1–2). However, Student #14 recognizes their lack of understanding of the learning content (jou-), stating “I don’t know what that is” (monitoring code; line 3). Following this, the student abruptly evaluates that the content is not particularly relevant to learning (ce-), expressing “I don’t think it matters too much” (monitoring code; line 4).
Table 6
An excerpt from a student in the progressing group during SRL processing
 
Speaker
Group
Coded transcript
TAP Macro
TAP Micro
1
Student #14
Progressing
If it’s completely dominant then you just double it, and if it’s heterozygous then you just add it
domain-general strategies
sum
2
Student #14
Progressing
0.8 expresses dominant right now
domain-general strategies
fnc
3
Student #14
Progressing
I don’t know what that is
monitoring
jou-
4
Student #14
Progressing
I don’t think it matters too much
monitoring
ce-
SUM Summarization, FNC Forming New Conclusion JOU-/ + Judgment of Understanding, CE-/ + Content Evaluation
As exemplified in this excerpt, unlike students in the mastery group, those in the progressing group appeared to engage more actively in monitoring their learning while accessing information from learning materials, rather than engaging in mathematical problem-solving during in-task quizzes. During the monitoring process, students in the progressing group seemed to focus primarily on evaluating the relevance of the content to their overall learning task, goal, or a question, rather than monitoring their own learning processes. Moreover, their evaluations of these contents often leaned towards the negative, with expressions such as “Not very helpful” or “This is not leading me to what I'm looking for.” Interestingly, this negative evaluation of relevance often followed students’ struggles to fully understand the contents.
These qualitative analyses align to the t-test results on the raw occurrences of macro- and micro-level codes between the two groups, which indicate that students in the progressing group produced twice as many instances of content evaluation (ce-) and judgment of understanding (jou-) compared to those in the mastery group (see the Appendix).

Discussion

Theoretical Contributions

In this paper, we investigated an understudied aspect of SRL theory: how temporal SRL processing, as captured by multimodal traces, differs between mastery and progressing groups. We examined count-based, duration-based, and sequence-based models of SRL processing to unveil diverse dimensions of the process across different achievement levels. Specifically, for the sequence-based model, we employed a multi-method approach, incorporating ENA to analyze the temporal and sequential dynamics of multimodal events during SRL processing and then qualitative analysis to delve deeper into those dynamics.
Our findings from RQ1, focusing on conventional frequency-based modeling of SRL processing (Model 1), revealed that students in the progressing group, characterized by relatively lower in-task quiz performance, exhibited more frequent occurrences of overall SRL-related utterances and log events compared to those in the mastery group. The verbalizations of progressing group students primarily focused on restating sub-goals, guessing an answer, and re-reading and summarizing. These tendencies appear to be associated with their increased efforts to correct homework submissions, as captured in digital traces, which prolonged their individual task times. Despite normalization of raw counts based on session durations and average session time across all students, frequency-based modeling remained highly sensitive to fluctuations in task durations and the frequency of learning activity attempts, often influenced by performance-related re-attempts. This suggests that if individual student’s task session durations fluctuate based on performance, frequency-based modeling may not provide an accurate depiction of SRL, nor fully capture the temporal and sequential aspects of the process (Saint et al., 2022).
For RQ2, focusing on duration-based modeling of SRL processing (Model 2), we found statistically significant differences in the time allocation for multimodal traces between the two groups. Specifically, students in the progressing group devoted a greater proportion of time to verbalizing task-specific processes, such as restating sub-tasks or sub-goals, or addressing errors in homework submissions, whereas they dedicated less time reviewing guided reading questions. Although some findings from Model 2 echoed those of Model 1, the statistically significant differences in the frequency of two TAP macro-level codes observed in Model 1, assessment strategies and domain-general strategies, vanished when adjusting for the proportions of time spent on SRL events. The discrepancy between the two models implies that although there may be differences between the mastery and progressing groups in the frequency of SRL events, especially for TAP codes that extend over time, these differences may not be reflective of the actual proportion of time allocated to those events during the learning process. In essence, differences in frequency may not necessarily translate into differences in the actual time spent on those events. This highlights the importance of examining SRL processing dynamics from both frequency and duration perspectives when utilizing multimodal traces, which encompass various time spans for occurrence.
For RQ3, focusing on sequence-based modeling of SRL processing using ENA (Model 3), statistically significant differences emerged in the patterns of multimodal traces during SRL processing between the progressing and mastery groups. Although the TRACE codes themselves did not present substantial differences, suggesting that the primary distinctions in SRL processing were more evident in verbalizations rather than digital trace events, important nuances were still observed. The mastery group showed stronger connections between digital trace events, such as guided reading questions and in-class tasks, and their SRL verbalizations, indicating a more intentional use of these digital resources to support their SRL processes. In contrast, the progressing group more frequently linked their SRL verbalizations with incorrect pre-class responses, suggesting that their SRL efforts were often focused on addressing the errors they encountered. Interestingly, when examining TAP codes, both groups frequently verbalized monitoring-related processes—the most prevalent among the SRL codes—yet no difference in frequency was detected between the two groups. However, the way this specific SRL process temporally and sequentially connected to other processes differed between the groups: it more frequently co-occurred with domain-specific strategies in the mastery group, but with domain-general strategies in the progressing group. Students in the mastery group demonstrated that they tended to primarily engage in monitoring their learning during mathematical problem-solving processes, contributing to their superior performance on in-task quizzes. In contrast, those in the progressing group appeared to be more actively engaged in monitoring their learning while accessing information from learning materials, primarily focusing on evaluating the relevance of the content to their overall learning task, goal, or question, rather than monitoring their own learning processes. Furthermore, their evaluations of these contents tended to be negative, often following struggles to fully comprehend the contents. These insights from our multi-method approach underscore the inherently temporal and sequential nature of SRL (Azevedo, 2014). As the SRL process unfolds over time (Bernacki, 2018), it is crucial to analyze these data in ways that take into account the temporal and sequential nature of SRL to gain an accurate understanding of how they dynamically interact and influence SRL processing.
In addition, our work contributes to advancing theoretical understanding of the mechanisms underlying SRL processes between mastery and progressing students in technology-enhanced STEM learning. By leveraging multichannel and multimodal data to assess and understand SRL processing (Azevedo & Gašević, 2019), which capture both conscious and non-conscious thoughts and actions during learning through think-aloud protocols and digital traces, we offer a nuanced understanding of mastery and progressing students’ SRL processing in STEM learning within technology-enhanced learning environments—the way specific SRL processes temporally and sequentially connect to other processes differed between the two groups, which lead to disparities in their academic performance. Doing so reveals how assumptions about differences in SRL processing between novices and experts (e.g., Zimmerman, 2002) require data collection and analysis techniques that go beyond what occurred, and instead can model how sequences of processing differ. This is evidence for more sophisticated modeling of the phenomenon of SRL, comprising more than individual processes (Greene, 2022). In addition, our findings underscore the need for targeted interventions to guide and assist low achieving students in effectively utilizing SRL strategies at various stages of the learning process to enhance their learning outcome.

Methodological Contributions

Our study also made contributions to methodologies for capturing and analyzing SRL. We conducted empirical investigations into various facets of the temporal SRL processing by collectively examining frequency-based, duration-based, and sequence-based modeling approaches in a unified manner. To our knowledge, no previous empirical study has undertaken such a comprehensive approach, and doing so allowed us to demonstrate how each modeling approach reveals unique aspects about temporal SRL processing.
Another methodological contribution is our empirical demonstration of implementing a novel analytic approach to model the temporal dynamics of SRL processes employing the interleaving approach and ENA. We systematically applied the interleaving approach to accurately model multimodal events occurring during the learning process, reflecting the temporal interconnections among them through temporal alignment of multiple data streams. By detailing how we operationalized the multimodal trace data, representing a “temporally entangled” data structure (Sung et al., 2022), our analytic approach provides insights into one method for modeling the dynamic interplay among multimodal interactions during the SRL process. Moreover, our study leverages ENA as a potent tool to model and visualize SRL process patterns while retaining sequential and temporal information of multimodal traces. Through a multi-methods approach that integrates quantitative findings with qualitative insights, we empirically showcased the analytical effectiveness of our approach, employing the interleaving and ENA, in temporally focused SRL analysis.

Limitations and Future Directions

This study has several limitations. First, due to the highly structured course design for individual learning, there were limited instances of interleaving between verbalization and digital traces during the learning process, which may have diminished the impact of the interleaving approach in modeling multimodal interactions. Given that the interleaving approach has shown its usefulness in modeling multimodal discourse in computer-supported collaborative learning context and revealing different patterns of collaborative multimodal interactions among students (e.g., Sung & Nathan, 2024), future researchers should explore its applicability in more typical technology-enhanced collaborative learning settings, where there are much more instances of students interacting with peers while exploring digital learning resources contemporaneously. Secondly, the current treatment of digital traces as momentary events ignores their duration. Future researchers should consider extracting temporal information, such as the duration of student engagement with digital resources, to better understand their impact and influence on the learning process.

Conclusion

Despite any limitations, our study offered valuable methodological contributions by empirically demonstrating how different modeling approaches unveil diverse aspects of temporal SRL processing captured in multimodal traces. Additionally, it makes theoretical contributions to advancing the understanding of the mechanisms underlying the temporal dynamics of SRL processing between mastery and progressing students in technology-enhanced STEM education. Future researchers can leverage these contributions to advance models of the dynamic and temporal nature of SRL.

Declarations

Ethical Approval

All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki declaration and its later amendments or comparable ethical standards.
Informed consent was obtained from all individual participants included in the study.

Conflict of Interest

The authors declare no conflicts of interest that could have influenced the design, execution, or reporting of this research.
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Title
Beyond Frequency: Using Epistemic Network Analysis and Multimodal Traces to Understand Temporal Dynamics of Self-Regulated Learning
Authors
Hanall Sung
Matthew L. Bernacki
Jeffrey A. Greene
Linyu Yu
Robert D. Plumley
Publication date
19-10-2024
Publisher
Springer Netherlands
Published in
Journal of Science Education and Technology / Issue 5/2025
Print ISSN: 1059-0145
Electronic ISSN: 1573-1839
DOI
https://doi.org/10.1007/s10956-024-10164-2

Appendix

See Table 7
Table 7
Counts of macro- and micro-level codes from TAP and Trace data (N = 48)
Codes
Progressing group (N = 27)
Mastery group (N = 21)
t
p-value
Cohen’s d
M
SD
M
SD
TAP: task-specific
27.56
10.26
19.43
9.56
2.83
0.01**
0.82
plan
0.52
0.80
0.24
0.62
1.36
0.18
0.38
recycle:target
19.63
9.35
13.76
8.17
2.32
0.03*
0.66
sg:target
3.48
3.25
2.19
1.97
1.70
0.10
0.47
tdfn
3.11
1.93
2.81
1.40
0.63
0.53
0.18
TAP: monitoring
46.74
27.94
36.57
17.25
1.55
0.13
0.43
ce-
2.48
3.57
0.81
0.93
2.33
0.03*
0.61
ce + 
2.96
3.08
2.76
2.95
0.23
0.82
0.07
eac-
0.22
0.58
0.19
0.40
0.22
0.82
0.06
eac + 
0.56
0.75
0.86
0.15
 − 1.04
0.31
 − 0.32
em
0.41
0.93
0.67
1.20
 − 0.82
0.42
 − 0.25
fok-
0.15
0.46
0.19
0.40
 − 0.34
0.73
 − 0.10
fok + 
0.04
0.19
0.05
0.22
 − 0.18
0.86
 − 0.05
for-
0.81
1.42
0.48
0.81
1.04
0.30
0.28
for + 
3.96
3.51
2.90
2.43
1.23
0.22
0.34
joc-
1.26
1.75
1.57
1.29
 − 0.71
0.48
 − 0.20
joc + 
1.56
1.69
1.71
1.27
 − 0.37
0.71
 − 0.10
jol-
0.07
0.27
0
0
1.44
0.16
0.37
jou-
8.30
6.41
4.43
3.36
2.69
0.01**
0.73
jou + 
4.78
6.02
2.43
3.01
1.76
0.09
0.48
mic-
1.81
1.59
1.95
2.06
 − 0.25
0.80
 − 0.08
mic + 
1.81
2.15
1.76
1.58
0.10
0.92
0.03
mma
5.22
4.35
5.86
3.86
 − 0.53
0.59
0.16
mp:target
3.67
2.59
3.29
2.22
0.55
0.59
0.16
mu:target
1.19
1.36
0.71
1.06
1.35
0.18
0.38
mus-
0.22
0.58
0.14
0.36
0.58
0.56
0.16
mus + 
0.15
0.36
0.10
0.30
0.55
0.58
0.16
ska
1.93
2.89
1.24
1.41
1.08
0.29
0.29
td
0.48
0.98
0.19
0.51
1.33
0.19
0.36
tm
1.96
3.29
0.90
1.45
1.50
0.14
0.40
TAP: assessment strategies
16.04
8.47
11
7.10
2.24
0.03*
0.64
changing Answer
0.33
0.70
0.57
0.75
 − 1.14
0.26
 − 0.34
guessing
1.41
1.25
0.43
0.75
3.37
0.00**
0.92
match
3.60
3.20
2.19
2.73
1.64
0.11
0.47
reviewing feedback
2.37
2.00
1.43
2.13
1.56
0.13
0.46
ruling out answers
4.56
3.09
3.62
1.91
1.29
0.20
0.35
skipping an item
0.37
0.49
0.24
0.54
0.88
0.39
0.26
testwiseness
0.07
0.27
0.10
0.30
 − 0.25
0.80
 − 0.08
TAP: domain-general strategies
27.07
13.19
18
9.85
2.73
0.01**
0.77
cc
0.07
0.27
0
0
1.44
0.16
0.37
cois
0.74
1.20
0.38
0.67
1.32
0.19
0.36
draw
0.04
0.19
0
0
1
0.33
0.26
fnc
5.33
3.95
4.10
2.59
1.31
0.20
0.36
hsb
0.81
1.21
0.48
1.08
1.02
0.31
0.29
isc
0.56
0.93
0.14
0.48
1.99
0.05
0.54
mem
0.04
0.19
0
0
1
0.33
0.26
pka:content
2.85
1.97
2.29
2.53
0.84
0.40
0.25
pka:course
1.30
2.33
0.24
0.44
2.30
0.03*
0.60
rn
2.33
1.92
2.76
3.35
 − 0.52
0.60
 − 0.16
rr
5.81
4.78
2.71
2.35
2.94
0.01**
0.79
search
0.33
1.36
0.05
0.22
1.07
0.29
0.28
sq
0.78
0.93
0.52
1.12
0.84
0.41
0.25
st
0.85
1.38
0.57
0.98
0.82
0.41
0.23
sum
2.59
2.80
0.95
1.12
2.77
0.01**
0.73
tn
0.85
2.35
1.29
3.39
 − 0.50
0.62
 − 0.15
TAP: domain-specific strategies
             
mps
27.59
8.89
28.05
7.11
 − 0.20
0.84
 − 0.06
Trace: guided reading questions
3.33
1.82
4.14
1.71
 − 1.58
0.12
 − 0.46
grq.dwnld
0.89
0.32
1
0
 − 1.80
0.08
 − 0.46
grq.newSubm
0.56
0.51
0.76
0.44
 − 1.51
0.14
 − 0.43
grq.readAssign
0.78
0.42
0.86
0.48
 − 0.60
0.55
 − 0.18
grq.reviseSubm
0.56
0.51
0.76
0.44
 − 1.51
0.14
 − 0.43
grq.submit
0.56
0.51
0.76
0.44
 − 1.51
0.14
 − 0.43
Trace: in-class tasks
2.89
0.32
2.76
0.70
0.77
0.45
0.24
lc.submrd1
2.81
0.40
2.76
0.70
0.31
0.76
0.10
lc.submrd2
0.07
0.27
0
0
1.44
0.16
0.37
Trace: pre-class tasks: incorrect
             
hw.submincorr
3.11
2.41
1.43
1.78
2.79
0.01**
0.78
Trace: pre-class tasks: correct
             
hw.submcorr
4.30
1.58
2.76
0.70
 − 1.17
0.25
 − 0.33
Total code occurrences
175.44
53.89
144.05
44.04
2.22
0.03*
0.63
Total time duration
4610.33
439.15
4276.81
541.76
2.30
0.03*
0.69
* p < 0.05, ** p < 0.01
PLAN Planning, RECYCLE:TARGET Recycling:[target], SG:TARGET Sub-Goal:[target], TDFN Task Definition, CE-/ + Content Evaluation, EAC-/ + Expectation of Adequacy of Content, EM Emotion Monitoring, FOK-/ + Feeling of Knowing, FOR-/ + Feeling of Recognition, JOC-/ + Judgment of Correctness, JOL-/ + Judgment of Learning, JOU-/ + Judgment of Understanding, MIC-/ + Monitor Information Coherence, MMA Monitoring Math Accuracy, MP:TARGET Monitoring Progress:[target], MU:TARGET Monitoring Understanding:[target], MUS-/ + Monitoring Use of Strategies, SKA Self-Knowledge Activation, TD Task Difficulty, TM Time Monitoring, CC Comparing and Contrasting, COIS Coordinating Informational Sources, DRAW, FNC Forming New Conclusion, HSB Help-Seeking Behavior, ISC Inferring Source Content, MEM Memorization, PKA:CONTENT/COURSE Prior Knowledge Activation of Content or Course, RN Read Notes, RR Re-reading, SEARCH, SQ Self-Questioning, ST Self-Testing, SUM Summarization, TN Taking Notes, MPS Mathematical Problem-Solving
go back to reference Azevedo, R. (2014). Issues in dealing with sequential and temporal characteristics of self- and socially-regulated learning. Metacognition and Learning, 9(2), 217–228. https://doi.org/10.1007/s11409-014-9123-1CrossRef
go back to reference Azevedo, R., & Gašević, D. (2019). Analyzing multimodal multichannel data about self-regulated learning with advanced learning technologies: Issues and challenges. Computers in Human Behavior, 96, 207–210.CrossRef
go back to reference Azevedo, R., Bouchet, F., Duffy, M., Harley, J., Taub, M., Trevors, G., ... & Cerezo, R. (2022). Lessons learned and future directions of MetaTutor: leveraging multichannel data to scaffold self-regulated learning with an intelligent tutoring system. Frontiers in Psychology, 13.
go back to reference Bannert, M., Reimann, P., & Sonnenberg, C. (2014). Process mining techniques for analysing patterns and strategies in students’ self-regulated learning. Metacognition and Learning, 9(2), 161–185. https://doi.org/10.1007/s11409-013-9107-6CrossRef
go back to reference Ben-Eliyahu, A., & Bernacki, M. L. (2015). Addressing complexities in self-regulated learning: A focus on contextual factors, contingencies, and dynamic relations. Metacognition and Learning, 10(1), 1–13. https://doi.org/10.1007/s11409-015-9134-6CrossRef
go back to reference Bernacki, M.L. (2018). Examining the cyclical, loosely sequenced, and contingent features of self-regulated learning: Trace data and their analysis. In D.H. Schunk & J.A. Greene (Eds.), Handbook of Self-Regulated Learning and Performance (pp. 370–387). Routledge.
go back to reference Bonwell, C. C., & Eison, J. A. (1991). Active learning: Creating excitement in the classroom. 1991 ASHE-ERIC highereducation reports. ERIC Clearinghouse on Higher Education, The George Washington University.
go back to reference Eddy, S. L., & Hogan, K. A. (2014). Getting under the hood: How and for whom does increasing course structure work? CBE Life Sciences Education, 13(3), 453–468. https://doi.org/10.1187/cbe.14-03-0050CrossRef
go back to reference Ericsson, K. A., & Simon, H. A. (1998). How to study thinking in everyday life: Contrasting think-aloud protocols with descriptions and explanations of thinking. Mind, Culture, and Activity, 5(3), 178–186.CrossRef
go back to reference Fan, Y., Saint, J., Singh, S., Jovanovic, J., & Gašević, D. (2021). A learning analytic approach to unveiling self-regulatory processes in learning tactics. ACM International Conference Proceeding Series, 184–195. https://doi.org/10.1145/3448139.3448211
go back to reference Fox, M. C., Ericsson, K. A., & Best, R. (2011). Do procedures for verbal reporting of thinking have to be reactive? A meta-analysis and recommendations for best reporting methods. Psychological Bulletin, 137(2), 316.CrossRef
go back to reference Greene, J. A. (2022). What can educational psychology learn from, and contribute to, theory development scholarship? Educational Psychology Review, 34(4), 3011–3035. https://doi.org/10.1007/s10648-022-09682-5CrossRef
go back to reference Greene, J. A., & Azevedo, R. (2007). A theoretical review of Winne and Hadwin’s model of self-regulated learning: New perspectives and directions. Review of Educational Research, 77(3), 334–372. https://doi.org/10.3102/003465430303953CrossRef
go back to reference Greene, J. A., & Azevedo, R. (2009). A macro-level analysis of SRL processes and their relations to the acquisition of a sophisticated mental model of a complex system. Contemporary Educational Psychology, 34(1), 18–29. https://doi.org/10.1016/j.cedpsych.2008.05.006CrossRef
go back to reference Greene, J. A., Deekens, V. M., Copeland, D. Z., & Yu, S. (2018). Capturing and modeling self-regulated learning using think-aloud protocols. In D. H. Schunk & J. A. Greene (Eds.), Handbook of self-regulation of learning and performance (2nd Ed, pp. 323–337). Routledge. https://doi.org/10.4324/9781315697048
go back to reference Greene, J. A., Bernacki, M. L., & Hadwin, A. F. (2024). Self-regulation. In Handbook of educational psychology (pp. 314–334). Routledge.
go back to reference Kizilcec, R. F., Pérez-Sanagustín, M., & Maldonado, J. J. (2017). Self-regulated learning strategies predict learner behavior and goal attainment in Massive Open Online Courses. Computers and Education, 104, 18–33. https://doi.org/10.1016/j.compedu.2016.10.001CrossRef
go back to reference Lombardi, D., Shipley, T. F., Bailey, J. M., Bretones, P. S., Prather, E. E., Ballen, C. J., Knight, J. K., Smith, M. K., Stowe, R. L., Cooper, M. M., Prince, M., Atit, K., Uttal, D. H., LaDue, N. D., McNeal, P. M., Ryker, K., St. John, K., van der Hoeven Kraft, K. J., & Docktor, J. L. (2021). The curious construct of active learning. Psychological Science in the Public Interest, 22(1), 8–43. https://doi.org/10.1177/1529100620973974CrossRef
go back to reference Maldonado-Mahauad, J., Pérez-Sanagustín, M., Kizilcec, R. F., Morales, N., & Munoz-Gama, J. (2018). Mining theory-based patterns from Big data: Identifying self-regulated learning strategies in Massive Open Online Courses. Computers in Human Behavior, 80, 179–196. https://doi.org/10.1016/j.chb.2017.11.011CrossRef
go back to reference Molenaar, I. (2014). Advances in temporal analysis in learning and instruction. Frontline Learning Research, 2(4), 15–24. https://doi.org/10.14786/flr.v2i4.118CrossRef
go back to reference Reimann, P. (2009). Time is precious: Variable- and event-centred approaches to process analysis in CSCL research. International Journal of Computer-Supported Collaborative Learning, 4(3), 239–257. https://doi.org/10.1007/s11412-009-9070-zCrossRef
go back to reference Saint, J., Fan, Y., Gašević, D., & Pardo, A. (2022). Temporally-focused analytics of self-regulated learning: A systematic review of literature. Computers and Education: Artificial Intelligence, 3(December 2021), 100060. https://doi.org/10.1016/j.caeai.2022.100060CrossRef
go back to reference Saint, J., Gaševic, D., Matcha, W., Uzir, N. A. A., & Pardo, A. (2020). Combining analytic methods to unlock sequential and temporal patterns of self-regulated learning. ACM International Conference Proceeding Series, 402–411. https://doi.org/10.1145/3375462.3375487
go back to reference Schunk, D. H., & Greene, J. A. (2018). Historical, contemporary, and future perspectives on self-regulated learning and performance. In D. H. Schunk & B. Zimmerman (Eds.), Handbook of Self-Regulation of Learning and Performance (2nd Ed, pp. 1–15). Routledge.
go back to reference Shaffer, D. W., Collier, W., & Ruis, A. R. (2016). A tutorial on epistemic network analysis: Analyzing the structure of connections in cognitive, social, and interaction data. Journal of Learning Analytics, 3(3), 9–45.CrossRef
go back to reference Shaffer, D. W. (2017). Quantitative ethnography. Madison, WI: Cathcart Press
go back to reference Siebert-Evenstone, A. L., Collier, W., Ruis, A. R., & Shaffer, D. W. (2017). In search of conversational grain size: Modelling semantic structure using moving Stanza Windows. Journal of Learning Analytics, 4(3), 123–139.CrossRef
go back to reference Sung, H., Swart, M. I., & Nathan, M. J. (2022). Methods for analyzing temporally entangled multimodal data. In A. Weinberger, W. Chen, D. Hernández-Leo, & B. Chen (Eds.), Proceedings of the 15th International Conference on Computer Supported Collaborative Learning - CSCL 2022 (pp. 242–249). https://doi.org/10.22318/cscl2022.242
go back to reference Sung, H., & Nathan, M. J. (2024). Your body tells how you engage in collaboration: Machine-detected body movements as indicators of engagement in collaborative math knowledge building. British Journal of Educational Technology, 55(5), 1950–1973. https://doi.org/10.1111/bjet.13473CrossRef
go back to reference Tan, Y., Ruis, A. R., Marquart, C., Cai, Z., Knowles, M. A., & Shaffer, D. W. (2022). Ordered network analysis. International Conference on Quantitative Ethnography, 101–116. https://doi.org/10.1007/978-3-031-31726-2_8
go back to reference Taub, M., Azevedo, R., Bouchet, F., & Khosravifar, B. (2014). Can the use of cognitive and metacognitive self-regulated learning strategies be predicted by learners’ levels of prior knowledge in hypermedia-learning environments? Computers in Human Behavior, 39, 356–367. https://doi.org/10.1016/j.chb.2014.07.018CrossRef
go back to reference Uzir, N. A. A., Gaševic, D., Jovanovic, J., Matcha, W., Lim, L. A., & Fudge, A. (2020). Analytics of time management and learning strategies for effective online learning in blended environments. ACM International Conference Proceeding Series, 392–401. https://doi.org/10.1145/3375462.3375493
go back to reference Winne, P. H., & Hadwin, A. F. (1998). Studying as self-regulated learning. In D. J. Hacker, J. Dunlosky, & A. C. Graesser (Eds.), Metacognition in educational theory and practice (pp. 277–304). Lawrence Erlbaum Associates.
go back to reference Winters, F. I., Greene, J. A., & Costich, C. M. (2008). Self-regulation of learning within computer-based learning environments: A critical analysis. Educational Psychology Review, 20(4), 429–444. https://doi.org/10.1007/s10648-008-9080-9CrossRef
go back to reference Xu, Z., Zhao, Y., Liew, J., Zhou, X., & Kogut, A. (2023). Synthesizing research evidence on self-regulated learning and academic achievement in online and blended learning environments: A scoping review. Educational Research Review, 39(December 2022), 100510. https://doi.org/10.1016/j.edurev.2023.100510CrossRef
go back to reference Zimmerman, B. J. (2002). Becoming a self-regulated learner: An overview. Theory into Practice, 41(2), 64–70. https://doi.org/10.1207/s15430421tip4102_2CrossRef
Image Credits
in-adhesives, MKVS, Ecoclean/© Ecoclean, Hellmich GmbH/© Hellmich GmbH, Krahn Ceramics/© Krahn Ceramics, Kisling AG/© Kisling AG, ECHTERHAGE HOLDING GMBH&CO.KG - VSE, Schenker Hydraulik AG/© Schenker Hydraulik AG