Files
opaque-lattice/papers_txt/How-AI-agents-transform-reflective-practices--A-three-se_2026_Computer-Stand.txt
2026-01-06 12:49:26 -07:00

897 lines
112 KiB
Plaintext
Raw Permalink Blame History

This file contains invisible Unicode characters
This file contains invisible Unicode characters that are indistinguishable to humans but may be processed differently by a computer. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
Computer Standards & Interfaces 97 (2026) 104094
Contents lists available at ScienceDirect
Computer Standards & Interfaces
journal homepage: www.elsevier.com/locate/csi
How AI agents transform reflective practices: A three-semester comparative
study in socially shared regulation of learning
Yumin Zheng a, Fengjiao Tu b , Fengfang Shu a,c , Chaowang Shang a,* , Lulu Chen a , Jiang Meng a
a
Faculty of Artificial Intelligence in Education, Central China Normal University, Wuhan 430079, China
b
Department of Information Science, University of North Texas, 3940 North Elm, Denton, Texas, 76203, USA
c
Institute of Open Education, Wuhan Vocational College of Software and Engineering, Wuhan Open University, Wuhan, China
A R T I C L E I N F O A B S T R A C T
Keywords: High-quality reflection has been a challenging barrier in the socially shared regulation of learning (SSRL).
Artificial intelligence agent Especially with the emergence of generative artificial intelligence (GAI), traditional methods such as reflection
Socially shared regulation of learning reports may increase the students risk of superficial reflection. This study uses an artificial intelligence agent (AI
Reflection quality
agent) to design a reflection assistant, which aims to enhance students reflection ability through continuous
Collaborative learning
Generative artificial intelligence
questioning and real-time, content-specific feedback based on their written reflections. Through a comparative
experiment conducted over three semesters, this study demonstrates the different impacts of three reflection
methods, reflection reports, reflection short-answer questions, and AI agents, on the quality of university stu­
dents reflections. The results indicate that there is a significant difference in the quality of reflection among the
three reflection methods. Students using AI agents show the highest levels of reflection, characterized primarily
by connective reflection and critical reflection. Epistemic network analysis further reveals that the AI agent
reflection method is more effective in improving the reflection quality of low-performance teams than that of
high-performance teams. This expands AI agents use in SSRL reflection, introduces new methods for the GAI era,
and provides practical experience and reflection intervention strategies for teachers and instructional designers
in SSRL.
1. Introduction Nowadays, these traditional methods fall short of addressing the chal­
lenges posed by GAI [9]. Students may easily rely on tools like ChatGPT
With the rapid advancement of generative artificial intelligence to complete short-answer questions, journals, and reports. Kiy [10] has
(GAI), numerous challenges in collaborative learning have been shown that 76 % of university students use ChatGPT for their assign­
addressed with innovative solutions [1,2]. GAI applications, represented ments, with the percentage being even higher among software engi­
by artificial intelligence agents (AI agents), have introduced revolu­ neering students, reaching 93 % [11]. The widespread use of GAI has
tionary transformations to education. These transformations are mainly profoundly transformed traditional methods of learning and teaching,
due to the powerful expert-level conversational abilities and and this era calls for new approaches to reflection.
user-friendly accessibility [3]. AI agents are computing systems with capabilities for autonomous
The socially shared regulation of learning (SSRL) strategy serves as a perception, decision making, and action [12]. They use GAI to learn,
crucial mechanism for enhancing learning outcomes in collaborative reason, and perform corresponding tasks or actions from the surround­
learning [4]. Through the SSRL strategy, learners collaboratively set ing environment and input information. To enable practical imple­
goals and monitor progress, thereby improving their performance [5]. mentation, rule-based AI agents have been developed that require no
Reflection is a critical component of SSRL, aiding learners in recognizing programming and can be deployed simply by defining task objectives
and refining their learning processes [6]. However, achieving and roles via prompts. In educational contexts, these rule-based AI
high-quality reflection remains a challenge [7]. agents are commonly used for personalized instruction and intelligent
There are various methods to enhance reflection quality in SSRL, tutoring due to their ability to engage in real-time dialogue and provide
such as providing prompts and templates in reflection reports [8]. immediate feedback [13].
* Corresponding author.
E-mail address: phdzhengyumin@mails.ccnu.edu.cn (C. Shang).
https://doi.org/10.1016/j.csi.2025.104094
Received 1 February 2025; Received in revised form 28 October 2025; Accepted 10 November 2025
Available online 11 November 2025
0920-5489/© 2025 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
The rule-based AI agent provides an effective approach for sup­ widely applied in education [16]. It can support collaborative learning
porting SSRL reflection. Instructors can set specific SSRL task directions, through personalized instruction, real-time feedback, and intelligent
and the agent guides students based on the reflection checklist while assessment [17]. AI agents, a form of GAI equipped with autonomous
adaptively generating questions according to students responses. Each learning and decision-making capacities, have emerged as key instruc­
follow-up question is dynamically generated based on the students tional tools in global educational research.
prior answers and the specific SSRL task, making it difficult for students Empirical studies have shown that AI agents significantly improve
to rely on external AI tools like ChatGPT to provide generic responses. student engagement [18,19], learning motivation [20,21], and aca­
This continuous dialogue mechanism supports deeper, more analytical demic performance [22]. AI agents exist in various forms, such as
reflection and reduces the risk of superficial reflection [14]. Despite AI chatbots [23], intelligent tutoring systems (ITS; [24]), embodied
agents having broad application prospects, current research on conversational agents (ECA; [25,26]), and intelligent virtual assistants
improving learners reflection quality by AI agents remains limited and (IVA; [13,27]). Among these, GAI-based chatbots have been widely
requires further in-depth exploration. adopted in education due to their customizable roles and flexible
Against this backdrop, this study introduces a rule-based AI agent deployment. The present study focuses on this type of conversational AI
reflection assistant within the SSRL framework to help learners enhance agent.
their reflection quality. This study aims to examine the impact of the AI In higher education, AI agents have been shown to support higher-
agent on SSRL reflection quality by comparing three reflection methods: order thinking skills, such as critical thinking, metacognition, and
reflection reports, short-answer reflection questions, and the AI agent- problem-solving [23,28,29]. In these studies, GAI was embedded within
based reflection. In addition, different methods may lead to different structured reflection activities, allowing students to engage in guided
reflection qualities among learners in high and low-performance teams reflective processes targeting specific cognitive skills. For example,
[15]. Therefore, we further explored the differences in reflection quality Hong et al. [29] employed AI to handle lower-level tasks in essay
between high and low-performance teams when using these three writing, enabling students to focus on evaluation and reflection, thereby
reflection methods. We proposed the following research questions: enhancing critical thinking. Chen et al. [28] implemented metacognitive
strategy-supported AI agents that prompted process-oriented reflection
RQ1: How does the AI agent reflection assistant affect learners and multi-perspective discussion, improving metacognitive skills. Zhou
reflection quality in SSRL? et al. [23] situated reflection within a self-regulated learning frame­
RQ2: What differences do high and low-performance teams show in work, showing that GAI-supported reflection indirectly benefits critical
reflection quality when using the three reflection methods? thinking and problem-solving.
Although these studies demonstrate that AI agents can enhance
This study conducted a three-semester comparative teaching exper­ higher-order thinking, reflection itself has often been treated merely as a
iment to evaluate the impact of AI agents and two traditional reflection learning process rather than a measurable skill. Reflection is a core
methods (reflection reports and short-answer questions) on university component of higher-order thinking and an essential learning compe­
students reflection quality. Using statistical analysis, content analysis, tency for 21st-century university students. Empirical evidence directly
and epistemic network analysis (ENA), this study examines the effec­ examining the impact of AI agents on learners reflective abilities,
tiveness of AI agents in enhancing university students reflection quality particularly in collaborative learning environments, remains scarce.
in SSRL. Investigating this relationship is therefore necessary to understand how
The main contributions of this study are summarized as follows: AI agents can effectively support the development of reflection.
- We introduce a practical SSRL activity, providing educators with a 2.2. Socially shared regulation of learning and reflection
valuable instructional framework for facilitating collaborative
learning. Collaborative learning includes three primary types of regulation:
- We integrated an AI agent reflection assistant in SSRL and provided a self-regulation (SR), co-regulation (CoR), and socially shared regulation
comprehensive debugging process, offering instructors examples and (SSR) [30,31]. Based on SSR theory, socially shared regulation of
considerations of AI agent implementation. learning (SSRL) is an emerging collaborative learning strategy empha­
- We revealed the reflection quality differences between high and low- sizing mutual support and feedback among team members. The strategy
performance teams in various reflection approaches and demon­ consists of four key stages: goal setting, task distribution, progress
strated the advantages of the AI agent for low-performance teams. monitoring, and reflection evaluation [3235]. Research indicates that
the SSRL strategy has a positive impact on collaborative learning [36].
The research is organized as follows: Section 2 reviews prior research Learners may enhance their awareness of the collaborative process and
on AI agents in education, SSRL theory, and reflection. Section 3 de­ facilitate the activation of regulatory processes through SSRL [4]. And
scribes the participants, research design, and methods for data collection SSRL helps to enhance learners cognitive and metacognitive abilities,
and analysis. Section 4 compares reflection quality across the three boosting learning motivation and engagement [37,38]. Additionally,
methods and examines differences between high and low-performance SSRL fosters communication among team members, improving collab­
teams using ENA. Section 5 discusses the results and implications. The orative efficiency [39]. Thus, SSRL has been widely incorporated into
paper concludes with a summary and potential directions for future collaborative learning and plays a significant role in enhancing various
research. learner abilities.
Reflection quality is a key indicator for assessing the success of SSRL
2. Literature review [39]. High-quality reflection is an indispensable component of SSRL, as
it enables learners to examine and evaluate their learning processes and
To explore the impact of AI agents on learning processes, it is outcomes [40]. Unlike conventional collaborative learning, the reflec­
essential to examine their application in education, followed by a dis­ tion content in SSRL emphasizes the process of mutual regulation and
cussion on SSRL and reflection. monitoring among group members. However, since reflection is the final
stage of SSRL, educators often overlook its significance [41]. Teachers
2.1. AI agents in teaching lack of emphasis on the reflection stage may lead to low-quality
reflection among students [42]. Achieving high-quality SSRL reflection
Generative Artificial Intelligence (GAI), defined as AI systems remains a persistent challenge for educators and students [43].
capable of autonomous learning and content generation, has been To enhance students reflective abilities, it is essential to focus on the
2
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
definition of reflection. Dewey [44] defined reflection as a continuous elaborated on the activities of SSRL and the design process of the AI
process of exploring and evaluating experiences, which helps in­ agent. Lastly, we discussed the coding scheme for reflection quality and
dividuals gain a deeper understanding of their behaviors and outcomes. provided the methodology for data collection and analysis.
Zimmerman [45] further emphasized that self-reflection is a complex
learning process involving various aspects of self-monitoring, such as 3.1. Participants
self-assessment and feedback on contributions. In the theory of SSRL,
reflection encompasses not only self-assessment but also shared moni­ The participants were from the course “Internet Thinking and Digital
toring processes with others [39]. These theories provide support for Self-Learning” over three semesters: Spring 2023, Fall 2023, and Spring
exploring and promoting the reflective process. 2024. A total of 97 undergraduate students, aged 18 to 22, took part in
In reflective activities, teachers can support students deep learning this study (Table 1).
and reflective abilities through various intervention strategies, such as At the beginning of each semester, students completed a pre-test
scaffolding, reflective prompts, and feedback [46]. Reflective scaf­ using the CThQ [63], which assesses six cognitive dimensions: mem­
folding involves providing structured guidance to help students more ory, comprehension, application, analysis, evaluation, and creation
effectively review and analyze their learning experiences [47]. When (overall reliabilityα= 0.87). According to Dewey [64], critical thinking
designing reflection tasks for SSRL, teachers often utilize the SSRL is a deepening and extension of reflective thinking, with high consis­
reflection scaffolds developed by Panadero et al. [48]. Additionally, tency in cognitive processing, reasoning, and evidence evaluation. The
reflective prompts and guiding questions steer students toward specific CThQ pre-test provides a valid proxy for students baseline reflection
directions for reflection, assisting them in identifying potential barriers levels. One-way ANOVA indicated no significant differences in pre-test
and challenges in their learning [49]. Feedback provides learners with total scores among the three groups (Group 1: M = 105.07, SD =
suggestions or information to improve task performance, helping them 6.13; Group 2: M = 103.72, SD = 4.19; Group 3: M = 105.22, SD =
optimize both their reflection and learning processes [50]. From a 4.24), F(2, 86) = 1.33, p = 0.27, suggesting comparable reflection
cognitive perspective, feedback serves as guidance to enhance students abilities across groups prior to the intervention.
task performance [51]. Timely feedback on students reflections not Participants were divided into 3 groups, each employing a different
only improves the quality of subsequent reflections but also deepens reflection method, and within each group, students were further divided
their understanding of reflective concepts [52]. into teams using random assignment to minimize potential biases arising
Reflection journals, reflection reports, and reflection short-answer from prior academic performance, familiarity, or interpersonal prefer­
questions have been explored to improve reflection quality [53,54]. ence. Random assignment was chosen over self-selection or instructor-
However, the traditional methods may not adapt to the advancements of based grouping to ensure group equivalence and to enhance the inter­
GAI. These require students to submit longer texts, which inevitably nal validity of the comparative analysis [65].
causes a risk of superficial reflections due to the use of GAI. Some The first group (G1), consisting of 31 students from the Spring 2023
scholars have also modified reflection methods from a technological semester, conducted reflection reports and were further divided into 7
perspective by using various reflection platforms, such as Google Docs teams. The second group (G2), consisting of 30 students from the Fall
[55], Flipgrid [56], the VEO app [57], and Wiki [58]. However, these 2023 semester, conducted short-answer reflections and were divided
platforms primarily offer static or limited interaction, which constrains into 7 teams. The third group (G3), consisting of 36 students from the
students ability to adaptively engage in reflective processes. The Spring 2024 semester, conducted reflections through continuous ques­
low-quality reflection issues in SSRL urgently require new solutions. tioning by an AI agent and were divided into 9 teams. Additional in­
Although GAI poses challenges to traditional reflection methods, it formation about the participants is provided in Table 1.
also offers new solutions. AI agents are increasingly regarded as effective
tools for supporting reflection practices. Research indicates that the use 3.2. Design of socially shared regulation of learning activities
of AI agents in reflection activities may enhance students learning
motivation and engagement [59]. Teachers can use AI agents to design During the 4-week activity, students collaborated in teams to pro­
reflection scaffolding, assisting learners in conducting more in-depth duce micro-lesson videos lasting 5 to 8 min. The activity was divided
and systematic reflections [60]. In addition, AI agents may enhance into 4 stages, each lasting one week (Table 2).
reflection quality through data analysis and intelligent feedback [61]. In the first week (goal setting), students were required to establish a
Therefore, AI agents demonstrate potential in addressing the issue of common goal, select the videos theme, and outline the content frame­
improving SSRL reflection quality. work. Then, they submitted a project proposal detailing the topic, ob­
Thus, this study designed a reflection assistant by AI agents to jectives, task distribution, and timeline. In the second week (task
enhance university students reflection quality in SSRL. Statistical distribution), the teams followed their project plan to allocate tasks and
analysis, content analysis, and ENA were employed to collect and begin executing the project. The instructor provided guidance and
analyze textual data related to reflection quality. By comparing the AI suggestions throughout this process. In the third week (progress moni­
agent reflection assistant with traditional SSRL strategy reflection scaf­ toring), each team submitted a video sample that was between 1 and 2
folding, this study analyzed the differences in reflection content and min long. The instructor conducted an initial evaluation based on the
reflection levels among university students across three methods. sample and suggested improvement. Students refined and adjusted their
Additionally, previous research suggests that high and low-performance video production based on the feedback. In the fourth week (reflection
teams may experience different effects from various reflection methods evaluation), students submitted their completed micro-lesson videos
[62]. Therefore, this study further explores the differences between high
and low-performance teams when using three reflection methods. This Table 1
study provides new theoretical evidence for using AI agents in SSRL Participant and group information.
reflection practices.
Group Course Reflection Team Participant Female Male
method
3. Methodology
G1 Spring Report 7 31 17 14
2023
This study employed a quasi-experiment to explore the differences G2 Fall 2023 Short-answer 7 30 19 11
among three reflection methods in SSRL. And examine whether AI questions
agents improve the reflection quality of university students. Firstly, we G3 Spring AI reflection 9 36 20 16
2024 assistant
provided information about the participants and the course. Then, we
3
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
Table 2 AI agent reflection assistant, Crystal, was developed using the Coze
The stages of SSRL. platform (https://www.coze.cn/). The AI agent consists of 4 core com­
Week SSRL stages Description ponents, with Part A being the AI agents name, Part B defining the role
setting and response logic, Part C specifying the conversational experi­
1 Goal setting Students discuss the goal, theme, and framework.
2 Task distribution Students allocate tasks and make the micro lesson ence, such as the opening dialogue, and Part D serving as the preview
videos. interface. Developing the AI agent requires following these operational
3 Progress Students monitor the task and submit a video sample. steps.
monitoring
4 Reflection Students submit completed micro-lesson videos and
evaluation individual reflection assignments.
Step 1: Create the AI agent and assign it to the name Crystal (as
shown in Fig. 1, Part A). Define it as the reflection assistant for the
course “Internet Thinking and Digital Self-Learning. Set its duty to
and individual reflection assignments (employing different reflection guide students in completing tasks (as shown in Fig. 1 Part B) and
methods for each of the three semesters). Finally, a reflection-sharing design the opening statement (as shown in Fig. 1 Part C).
session was held in class, where students exchanged learning experi­ Step 2: Set up the reflection task (as shown in Fig. 1, Part B). Input all
ences and insights. the questions from the SSRL reflection scaffolding developed by
Panadero et al. [48] into the AI agents as the question base. This
ensures a logical flow of questions from the AI agent to the students,
3.3. Design of the three reflection methods preventing task misdirection. In addition, the AI agent was not
restricted to this fixed list but generated follow-up questions,
Prior to the reflection phase, all students completed a four-week particularly “Why” questions, based on the students specific an­
SSRL activity in which the instructor introduced and practiced the swers, which reflected its adaptiveness.
four SSRL stages. Consequently, all reflections were anchored in the Step 3: Set up the response rule (as shown in Fig. 1, Part B). Establish
teams performance across these four stages. In G1, the reflection the response rules for the AI agent:
remained open-ended within this framework and only specified a min­ a. Ask only one reflection question per interaction.
imum length of at least 200 words (no SSRL question list was provided). b. Provide encouraging feedback that adapts dynamically after each
In G2, students conducted individual reflections through short- response (e.g., “You did a great job”, “Your reflection is very
answer questions. The guiding questions were derived from the SSRL insightful”).
reflection scaffolding [48]. For example, questions included “What is the c. Avoid using academic terms.
groups current assignment?” and “What obstacles might the group d. Use only special interrogative questions (e.g., “What, ” “Why”),
encounter?” with follow-up questions adjusted according to students responses.
G3 students used the AI agent reflection assistant for their re­ e. After answering all questions, conclude the conversation and ex­
flections. After the SSRL task, the instructor provided students with a press gratitude.
quick response code (QR code) linking to the AI agents website. Stu­ Step 4: Testing and deployment (as shown in Fig. 1, Part D). Check
dents scanned the QR code with their phones to initiate a conversation the conversation flow and ensure the AI agents smooth and effective
with the AI agent. Each student completed the reflection task through interactions. Select 5 students for a second round of testing to ensure
the dialogue.
The development process of the AI agent is illustrated in Fig. 1. The
Fig. 1. AI agent development interface on the Coze platform.
4
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
the conversation flows smoothly. Once confirmed, the AI agent can Table 3
be deployed and available to all students. Learner reflection quality coding scheme.
Categories Coding Description
3.4. Experimental procedure Reflection NOR Lacking a reflection mindset.
level LOWR Having a reflective mindset involves reviewing
experiences, describing facts and feelings, and
The experimental procedure is illustrated in Fig. 2. As described in
reflecting on what has been learned. It also
the Participants section, all students completed the CThQ [63] as a encompasses the ability to connect new knowledge
pre-test before the course. They then attended a 16-week course with existing knowledge and to improve learning
covering basic concepts. All students were taught by the same instructor, strategies.
with the course content, teaching methods, and learning resources HIGHR Critically analyzing the current situation, attempting to
view problems from different perspectives, forming
remaining entirely consistent across the three semesters. Students new viewpoints from available resources, and seeking
participated in a 4-week group collaboration activity, “creating micro to test hypotheses.
lesson videos”, conducted using the SSRL strategy. After the group ac­ Reflection DESR A description of “what” the object of reflection is.
tivity finished, each student was assigned an individual reflection task. content EXPR An explanation of the causes behind the object of
reflection, addressing the “why” often indicated by
G1 and G2 used traditional reflection methods, with G1 completing
keywords such as “in order to”, "due to", or "so as to".
reflection reports and G2 answering short-answer questions. G3 CONR Understanding whether the object of reflection has
employed a new reflection method, utilizing the AI agent reflection changed across different times and contexts, coupled
assistant. with an analysis of the reasons for these changes and
their impact on behavior, represents a higher level of
analysis concerning the “what” and “why”.
3.5. Data collection and analysis CRIR It identifies personal or team issues and analyzes them
with theory and practice to solve problems, focusing on
“how” to achieve self-reconstruction. This may include
After the three semesters, the reflection texts of all students were keywords like “needs improvement” or “next stage”.
collected and anonymized. G1 produced 31 reflection reports totaling
8032 words. G2 submitted 30 reflection short-answer texts, totaling
15,468 words. G3s AI agent reflection assistant dialogues comprised 36 To ensure reliability, a coding discussion group comprised two ex­
submissions, totaling 16,801 words (excluding the AI agents questions). perts and two professional coders. First, the two coders preliminarily
Content analysis was used to process the reflection texts. Through coded the first 10 % of the reflection texts. In cases of disagreement, they
systematic coding rules, this method reduced the influence of subjective consulted with the experts to reach a consensus. After training and
judgment and personal bias, thereby providing more objective results. repeated practice, the coders achieved a high level of consistency. The
The coding scheme consists of two parts: reflection level and reflection coders strictly adhered to the revised coding scheme during the formal
content, as shown in Table 3. The reflection level coding scheme is based coding process. After coding, inter-coder reliability was calculated,
on Plack et al. [66], and it is used to assess the overall reflection level of yielding a Cohens kappa coefficient of 0.87, indicating that the coding
learners, categorized into no reflection (NOR), low reflection (LOWR), process had a high level of reliability. The coders consulted with experts
and high reflection (HIGHR). The reflection content coding scheme is for different coding results and ultimately reached an agreement.
based on Wang et al. [67] and is used to explore the differences in the After coding the reflection texts using the content analysis, ENA was
types of learners reflection content. The reflection content is catego­ employed to conduct a fine-grained analysis of the reflection data.
rized into 4 types: descriptive reflection (DESR), explanatory reflection Content analysis excels at systematically and objectively analyzing large
(EXPR), connected reflection (CONR), and critical reflection (CRIR), volumes of textual content. ENA focuses on uncovering the complex
with reflection quality progressively increasing across these categories. relational networks between elements, such as reflection levels. The
The reflection texts in the reflection reports and short-answer re­ combination of the two methods allows for attention to both the char­
flections were relatively longer, while those in the AI agent dialogues acteristics of the text itself and the internal relationships between the
were shorter. To mitigate the differences caused by these length dis­ content elements. Additionally, the ENA Webkit (http://www.epist
crepancies, this study used a single complete sentence as the minimum emicnetwork.org/) provides a stable environment for data analysis.
coding unit. For example, the statement “As the group leader, I am quite To investigate the differences in reflection quality between the high
decisive. I directly assigned tasks to everyone, and the group was sup­ and low-performance teams, we assessed the micro lesson videos
portive.” should be coded as two separate sentences. completed by students in SSRL. The videos were assessed by two experts
Fig. 2. Experimental procedure.
5
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
in education, each with over 10 years of teaching experience. The Table 5
evaluation criteria included the following categories, with topic selec­ The result of the Kruskal-Wallis H test.
tion worth 10 points, instructional design 40 points, content complete­ Codes Mean score χ² p
ness 20 points, audio-visual quality 20 points, and artistry 10 points.
G1 G2 G3
Each group received a score ranging from 0 to 100 points. The two ex­
perts thoroughly discussed the evaluation criteria to ensure consistency Reflection quality NOR 0.018 0.005 0.088 6.557 0.038
LOWR 0.267 0.163 0.232
in scoring and then individually assessed all instructional designs and
HIGHR 0.018 0.044 0.218
materials. The scoring consistency between the two experts (Spearman DESR 0.229 0.197 0.262
correlation coefficient) was 0.86 (p < 0.01). EXPR 0.100 0.103 0.264
The average score from both experts was used as the final score for CONR 0.038 0.028 0.221
CRIR 0.006 0.037 0.200
each group (Table 4). The grouping criteria for high and low performing
teams proposed by Hou [68] have been widely adopted by scholars [69].
In this study, based on those criteria, the top 15 % of teams were clas­ significance of 0.038. The mean ranks for the 3 groups were G1 = 9.14,
sified as the high-performance teams, including G1-team7, G2-team2, G2 = 8.00, and G3 = 15.86. The results indicate a statistically significant
and G3-team1. The bottom 15 % of teams were classified as the difference in reflection scores between the groups (p = 0.038). Specif­
low-performance teams, including G1-team5, G2-team6, and G3-team4. ically, G3s mean rank was significantly higher than G1 and G2, indi­
Using ENA, we further explored the differences between the high and cating that using the AI agent is associated with higher performance.
low-performance teams of students. To further investigate the observed differences, we applied ENA for a
fine-grained analysis of the students reflections across the 3 reflection
3.6. IRB approval and AI agent data privacy methods. This analysis aims to uncover the epistemic structures and
patterns, providing deeper insights into how different reflection
This study has received approval from the Institutional Review Board methods influence the quality and complexity of students reflection
(IRB) of the university, ensuring that all ethical standards are met. All processes. By analyzing epistemic networks, we may better understand
students participated voluntarily, fully aware of the studys purpose and the specific epistemic factors and relationships underlying the differ­
procedures, and signed informed consent forms prior to the ences observed in the statistical results.
commencement of the experiment. In addition, to protect participants Fig. 3 presents a comparative ENA network model of reflection
privacy, all data collected during the study were anonymized. content for the three groups using different reflection methods. In this
All conversations on the Coze platform were fully anonymized, and model, nodes represent individual reflection codes, and edges indicate
students were reminded before using the platform not to enter any the co-occurrence of codes within each unit of analysis. Blue, red, and
personal or sensitive information (such as name, student ID, gender, or purple dots denote the centroids of students in G1, G2, and G3,
school). Data was labeled only with class sequence numbers (e.g., Stu­ respectively, while the four black dots represent the four categories of
dent 1, Student 2), and access was strictly limited to the research team. reflection content (DESR, EXPR, CRIR, CONR). ENA applies singular
In addition, all students signed the Coze platforms privacy protection value decomposition (SVD) to reduce the network model to two di­
agreement, and the platform further ensures data security through mensions, which together account for 70.1 % of the variance (SVD1 =
anonymization and encryption techniques. 51.5 %, SVD2 = 18.6 %). The x-axis in the ENA space (SVD1) defines the
dimension of reflection content, with the right side (higher x-values)
4. Results representing DESR codes and the left side (lower x-values) representing
CONR codes. The y-axis (SVD2) in the ENA space defines the dimension
The results are organized to address the key research questions of reflection content, where the CRIR and EXPR codes are positioned
regarding the effectiveness of the AI agent and the differences in higher (with higher y-values). The DESR code is located lower in the
reflection quality across various reflection methods. ENA space (with lower y-values). This model allows comparison across
students and groups, showing which types of reflection are more
4.1. How does the AI agent reflection assistant affect learners reflection dominant and how reflection content patterns differ between groups.
quality in SSRL? The right side of Fig. 3 displays the mean networks of the 3 groups.
Overall, the reflection content of all 3 groups predominantly features
A Kruskal-Wallis H test was conducted to assess the differences in EXPR and DESR, with a strong association observed between these two
SSRL reflection scores among the 3 groups of students using different points. The reflection content network of G1 is the sparsest, with only a
reflection methods, as shown in Table 5. The test compares independent few occurrences of CRIR, aside from the relatively frequent appearances
samples without assuming a normal data distribution. This makes it of EXPR and DESR. The network of G2 is more concentrated, with dis­
highly suitable for analyzing the multiple groups of non-normally tribution across all 4 reflection types and a stronger CRIR-DESR
distributed reflection data in this study. connection (value of 0.10). The reflection content of G3 is the most
For this analysis, an overall reflection quality score was calculated densely connected, with all 4 types having a relatively high proportion
for each student by taking the meaning of all seven reflection codes of representation. The CRIR-CONR (0.23) and CONR-EXPR (0.13) con­
(NOR, LOWR, HIGHR, DESR, EXPR, CONR, CRIR). This composite score nections are relatively strong. In contrast, the other pairs based on
was used for the Kruskal-Wallis H test, while the mean scores for indi­ traditional SSRL reflection did not exhibit strong correlations.
vidual codes presented in Table 5 are provided only for descriptive Table 6 demonstrates how the AI agent, through guided dialogue,
purposes. facilitated the transition of G3 students from connective reflection
The results showed a chi-square value of 6.557, and an asymptotic (CONR) to critical reflection (CRIR), thereby deepening the SSRL
Table 4
Scores of the SSRL performance for the 3 groups.
Group team1 team2 team3 team4 team5 team6 team7 team8 team9
G1 86.0 90.0 88.5 76.5 68.5 87.0 92.0 NA NA
G2 83.5 93.5 87.0 90.5 81.0 71.5 76.5 NA NA
G3 94.0 84.5 75.0 71.5 88.0 76.0 89.5 84.5 90.5
6
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
Fig. 3. Comparison of reflection content.
then proposed a concrete plan for deeper learning of AI technology
Table 6
(CONR). The AI agents continuous questioning and feedback prompted
Transcribed example of the G3 reflection content.
students to progress from simple descriptive reflection to more critical
Student Transcript and in-depth reflection throughout the reflection process.
1 AI agent: What problems did you encounter, and how did you solve them? Table 7 presents reflection examples from some G1 and G2 students,
Student 9: We initially struggled, but by discussing with our team and highlighting the impact of different reflection forms and guidance
leader, we solved the problems and saw how teamwork improved our
process over time. (CONR)
AI agent: What are your plans moving forward?
Table 7
Student 9: I hope for more team projects to use everyones strengths and
improve how I contribute next time. (CRIR)
Transcript example of the G1 and G2 reflection content.
2 AI agent: What were the shortcomings of this micro-lecture task? Group Transcript
Student 18: Our PPT didnt cover everything, so next time Ill make it
more complete. (CRIR) G1 G1-Student 4: Our group chose a radio show format for this Himalaya
AI agent: I see, thats a very insightful analysis. What plans do you have for assignment. (DESR) Ive always been a fan of radio shows, so I was very
your future studies? happy to have the opportunity to create one this time. (DESR) Of course, I
Student 18: I plan to learn more about AI video production to improve my also faced some challenges during the production process (DESR), such as
skills for future jobs. (CONR) the tone not fitting the storyline and the quality of the program needing to
3 AI agent: Did you encounter any difficulties and challenges while be better. (EXPR)
completing the task? If so, what were they, and how did you solve them? G1 G1-Student 30: Regarding this task, firstly, we didnt do well in the
Student 31: Our biggest challenge was managing time, but adjusting our presentation aspect. The presentation was only in the form of a document,
schedule helped us complete the tasks. (CONR) which needed to ensure a smooth connection between the presentation and
AI agent: Great. How did you overcome these difficulties and challenges? the work, making it difficult to access the content. (CONR) Secondly, the
Student 31: We held weekly meetings to discuss. (DESR) content presentation was poorly executed and lacked a logical structure.
4 AI agent: Please summarize the successful aspects of this task. (EXPR) Finally, the speech was not coherent during the presentation, and
Student 36: The tasks were well-organized, and because our team the preparation was insufficient. (EXPR)
cooperated closely, we were able to complete the work more efficiently G2 G2-Student 6:
than at the beginning. (CONR) Task: We approached the task mainly in two aspects. (DESR) The first part
determined the theme and type of work, and the second part recorded the
work. (DESR)
Division of labor: Our division of labor and cooperation were very
reflection process. Under the guidance of the AI agent, student 9 and
reasonable, and each member completed their assigned tasks. (EXPR)
student 31 shifted from describing the current state of teamwork and Self-evaluation: Very successful. (DESR)
time management, such as “We solved problems through communica­ Outlook: We plan to work more collaboratively on each task and strive to do
tion with team members” (CONR), to deeper reflections on self- our best. (CRIR)
improvement and future learning plans, exemplified by “I hope for G2 G2-Student 27:
Task: This task enhanced our understanding of content production and
more team projects to utilize everyones potential” (CRIR). Prompted by strengthened the collaboration among team members. (EXPR)
the AI agents questioning, student 10 and student 36 reflected on the Division of labor: Our team had a clear division of responsibilities, and
shortcomings of the SSRL tasks, noting that “The resources were not everyone had their tasks (EXPR). I was responsible for the recording, which
comprehensive, and most content lacked innovation” (CONR), and was quite challenging. (EXPR)
Self-evaluation: Although our team may not have been the best among all
further analyzed the root causes of these issues, along with potential
the teams, we had unique messages to convey. (CONR) If there is a next
improvement measures (CRIR). Inspired by the AI agent, student 18 first time, we will strive to improve it. (CRIR)
identified the issue of inadequate presentation in the task (CRIR) and Outlook: We should promote our work more effectively. (CRIR)
7
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
methods on students reflection quality. Two G1 students (student 4 and study, the U values are relatively high; however, they remain within the
student 30) conducted their reflections in the form of reports. Due to the acceptable range for statistical analysis. Some of these differences
lack of specific guidance from the instructor, who only provided general showed relatively small effect sizes, which will be further addressed in
requirements, their reflections remained superficial, primarily involving the discussion section.
DESR and EXPR. For example, student 4 wrote, “I have always enjoyed
radio shows, so I was very pleased to have the opportunity to create one 4.2. What differences do high and low-performance teams show in
this time. “Student 30 mentioned, "The tone did not match the storyline, reflection quality when using the three reflection methods?
and the sound quality of the program was poor. These reflections remain
limited to mere descriptions of the phenomena, needing more in-depth Fig. 4 illustrates the distribution of students from the 3 reflection
analysis of the underlying causes and offering no insights for future methods (G1, G2, G3) along the two principal component axes (SVD1
improvement. This tendency may be related to the relatively broad and SVD2). The points of different colors and shapes in the figure
scope of the reports. These examples demonstrate that structured represent high and low-performance teams within each group, indi­
guidance exerts a positive effect on the quality of reflection. In addition, cating their performance across various reflection categories, such as
they highlight the importance of timely feedback and question DESR, EXPR, CONR, and CRIR. The SVD1 axis accounts for 77.3 % of the
prompting. Providing students with immediate feedback based on their total variance, while the SVD2 axis explains 16.8 %. The position of each
responses and guiding them toward more elaborated answers contrib­ point represents the students tendencies in reflection content, with
utes to fostering deeper levels of reflection. points closer to a specific reflection category indicating that the groups
In contrast, two students from Group G2 (student 6 and student 27), performance is more concentrated in that category.
guided by the 4 aspects provided by the instructor and reflecting In Fig. 4, the centroids of the low-performance teams in G1 and G2
through short-answer questions, demonstrated a higher reflection are positioned relatively close to each other, with the low-performance
quality. The instructor guided students to reflect on four dimensions, teams located higher near DESR. Conversely, the high-performance
including task, division of labor, self-evaluation, and outlook. This teams are situated lower, closer to CRIR. This indicates a certain de­
approach, particularly in the latter two areas, effectively fostered CRIR gree of similarity in the reflection content between the low-performance
and CONR. For example, student 6 mentioned, “We plan to collaborate teams in G1 and G2. G3 is distributed on the right side of the figure, with
more effectively in completing each future learning task, striving to a greater distance between the high and low-performance teams, indi­
achieve the best outcome” (CRIR). At the same time, student 27 stated, cating a more pronounced difference in reflection content than the other
“Although our team may not be the best among all teams, we conveyed teams. Unlike G1 and G2, the G3 high-performance teams are positioned
our unique message. If there is a next time, we will work harder to at the top, closer to CONR, while the low-performance teams are located
improve” (CONR and CRIR). This structured guidance enhanced the at the bottom, near CRIR and EXPR. This suggests that the high-
depth of reflection. However, since short-answer questions are a one- performance teams in G3 tend to engage more in connective reflec­
way form of reflection for students, the instructor may not intervene tion, whereas the low-performance teams focus more on critical and
in their responses. As a result, there may be instances where students explanatory reflection.
provide irrelevant answers or overly brief responses, which can affect The study employed the Mann-Whitney U test to elucidate further
the overall reflection quality. For instance, student 6 responded with the scaling characteristics of the differences in reflection content be­
“Very successful” in the self-evaluation section (DESR), which lacked tween the high and low-performance teams across the 3 cohorts
depth in reflection. The AI agent could address this shortcoming by (Table 8). According to the results of the Mann-Whitney U test, there are
facilitating continuous interaction and feedback, encouraging students differences in the reflection content performance between the high and
to engage in deeper reflection.
When comparing the effectiveness of the reflection methods in G1,
G2, and G3, G1s reflection reports were of lower quality, primarily
focusing on DESR and EXPR. Due to the absence of specific guidance, the
reflections needed more depth. The short-answer questions format in G2
improved reflection quality to some extent. Students reflections became
more focused with the instructors guidance, particularly improving
CRIR and CONR. However, this approach is still constrained by the
limitations of outcome-based assessment. The AI agent guidance in G3
further enhanced reflection quality. Through real-time feedback and
targeted questioning, students could engage in deeper levels of CRIR and
CONR.
To scale these differences, the Mann-Whitney U test was employed to
evaluate the distribution of the projection points of the 3 groups of
students within the ENA space. The results indicated that at the α = 0.05
significance level, G1 and G2 showed significant differences in both the
first dimension (U = 147,537, p = 0.01, r = 0.09) and the second
dimension (U = 147,204, p = 0.01, r = 0.08). This suggests that the
structured guidance provided by short-answer questions enhances
reflection quality. G1 and G3 also showed a significant difference in the
first dimension (U = 99,595.5, p = 0.00, r = 0.34), highlighting the
impact of integrating the AI agent in G3 to enhance reflection quality.
However, no difference was observed in the second dimension (U =
147,049.5, p = 0.42, r = 0.03). Additionally, G2 and G3 exhibited dif­
ferences in both the first dimension (U = 127,246.5, p = 0.00, r = 0.36)
and the second dimension (U = 215,386.5, p = 0.01, r = 0.08), further
demonstrating the effectiveness of the AI agent in fostering deeper
reflection. This effect surpasses that of the structured short-answer Fig. 4. The centroid distribution of high and low group students across the
questions approach alone. Notably, due to the large sample size in this three reflection methods.
8
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
Table 8
The reflection content distribution of high and low-performance teams across the three methods.
low-performance teams across different reflection approaches. In G1, highlighted, AI may assist learners in constructing their learning pro­
the high and low-performance teams did not exhibit significant differ­ cesses, thereby enhancing critical thinking. In higher education, Xia and
ences in either dimension (MR1: U = 4932.00, p = 0.41, r = 0.05; MR2: Li [73] also suggested that AI assistants have a positive impact on stu­
U = 5463.00, p = 0.44, r = 0.05). In G2, the high and low-performance dents imagination, creativity, critical thinking, and autonomous
teams showed a significant difference in the MR1 dimension (U = learning. Zang et al. [69] experimentally confirmed the role of AI agents
3303.00, p = 0.03, r = 0.19) but no difference in the MR2 dimension (U in enhancing students critical thinking in English learning. However,
= 3051.00, p = 0.26, r = 0.10). For G3 (students using AI agent-driven the systematic review by Mohamud et al. [74] indicated that the
continuous questioning), the high and low-performance teams showed a introduction of AI in higher education may diminish students critical
significant difference in the MR1 dimension (U = 1136.50, p < 0.001, r thinking. This conclusion contradicts the findings of this study. The
= 0.45). In contrast, the difference in the MR2 dimension was insignif­ differences may be due to a lack of proper instructional design by
icant (U = 2187.50, p = 0.54, r = 0.06). teachers when using AI [74]. Cronje [75] argued that AI may serve as a
In G3, the differences between the high and low-performance teams teaching assistant to facilitate learning, but it should be integrated with
were the most pronounced, particularly on the MR1 dimension. Further instructional design and necessary prompts. In this study, the SSRL
analysis of the ENA diagram revealed that low-performance teams reflection checklist was operationalized as structured prompts to cali­
exhibited stronger connections in EXPR-CRIR (0.46) and EXPR-CONR brate the AI agent, enabling it to scaffold students reflections across the
(0.61). This suggests that the AI agent-driven reflection method may four phases of SSRL. By embedding SSRL principles into its dialogic
help low-performance teams focus more on specific reflection content. design, the agent acted as both a facilitator of reflection and a medium
for delivering theoretical scaffolds. This underscores the importance for
5. Discussion educators and researchers to apply instructional theory and design
thoughtfully when integrating AI into the classroom.
This section analyzes the findings based on the research questions. It In addition to SSRL theoretical guidance, the AI agent leveraged its
covers the positive impact of AI agents on students SSRL reflection, technological capabilities, including continuous questioning and real-
differences in reflection quality between high and low-performance time feedback, to actively scaffold deeper student reflections. Wolf­
teams, and key considerations for using AI agents effectively in SSRL. bauer et al. [76] noted that continuous dialogue with intelligent assis­
tants enhances students levels of reflection. In the G3 group, the AI
agents not only guided students to explore the root causes of issues but
5.1. The positive role of AI agents in students SSRL reflection
also helped them develop specific improvement plans. This guiding
process is similar to the “Socratic method” in educational psychology.
In SSRL, the AI agent reflection assistant enhanced the quality of
Through a series of targeted questions, students are encouraged to
students reflections. This outcome aligns with previous research [70,
engage in deep thinking and gain a more profound understanding of the
71]. For instance, Maedche et al. [70] demonstrated the positive role of
knowledge [77]. In addition, the timely feedback function of AI agents
AI agents in fostering deeper reflection among students. Sigman et al.
plays a crucial role in enhancing the quality of students SSRL re­
[71] also found that AI assistants emulate and augment human cogni­
flections. Self-determination theory suggests that providing positive
tion, thereby promoting reflection. These studies provide more evidence
emotional support through feedback helps students gain a sense of
of the positive impact AI agents have on facilitating reflective practices
belonging, thereby enhancing their motivation to learn and willingness
in education.
to reflect [78]. Uygur et al. [79] suggested that timely feedback
This study further clarifies how AI agents enhance the quality of
enhanced students reflection and learning. However, traditional SSRL
student reflection in the SSRL process through ENA. In these activities,
reflection reports and short-answer questions are one-way reflective
student reflections guided by AI agents exhibited higher levels of critical
activities, lacking immediate feedback and guidance. The AI agent
thinking and coherence. In contrast, the other two traditional reflective
reflection assistant compensates for the shortcomings of teachers in
texts displayed lower levels of reflection, focusing primarily on
providing timely feedback, enhancing the effectiveness of collaborative
descriptive and exploratory reflection. As Rusandi et al. [72]
9
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
learning. examine how to fine-tune AI guidance so that it benefits high performers
This study indicates that the level of reflection guidance directly without disrupting their existing strategies.
affects learners reflection quality, which is consistent with previous Additionally, there was no significant difference in performance
research [8082]. G1, with minimal guidance, showed the lowest between high and low-performance student teams in reflective reports,
quality, while G2, guided by the SSRL reflection checklist, exhibited with both showing low quality reflections. This may be due to learners
higher-quality reflections, demonstrating the importance of SSRL scaf­ lacking clear guidance in the reflection process. Maedche et al. [70]
folds. G3 combined SSRL scaffolding with real-time feedback and found that in reflective environments lacking external feedback or
encouragement for deeper reflection. Comparisons suggest that while structured guidance, the quality of students reflections is constrained.
structured short-answer questions had a limited impact, the AI agent This suggests that instructors should provide the necessary scaffolding
provided a practically meaningful enhancement of students reflective when designing reflective tasks. The SSRL scaffolding demonstrated
practices. However, these findings are based primarily on qualitative significant value in this study and is well-suited for broader application
data, and further quantitative research is needed to validate them. in collaborative settings.
In summary, AI agents play a substantial role in promoting student
reflection. Although the comparison between structured short-answer 5.3. Considerations for the effective use of AI agents in SSRL
questions and traditional reflective reports showed statistically signifi­
cant but very small effects, this suggests that short-answer questions Although experiments have demonstrated that AI agents enhance
alone had a limited impact on enhancing students reflection quality. In SSRL reflection quality, there are several limitations in their usage. To
contrast, the AI agent had a substantially greater impact on students better promote the outcomes of this study, we offer considerations for
reflective practices. It is essential for educators and instructional de­ teachers and instructional designers regarding the use of AI agents.
signers to integrate AI agents into classrooms and develop more Firstly, the quality and reliability of feedback provided by AI agents
instructional design case studies. Moreover, teachers should prioritize still present limitations. This finding aligns with the studies of Maloney
the importance of instructional theories and provide essential design et al. [91] and Fedus et al. [92], which suggest that the accuracy and
guidance when applying AI agents. effectiveness of AI agents depend on algorithm design and data quality.
In this study, the AI agent exhibited two primary issues: repeated
5.2. Differences between high and low-performance teams under various questioning and unexpected interruptions during conversations. To
SSRL reflection methods address the issue of repeated questioning, adjustments to the prompt
design can be implemented. For example, the prompts specify that each
The results indicate a significant difference in the high and low- question should be asked only once and repeated only if the student
performance teams that utilized reflective short-answer questions and responds off-topic or does not answer. For unexpected interruptions,
the AI agent reflection assistant. In short-answer questions, high- teachers need to guide students in testing their network environment
performance teams performed better. This aligns with the conclusions and re-engaging with the task. These observations show that AI agents
of Knight et al. [83], who found that high-performance students out­ need improvement in handling complex contexts and dynamic learning
performed low-performance students in reflective questions. The needs.
disparity in reflection between high and low-performance learners is In addition, data privacy and ethical concerns pose another chal­
primarily attributed to their metacognitive levels and learning strategies lenge in the application of AI agents. AI agents require extensive data
[8486]. For instance, Safari and Fitriati [85] found that collection, including students reflection content, behavioral patterns,
high-performance learners were able to use all strategies equally, but and learning habits [93]. To mitigate this issue, this study incorporated
low-performance learners more frequently relied on metacognitive and an opening message in the AI agents script. The message advised stu­
social strategies. These differences may impact learners outcomes, dents: “Please do not disclose personal sensitive information, such as
including their learning effectiveness and reflection [84]. your name or school, during the interaction.” Furthermore, before
In contrast, the reflection quality of low-performance teams using the implementing the AI agent, teachers need to raise students awareness of
AI agent reflective assistant was better than that of the high- data security and privacy protection [94].
performance teams. This is a novel finding of the study, suggesting The risks associated with over-reliance on AI technology should also
that the AI reflective assistant played a positive role in guiding low- be carefully evaluated. Although AI agents can provide personalized
performance learners through the reflection process. This finding support, they cannot fully replace the role of human teachers, particu­
aligns with previous evidence showing that AI technologies tend to larly in offering emotional support and fostering social interaction [95].
provide greater benefits for lower performers [8790]. Prior studies In this study, AI agents were utilized exclusively in the post-class
have suggested that such differential effects often occur because an AI reflection phase. The remaining instructional time relied on
chatbot can use adaptive strategies and personalized feedback to address face-to-face interactions between teachers and students. As GAI tech­
the strategic gaps of low performers [88]. AI tutoring can also offer both nology becomes increasingly accessible, preventing students from
cognitive and emotional support [89]. Xu et al. [90] further found that developing dependency behaviors may become more challenging.
low-performing learners become more engaged when they receive im­ Future research could explore strategies to prevent learners from
mediate feedback and external help. This engagement encourages them becoming overly reliant on GAI technologies.
to apply higher-order thinking strategies more actively. While AI agents have demonstrated advantages in enhancing stu­
These mechanisms may also explain the current results in our SSRL dents SSRL reflection quality, their widespread applicability is con­
reflection task. The AI reflection assistant provided structured guidance strained by feedback quality, data privacy, and ethical considerations.
in real time and reduced the cognitive load of producing reflections. This Future research should emphasize these limitations, refining the appli­
allowed low-performing learners to focus more on critical and creative cation framework of AI to ensure its effectiveness and sustainability in
thinking. In contrast, high-performing learners may already have the educational domain.
established reflection routines. Extra guidance could interfere with these
processes, leading to smaller gains in reflection quality [87]. 6. Conclusion, limitations, and future research
This study, therefore, not only confirms that differential effects exist
in reflection tasks but also highlights the potential of AI support to This study explores methods to enhance student reflection quality by
promote higher-order thinking in low-performing learners. In educa­ designing an AI agent that supports reflection through continuous
tional practice, this suggests that AI reflection assistants could be stra­ questioning and real-time feedback. Using content analysis and ENA,
tegically deployed to close performance gaps. Future research could this study conducted a three-semester experiment comparing reflection
10
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
reports, short-answer questions, and an AI agent reflection assistant. The National Natural Science Foundation of China (Grant Number:
results indicate that AI agents improve reflection quality, particularly for 62577035). The other authors declare that they have no known
low-performance teams. The study offers practical guidance for inte­ competing financial interests or personal relationships that could have
grating AI into SSRL-based instruction. appeared to influence the work reported in this paper.
Although this study contributes to understanding students reflection
behaviors in SSRL, several limitations remain. The first limitation arises Appendix A. The Critical Thinking Questionnaire (CThQ)
from the study participants. Conducted within a higher education
setting, this research primarily examines the effectiveness of using AI Instructions: For each statement below, please indicate how much
agents to facilitate reflection among university students. Only 97 stu­ you agree using a 5-point Likert scale (1 = Strongly disagree, 2 =
dents from the “Internet Thinking and Digital Self-Learning” course Disagree, 3 = Neutral, 4 = Agree, 5 = Strongly agree).
participated, so the findings may not be generalizable to other courses or
age groups. Further research is needed to explore the potential impact 1. After reading a text, I check important information, even if it
and adaptability of AI agents in secondary and primary education set­ seems to be true.
tings [96]. Secondly, the AI agent still has limitations in the quality and 2. I like combining information from different texts.
reliability of feedback, which may affect the depth and quality of stu­ 3. I am willing to share newly acquired information.
dents reflections. Addressing this issue relies on rapidly updating and 4. In-depth analyses of reality is a waste of time.
optimizing large AI model algorithms to provide higher-quality and 5. After reading a text, I can recall important points.
more targeted feedback. The third limitation is that the three reflection 6. The same content can be expressed in many different ways.
methods used in this experiment all fall under outcome-based reflection, 7. I can understand texts from various fields.
overlooking the dynamic process of students reflections at different 8. I form my impressions based on various pieces of information that
stages of collaborative learning. Additionally, the proposed mechanisms I combine.
underlying the AI agents impact on reflection quality, particularly for 9. Everything already exists, so nothing completely new can be
low-performance teams, remain hypothetical and require further created.
empirical validation through quantitative studies. Lastly, this study did 10. When I talk, I give many examples.
not differentiate the specific contributions of individual design elements 11. In discussions, I care about justifying my stance while under­
in the AI agents interaction strategy (e.g., sequential questioning, standing the other party.
encouraging feedback, simplified language). More research could adopt 12. I like finding connections between seemingly different
ablation analysis to examine how these elements independently influ­ phenomena.
ence students reflective practices. 13. I can see the structure of a text, and I could reorganize it.
Based on the limitations identified in this study, future research 14. When discussing, I try to use practical examples to justify my
could expand the study to more diverse educational contexts, including stance.
secondary and primary education, to examine the generalizability and 15. If necessary, I can recall information I have read before.
adaptability of AI agents. Incorporating multi-modal data, such as stu­ 16. I do not remember much of what I learned at school.
dents facial expressions, gestures, and dialogue, may offer a more 17. When I am interested in some information, I try to verify whether
comprehensive understanding of reflective behaviors in SSRL. Im­ it is true.
provements in AI models are needed to enhance the quality and reli­ 18. I can extract the most relevant parts of a text.
ability of feedback, supporting deeper and higher-quality student 19. To evaluate information, I check multiple sources.
reflections. In addition, investigating the individual contributions of 20. I like discussing new interpretations of texts I already know.
specific design elements in AI agents interaction strategies, for example, 21. I like to collate different opinions and compare them.
through ablation-style comparisons, could clarify which features most 22. I have difficulties with paraphrasing.
effectively promote high-order reflection, particularly among low- 23. I try to apply the information I have learned in everyday life.
performance teams. We therefore urge more researchers to focus on 24. When I read, I look for relationships between its information and
this area of study, exploring the impact of GAI on educational outcomes other texts I have read.
to better understand and harness its potential for improving educational 25. I pay attention to the contexts, nuances, and overtones of
practices. statements.
Declaration of generative AI in the writing process
Data availability
During the preparation of this work, the authors used Kimi (https:
//kimi.moonshot.cn/) to improve language and readability. After The datasets generated and analyzed during the current study are
using this tool, the authors reviewed and edited the content as needed available from the corresponding author on reasonable request.
and take full responsibility for the content of the publication.
References
CRediT authorship contribution statement
[1] S. Ahmad, M. Rahmat, M. Mubarik, M. Alam, S. Hyder, Artificial intelligence and
its role in education, Sustainability 13 (22) (2021) 12902.
Yumin Zheng: Writing original draft, Conceptualization. Fengjiao [2] X. Gong, Z. Li, A. Qiao, Impact of generative AI dialogic feedback on different
Tu: Investigation, Data curation. Fengfang Shu: Investigation, Data stages of programming problem solving, Educ. Inf. Technol. 30 (7) (2025)
curation. Chaowang Shang: Formal analysis, Data curation. Lulu 96899709.
[3] O. Tapalova, N. Zhiyenbayeva, D. Gura, Artificial Intelligence in Education: aIEd
Chen: Writing review & editing, Formal analysis. Jiang Meng: for personalised learning pathways, Electron. J. e-Learn. 20 (5) (2022) 639653.
Investigation. [4] S. Järvelä, P. Kirschner, E. Panadero, J. Malmberg, C. Phielix, J. Jaspers,
M. Koivuniemi, H. Järvenoja, Enhancing socially shared regulation in collaborative
learning groups: designing for CSCL regulation tools, in: Educ. Technol. Res. Dev.,
Declaration of competing interest 63, 2014, pp. 125142.
[5] D. Bransen, M.J.B. Govaerts, E. Panadero, et al., Putting self-regulated learning in
The authors declare the following financial interests/personal re­ context: integrating self-, co-, and socially shared regulation of learning, Med.
Educ. 56 (1) (2022) 2936.
lationships which may be considered as potential competing interests:
Chaowang Shang acknowledges the financial support from the
11
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
[6] E. Eshuis, J. Vrugte, A. Anjewierden, L. Bollen, J. Sikken, T. Jong, Improving the [36] E. Panadero, S. Järvelä, Socially shared Regulation of Learning: a review, Eur.
quality of vocational students collaboration and knowledge acquisition through Psychol. 20 (2015) 190203.
instruction and joint reflection, Int. J. Comput.-Support. Collab. Learn. 14 (2019) [37] J. Isohätälä, H. Järvenoja, S. Järvelä, Socially shared regulation of learning and
5376. participation in social interaction in collaborative learning, Int. J. Educ. Res. 81
[7] C. Chan, K. Lee, Reflection literacy: a multilevel perspective on the challenges of (2017) 1124.
using reflections in higher education through a comprehensive literature review, [38] J. Li, Y. Lin, M. Sun, R. Shadiev, Socially shared regulation of learning in game-
Educ. Res. Rev. 32 (2020) 100376. based collaborative learning environments promotes algorithmic thinking,
[8] L. Guo, How should reflection be supported in higher education? — A meta- learning participation, and positive learning attitudes, Interact. Learn. Environ. 31
analysis of reflection interventions, Reflective Pract. 23 (2021) 118146. (2020) 17151726.
[9] S. Popenici, S. Kerr, Exploring the impact of artificial intelligence on teaching and [39] J. Malmberg, S. Järvelä, H. Järvenoja, E. Panadero, Promoting socially shared
learning in higher education, Res. Pract. Technol. Enhanc. Learn. 12 (1) (2017) 22. regulation of learning in CSCL: progress of socially shared regulation among high-
[10] H. Kiy, A Study on Writing Experience With ChatGPT of College Students, J. Korea and low-performing groups, Comput. Hum. Behav. 52 (2015) 562572.
Converg. Soc. 14 (9) (2024) 976. [40] J. Yukawa, Co-reflection in online learning: collaborative critical thinking as
[11] K. Hanifi, O. Cetin, C. Yilmaz, On ChatGPT: perspectives from software engineering narrative, Int. J. Comput.-Support. Collab. Learn. 1 (2006) 203228.
students, in: Proc. 2023 IEEE 23rd Int. Conf. Softw. Qual. Reliab. Secur. (QRS), [41] A. Głowala, M. Kołodziejski, T. Butvilas, Reflection as a basic category of a
2023, pp. 196205. teachers thinking and action, Multidiscip. J. Sch. Educ. 12.1(2023):229250.
[12] Zhiheng Xi, et al., The rise and potential of large language model based agents: A [42] J. Buck, Reflecting on reflections: a case study of disappointment in student writing
survey, Sci. China Inf. Sci. 68 (2) (2025) 121101. assignments, J. Acoust. Soc. Am. (2023). A273-A273.
[13] E. Katsarou, F. Wild, A. Sougari, P. Chatzipanagiotou, A systematic review of voice- [43] N Rahmi, C M Zubainur, Students mathematical reflective thinking ability through
based intelligent virtual agents in EFL education, Int. J. Emerg. Technol. Learn. scaffolding strategies[C]//Journal of Physics: Conference Series, IOP Publishing
(iJET) 18 (10) (2023) 6585. 1460 (1) (2020) 012022.
[14] P.R. Lewis, Ş. Sarkadi, Reflective artificial intelligence, Minds Mach. 34 (2) (2024) [44] J. Dewey, Education Democracy, The elementary school teacher 4 (4) (1903)
14. 193204.
[15] Z. Xu, P. Zhang, M. Tu, M. Zhang, Y. Lai, Brain optimization with additional study [45] B.J. Zimmerman, Self-regulated learning and academic achievement: an overview,
time: potential brain differences between high- and low-performance college Educ. Psychol. 25 (1) (1990) 317.
students, Front. Psychol. 14 (2023) 1209881. [46] D. Coulson, M. Harvey, Scaffolding student reflection for experience-based
[16] UK Government, Generative Artificial Intelligence (AI) in Education, GOV.UK, learning: a framework, Teach. High. Educ. 18 (2013) 401413.
2023. https://www.gov.uk/government/publications/generative-artificial-i [47] S. Lajoie, Extending the scaffolding metaphor, Instr. Sci. 33 (2005) 541557.
ntelligence-in-education/generative-artificial-intelligence-ai-in-education. [48] E. Panadero, P.A. Kirschner, S. Järvelä, J. Malmberg, H. Järvenoja, How individual
[17] M. Dogan, T. Dogan, A. Bozkurt, The use of artificial intelligence (AI) in online self-regulation affects group regulation and performance: a shared regulation
learning and distance education processes: a systematic review of Empirical intervention, Small Group Res. 46 (4) (2015) 431454.
Studies, Appl. Sci. 13 (5) (2023) 3056. [49] E. Davis, Prompting middle school science students for productive reflection:
[18] L. Shi, The integration of advanced AI-enabled emotion detection and adaptive generic and directed prompts, J. Learn. Sci. 12 (2003) 142191.
learning systems for improved emotional regulation, J. Educ. Comput. Res. 63 [50] J. Hattie, H. Timperley, The Power of Feedback, Rev. Educ. Res. 77 (2007)
(2024) 173201. 112181.
[19] B. Tang, J. Liang, W. Hu, H. Luo, Enhancing programming performance, learning [51] R. Ajjawi, F. Kent, J. Broadbent, J. Tai, M. Bearman, D. Boud, Feedback that works:
interest, and self-efficacy: the role of large language models in middle school a realist review of feedback interventions for written tasks, Stud. High. Educ. 47
education, Systems 30 (6) (2025) 81098138. (2021) 13431356.
[20] L. Feng, Investigating the effects of artificial intelligence-assisted language learning [52] U. Krause, R. Stark, Reflection in example- and problem-based learning: effects of
strategies on cognitive load and learning outcomes: a comparative study, J. Educ. reflection prompts, feedback, and cooperative learning, Eval. Res. Educ. 23 (2010)
Comput. Res. 62 (8) (2025) 17411774. 255272.
[21] Q. Huang, W. Li, Y. Zhao, Enhancing deep learning and motivation in university [53] J. Contreras, S. Edwards-Maddox, A. Hall, M. Lee, Effects of reflective practice on
English education through AI technology: a quasi-experimental study, Asian J. baccalaureate nursing students Stress, Anxiety, and competency: an integrative
Educ. Soc. Stud. 51 (4) (2025) 452463. review, Worldviews Evid.-Based Nurs. 17 (3) (2020) 239245.
[22] Ó. Cuéllar, M. Contero, M. Hincapié, Personalized and Timely Feedback in Online [54] H. Gadsby, Fostering reflective practice in Post Graduate Certificate in Education
education: Enhancing learning With Deep Learning and Large Language Models, students through reflective journals. Developing a typology for reflection,
MTI. 9 (5) (2025) 45. Reflective Pract. 23 (2022) 357368.
[23] X. Zhou, D. Teng, H. Al-Samarraie, The mediating role of generative AI self- [55] S. Rabu, N. Badlishah, Levels of students Reflective thinking skills in a
regulation on students critical thinking and problem-solving, Educ. Sci. 14 (12) collaborative learning environment using Google Docs, TechTrends 64 (2020)
(2024) 1302. 533541.
[24] S. Steenbergen-Hu, H. Cooper, A meta-analysis of the effectiveness of intelligent [56] J. Stoszkowski, A. Hodgkinson, D. Collins, Using Flipgrid to improve reflection: a
tutoring systems on college students academic learning, J. Educ. Psychol. 106 collaborative online approach to coach development, Phys. Educ. Sport Pedagogy
(2014) 331347. 26 (2020) 167178.
[25] C. Moridis, A. Economides, Affective learning: empathetic agents with emotional [57] E. Liesa, P. Mayoral, M. Giralt-Romeu, S. Angulo, Video-Based Feedback for
facial and tone of voice expressions, IEEE Trans. Affect. Comput. 3 (2012) Collaborative Reflection Among Mentors, University Tutors, and Students, Edu.
260272. Sci. 13 (9) (2023) 879.
[26] S. Nelekar, A. Abdulrahman, M. Gupta, D. Richards, Effectiveness of embodied [58] M. Alghasab, J. Hardman, Z. Handley, Teacher-student Interaction On wikis:
conversational agents for managing academic stress at an Indian university (ARU) Fostering collaborative Learning and Writing, Learn Cult. Soc. Inter. 21 (2019)
during COVID-19, Br. J. Educ. Technol. 53 (2021) 491511. 1020.
[27] W. Sun, Q. Chen, The design, implementation, and evaluation of Gamified [59] R. Gubareva, R. Lopes, Virtual Assistants for learning: a systematic literature
Immersive Virtual Reality (IVR) for learning: a review of Empirical Studies, Proc. review, CSEDU (1) (2020) 97103.
Eur. Conf. Games-Based Learn. 17 (1) (2023) 789797. [60] L. González, H. Neyem, I. Contreras-McKay, D. Molina, Improving learning
[28] M. Chen, L. Wu, Z. Liu, X. Ma, The impact of metacognitive strategy-supported experiences in software engineering capstone courses using artificial intelligence
intelligent agents on the quality of collaborative learning from the perspective of virtual assistants, Comput. Appl. Eng. Educ. 30 (2022) 13701389.
the community of inquiry, in: Proc. 2024 4th Int. Conf. Educ. Technol. (ICET), [61] B. Renner, G. Wesiak, V. Pammer-Schindler, M. Prilla, L. Müller, D. Morosini,
2024, pp. 1117. S. Mora, N. Faltin, U. Cress, Computer-supported reflective learning: how apps can
[29] H. Hong, C. Viriyavejakul, P. Vate-U-Lan, Enhancing critical thinking skills: foster reflection at work, Behav. Inf. Technol. 39 (2019) 167187.
exploring generative AI-enabled cognitive offload instruction in English essay [62] A. Freiberg-Hoffmann, A. Romero-Medina, B. López-Fernández, M. Fernández-
writing, 4, ECOHUMANISM Учредители: Transnational Press, London, p. 2024. Liporace, Learning approaches: cross-cultural differences (SpainArgentina) and
[30] D.H. Schunk, B.J. Zimmerman, Motivation and Self-Regulated learning: Theory, academic achievement in college students, Span. J. Psychol. 26 (2023) e16.
research, and Applications, Routledge, 2012. [63] A. Kobylarek, K. Błaszczyński, L. Ślósarz, M. Madej, Critical thinking Questionnaire
[31] P.H. Winne, A.F. Hadwin, N.E. Perry. Metacognition and computer-supported (CThQ)construction and application of a critical thinking test tool, Andragogy
collaborative learning, The International Handbook of Collaborative Learning, Adult Educ. Soc. Mark. 2 (2) (2022), 1-1.
Routledge, 2013, pp. 462479. [64] J. Dewey, An analysis of reflective thought, J. Philos. (1922) 2938.
[32] Y. Su, Y. Li, H. Hu, et al., Exploring college English language learners self and [65] D.T. Campbell, J.C. Stanley, Experimental and Quasi-Experimental Designs For
social regulation of learning during wiki-supported collaborative reading activities, Research, Ravenio Books, 2015.
Int. J. Comput.-Support. Collab. Learn. 13 (2018) 3560. [66] M.M. Plack, M. Driscoll, S. Blissett, R. McKenna, T.P. Plack, A method for assessing
[33] F. Tu, L. Wu, Kinshuk, et al., Exploring the influence of regulated learning reflective journal writing, J. Allied Health 34 (4) (2005) 199208.
processes on learners prestige in project-based learning, Educ. Inf. Technol. 30 (2) [67] L. Wang, G. Wu, J. Wu, A Study on the Reflective Level of Teachers
(2025) 22992329. Autobiography, Global Education Outlook (01), (2018) 93105.
[34] S. Zhang, J. Chen, Y. Wen, H. Chen, Q. Gao, Q. Wang, Capturing regulatory [68] H.T. Hou, Integrating cluster and sequential analysis to explore learners flow and
patterns in online collaborative learning: a network analytic approach, Int. J. behavioral patterns in a simulation game with a situated-learning context for
Comput.-Support. Collab. Learn. 16 (2021) 3766. science courses: a video-based process exploration, Comput. Human Behav. 48
[35] J. Zheng, W. Xing, G. Zhu, Examining sequential patterns of self-and socially (2015) 424435.
shared regulation of STEM learning in a CSCL environment, Comput. Educ. 136
(2019) 3448.
12
Y. Zheng et al. Computer Standards & Interfaces 97 (2026) 104094
[69] G. Zang, M. Liu, B. Yu, The application of 5G and artificial intelligence technology [83] J. Knight, D. Weaver, M. Peffer, Z. Hazlett, Relationships between prediction
in the innovation and reform of college English education, Comput. Intell. accuracy, metacognitive reflection, and performance in introductory genetics
Neurosci. 2022 (1) (2022) 9008270. students, CBE Life Sci. Educ. 21 (3) (2022) ar45.
[70] A. Maedche, C. Legner, A. Benlian, B. Berger, H. Gimpel, T. Hess, O. Hinz, [84] D. Difrancesca, J. Nietfeld, L. Cao, A comparison of high and low achieving
S. Morana, M. Söllner, AI-based digital assistants, Bus. Inf. Syst. Eng. 61 (2019) students on self-regulated learning variables, Learn. Individ. Differ. 45 (2016)
535544. 228236.
[71] M. Sigman, D. Slezak, L. Drucaroff, S. Ribeiro, F. Carrillo, Artificial and Human [85] S A Gani, D Fajrina, R Hanifa, Students learning strategies for developing speaking
intelligence in mental health, AI Mag. 42 (2021) 3946. ability[J], Stud. Eng. lang. educ. 2 (1) (2015) 1628.
[72] M.A. Rusandi, I. Saripah, D.M. Khairun, No worries with ChatGPT: building bridges [86] M. Yip, Differences between high and low academic achieving university students
between artificial intelligence and education with critical thinking soft skills, in learning and study strategies: a further investigation, Educ. Res. Eval. 15 (2009)
J. Public Health. 45 (3) (2023) e602e603. 561570.
[73] X. Xia, X. Li, Artificial intelligence for higher education development and teaching [87] H.K. Etkin, K.J. Etkin, R.J. Carter, C.E. Rolle, Differential effects of GPT-based tools
skills, Wirel, Commun. Mob. Comput. 2022 (1) (2022) 7614337. on comprehension of standardized passages, Front. Educ. 10 (2025) 1506752.
[74] Y. Mohamud, A. Marof, A. Mohamed, M. Uzir, A narrative review on the impact of [88] S. Ruan, A. Nie, W. Steenbergen, J. He, J.Q. Zhang, M. Guo, et al., A reinforcement
applied artificial intelligence tools on higher secondary students, Int. J. Acad. Res. learning tutor better supported lower performers in a math task, Mach. Learn. 113
Bus. Soc. Sci. 13 (14) (2023) 3442. (2024) 30233048.
[75] J. Cronje, Exploring the role of ChatGPT as a peer coach for developing research [89] D.R. Thomas, J. Lin, E. Gatz, A. Gurung, S. Gupta, K. Norberg, et al., Improving
proposals: feedback quality, prompts, and student reflection, Electron. J. (2024) student learning with hybrid human-AI tutoring: a three-study quasi-experimental
22.2, e-Learn. investigation, in: Proc. 14th Learn. Anal. Knowl. Conf., 2024, pp. 404415. New
[76] I. Wolfbauer, V. Pammer-Schindler, K. Maitz, C. Rosé, A script for conversational York, NY, USA: Association for Computing Machinery (LAK 24).
reflection guidance: a field study on developing reflection competence with [90] Y Xu, J Zhu, M Wang, et al., The impact of a digital game-based AI chatbot on
apprentices, IEEE Trans. Learn. Technol. 15 (2022) 554566. students academic performance, higher-order thinking, and behavioral patterns in
[77] F. Leigh, Platonic dialogue, maieutic method, and critical thinking, J. Philos. Educ. an information technology curriculum[J], App. Sci. 14 (15) (2024) 6418.
41 (2008) 309323. [91] Maloney, A., Roberts, D. A., & Sully, J. (2022). A solvable model of neural scaling
[78] E. Deci, R. Ryan. Intrinsic motivation and self-determination in Human behavior, laws. arXiv preprint arXiv:2210.16859.
1975, pp. 1371. [92] W. Fedus, B. Zoph, N. Shazeer, Switch transformers: scaling to trillion parameter
[79] J. Uygur, E. Stuart, M. Paor, E. Wallace, S. Duffy, M. OShea, S. Smith, models with simple and efficient sparsity, J. Mach. Learn. Res. 23 (120) (2022)
T. Pawlikowska, The Best evidence in Medical Education systematic review to 139.
determine the most effective teaching methods that develop reflection in medical [93] K. Seo, J. Tang, I. Roll, S. Fels, D. Yoon, The impact of artificial intelligence on
students: BEME Guide No. 51, Med. Teach. 41 (2019) 316. learnerinstructor interaction in online learning, Int. J. Educ. Technol. High. Educ.
[80] K. Arendt, L. Stark, A. Friedrich, R. Brünken, R. Stark, Quality of reflections on 18 (1) (2021) 54.
teaching: approaches to its measurement and low-threshold promotion, Educ. Sci. [94] B. Klimova, M. Pikhart, J. Kacetl, Ethical issues of the use of AI-driven mobile apps
15 (7) (2025) 884. for education, Front. Public Health 10 (2023) 1118116.
[81] J. Jung, Y. Lu, A. Ding, How do prompts shape preservice teachers reflections? A [95] T. Adiguzel, M. Kaya, F. Cansu, Revolutionizing education with AI: exploring the
case study in an online technology integration class, J. Teach. Educ. 73 (3) (2021) transformative potential of ChatGPT, Contemp. Educ. Technol. 15 (3) (2023).
301313. [96] M. Thottoli, B. Alruqaishi, A. Soosaimanickam, Robo academic advisor: can
[82] A. Sturgill, P. Motley, Methods of reflection about service learning: guided vs. free, chatbots and artificial intelligence replace human interaction? Contemp. Educ.
dialogic vs. expressive, and public vs. private. Teaching and learning inquiry, Technol. 16 (1) (2024) ep485.
ISSOTL J. 2 (1) (2014) 8193.
13