AI Scaffolding for Evidence-Based Critical Thinking in a Microcontroller Learning-Media Design Project

Harjito Harjito, Wahyu Hardyanto, Woro Sumarni, Sri Wardani

Abstract


The rapid integration of artificial intelligence (AI) into education has increased interest in how digital tools can support higher-order thinking rather than simply automate tasks. Within STEAM learning, this issue is especially important because students must combine evidence, reasoning, and design decisions in authentic project work. Generative AI tools based on large language models were used in this study as learning scaffolds to assist students with keyword expansion, scientific query construction, journal abstract summarization, hypothesis development, and revision of written explanations. The study examined whether such AI-supported learning improved students’ critical thinking during a STEAM microcontroller media-design project. Two intact classes completed the same project workflow: one class used AI support, while the other completed the project without AI. Student performance was evaluated using a five-dimension critical-thinking rubric covering information and keywords, concepts and logic, journal abstract interpretation, hypothesis and reasoning, and academic formatting and compliance. Results showed that the AI-supported class performed better overall, scoring about five points higher on average and showing a large overall advantage. Improvements appeared across all rubric dimensions, with the strongest gains found in journal abstract interpretation and hypothesis and reasoning, suggesting that AI was most helpful when students had to interpret evidence, connect ideas, and justify decisions. The weakest-link bottleneck analysis, defined as the rubric dimension in which each student performed worst, showed that journal abstract interpretation remained the main constraint in both groups. These findings indicate that AI can serve as a productive scaffold for critical thinking and decision-making in STEAM projects, especially by supporting evidence use and iterative reasoning, while also highlighting the need for explicit instruction in reading and interpreting scientific abstracts.


Keywords


Artificial intelligence in education; Critical thinking; STEAM learning; AI scaffolding; Project-based learning

Full Text:

PDF
rticle

References


Albadarin, Y., Saqr, M., Pope, N., & Tukiainen, M. (2024). A systematic literature review of empirical research on ChatGPT in education. Discover Education, 3, Article 60. https://doi.org/10.1007/s44217-024-00138-2

Berland, L. K., & Reiser, B. J. (2009). Making sense of argumentation and explanation. Science Education, 93(1), 26–55. https://doi.org/10.1002/sce.20286

Blikstein, P. (2013). Digital fabrication and “making” in education: The democratization of invention. In J. Walter-Herrmann & C. Büching (Eds.), FabLabs: Of machines, makers and inventors (pp. 1–22). Transcript Verlag. https://share.google/76G60acxkrzaFKJsi

Brookhart, S. M. (2010). How to assess higher-order thinking skills in your classroom. ASCD. https://share.google/1BgX7tSXmRR8pckeS

Butler, D. L., & Winne, P. H. (1995). Feedback and self-regulated learning: A theoretical synthesis. Review of Educational Research, 65(3), 245–281. https://doi.org/10.3102/00346543065003245

Cavagnetto, A. R. (2010). Argument to foster scientific literacy: A review of argument interventions in K–12 science contexts. Review of Educational Research, 80(3), 336–371. https://doi.org/10.3102/0034654310376953

Chernozhukov, V., Fernández-Val, I., & Melly, B. (2013). Inference on counterfactual distributions. Econometrica, 81(6), 2205–2268. https://doi.org/10.3982/ECTA10582

Cumming, G. (2014). The new statistics: Why and how. Psychological Science, 25(1), 7–29. https://doi.org/10.1177/0956797613504966

Dignath, C., Büttner, G., & Langfeldt, H. P. (2008). How can primary school students learn self-regulated learning strategies most effectively? A meta-analysis on self-regulation training programmes. Educational Research Review, 3(2), 101–129. https://doi.org/10.1016/j.edurev.2008.02.003

Efron, B. (1979). Bootstrap methods: Another look at the jackknife. The Annals of Statistics, 7(1), 1–26. https://doi.org/10.1214/aos/1176344552

Ennis, R. H. (1993). Critical thinking assessment. Theory Into Practice, 32(3), 179–186. https://doi.org/10.1080/00405849309543594

Facione, P. A. (1990). Critical thinking: A statement of expert consensus for purposes of educational assessment and instruction (The Delphi Report). American Philosophical Association. https://share.google/V6BILm5KABRlwY3Db

Firpo, S. (2007). Efficient semiparametric estimation of quantile treatment effects. Econometrica, 75(1), 259–286. https://doi.org/10.1111/j.1468-0262.2007.00738.x

Halpern, D. F. (2014). Thought and knowledge: An introduction to critical thinking (5th ed.). Psychology Press. https://doi.org/10.4324/9781315885278

Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77(1), 81–112. https://doi.org/10.3102/003465430298487

Hedges, L. V. (1981). Distribution theory for Glass’s estimator of effect size and related estimators. Journal of Educational and Behavioral Statistics, 6(2), 107–128. https://doi.org/10.3102/10769986006002107

Holmes, W., Bialik, M., & Fadel, C. (2019). Artificial intelligence in education: Promises and implications for teaching and learning. Center for Curriculum Redesign. https://share.google/Em6IhOHThMk1kV4Vm

Honey, M., Pearson, G., & Schweingruber, H. (Eds.). (2014). STEM integration in K–12 education: Status, prospects, and an agenda for research. National Academies Press. https://share.google/Xe8Y8R8v2usXbmX3X

Kasneci, E., Sessler, K., Küchemann, S., Bannert, M., Dementieva, D., Fischer, F., Gasser, U., Groh, G., Günnemann, S., Hüllermeier, E., Krusche, S., Kutyniok, G., Michaeli, T., Nerdel, C., Pfeffer, J., Poquet, O., Sailer, M., Schmidt, A., Seidel, T., & Kasneci, G. (2023). ChatGPT for good? On opportunities and challenges of large language models for education. Learning and Individual Differences, 103, 102274. https://doi.org/10.1016/j.lindif.2023.102274

Koenker, R., & Bassett, G. (1978). Regression quantiles. Econometrica, 46(1), 33–50. https://doi.org/10.2307/1913643

Lakens, D. (2013). Calculating and reporting effect sizes to facilitate cumulative science: A practical primer for t-tests and ANOVAs. Frontiers in Psychology, 4, Article 863. https://doi.org/10.3389/fpsyg.2013.00863

Lakens, D., Scheel, A. M., & Isager, P. M. (2018). Equivalence testing for psychological research: A tutorial. Advances in Methods and Practices in Psychological Science, 1(2), 259–269. https://doi.org/10.1177/2515245918770963

McNeill, K. L., & Krajcik, J. (2008). Inquiry and scientific explanations: Helping students use evidence and reasoning. Science Education, 92(2), 223–254. https://doi.org/10.1002/sce.20205

National Research Council. (2012). A framework for K–12 science education: Practices, crosscutting concepts, and core ideas. National Academies Press. https://share.google/iCVajQWmDPdVIA9NW

Nicol, D. J., & Macfarlane-Dick, D. (2006). Formative assessment and self-regulated learning: A model and seven principles of good feedback practice. Studies in Higher Education, 31(2), 199–218. https://doi.org/10.1080/03075070600572090

Osborne, J., Erduran, S., & Simon, S. (2004). Enhancing the quality of argumentation in school science. Journal of Research in Science Teaching, 41(10), 994–1020. https://doi.org/10.1002/tea.20035

Panadero, E., & Jonsson, A. (2013). The use of scoring rubrics for formative assessment purposes revisited: A review. Educational Research Review, 9, 129–144. https://doi.org/10.1016/j.edurev.2013.01.002

Papert, S. (1980). Mindstorms: Children, computers, and powerful ideas. Basic Books. https://share.google/r62JMhoM89licb8b2

Paul, R., & Elder, L. (2006). The miniature guide to critical thinking: Concepts & tools (4th ed.). Foundation for Critical Thinking. https://share.google/Sijbdb8redRwOfBd0

Sandoval, W. A., & Millwood, K. A. (2005). The quality of students’ use of evidence in written scientific explanations. Cognition and Instruction, 23(1), 23–55. https://doi.org/10.1207/s1532690xci2301_2

Sweller, J. (1988). Cognitive load during problem solving: Effects on learning. Cognitive Science, 12(2), 257–285. https://doi.org/10.1016/0364-0213(88)90023-7

Sweller, J., Ayres, P., & Kalyuga, S. (2011). Cognitive load theory. Springer. https://doi.org/10.1007/978-1-4419-8126-4

Thomas, J. W. (2000). A review of research on project-based learning. Autodesk Foundation. https://share.google/n9PuFKu4bh6ZLHGnF

Toulmin, S. (1958). The uses of argument. Cambridge University Press. https://doi.org/10.1017/CBO9780511840005

UNESCO. (2021). Recommendation on the ethics of artificial intelligence. UNESCO. https://share.google/057C8b8YaNYkI0wre

UNESCO. (2023). Guidance for generative AI in education and research. UNESCO. https://share.google/UBqDTqbkWcTD3H3f9

van Merriënboer, J. J. G., & Sweller, J. (2005). Cognitive load theory and complex learning: Recent developments and future directions. Educational Psychology Review, 17(2), 147–177. https://doi.org/10.1007/s10648-005-3951-0

Vygotsky, L. S. (1978). Mind in society: The development of higher psychological processes. Harvard University Press. https://share.google/IlYf9CvWuhmioBv5V

Wasserstein, R. L., & Lazar, N. A. (2016). The ASA statement on p-values: Context, process, and purpose. The American Statistician, 70(2), 129–133. https://doi.org/10.1080/00031305.2016.1154108




DOI: https://doi.org/10.20961/ijpte.v10i1.115698

Refbacks

  • There are currently no refbacks.


Copyright (c) 2026 Harjito, Wahyu Hardyanto, Woro Sumarni, Sri Wardani

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 International License.

  

International Journal of Pedagogy and Teacher Education

Print ISSN: 2597-7792
Online ISSN: 2549-8525
Website: https://jurnal.uns.ac.id/ijpte/index
Email: ijpte@mail.uns.ac.id
Published by: Faculty of Teacher Training and Education, Universitas Sebelas Maret
Ir. Sutami Street, No. 36A, Surakarta, Jawa Tengah Indonesia