Mapping research trends on development of scientific literacy assessment: A bibliometric analysis

Authors

  • Yasmine Hadiastriani Science Education Program, Postgraduate School, Pakuan University, Indonesia
  • Prasetyorini Djarot Biology Program, Mathematics and Science Faculty, Pakuan University, Indonesia

DOI:

https://doi.org/10.22219/jpbi.v10i1.31136

Keywords:

Kurikulum Merdeka, science literacy, multiple assessment, Critical Thinking, student learning interdependence

Abstract

This study conducts a bibliometric analysis to map the research trends and intellectual landscape surrounding the development of scientific literacy assessments in educational contexts from 2019 to 2023. Scientific literacy, which encompasses the ability to comprehend, evaluate, and apply scientific knowledge, has emerged as a crucial competency in science education. However, existing assessment tools often prioritize factual recall over deeper conceptual understanding and critical thinking abilities. To address this gap, there is a need for innovative assessment approaches that holistically evaluate students' scientific literacy across multiple dimensions. Through a systematic analysis of 185 relevant publications from the Scopus database, this study investigates the numerical distribution of literature by year, document type, journal, author, institution, and country. Additionally, it explores the prominent themes, evolving topic trends, and emerging clusters within the scientific literacy assessment research domain. Bibliometric techniques, including citation analysis, co-authorship networks, and keyword co-occurrence mapping, are employed to provide a comprehensive perspective. The findings reveal a steady increase in publications focused on scientific literacy assessments, reflecting the growing recognition of this research area. The analysis identifies highly influential journals, articles, authors, and institutions actively shaping the field. Furthermore, it highlights key research themes, such as secondary education, teacher professional development, and information literacy. However, the integration of scientific literacy assessments with artificial intelligence remains an underexplored area, presenting opportunities for future research.

Downloads

Download data is not yet available.

References

Allen, C., Nejdawi, R., El-Baba, J., Hamati, K., Metternicht, G., & Wiedmann, T. (2017). Indicator-based assessments of progress towards the sustainable development goals (SDGs): a case study from the Arab region. Sustainability Science, 12(6), 975–989. https://doi.org/10.1007/s11625-017-0437-1

Aria, D., Sukyadi, D., & Kurniawan, E. (2021). Teacher assessment literacy: Indonesian EFL secondary teachers’ self-perceived on classroom-based assessment practice. English Review: Journal of English Education, 10(1), 15–26. https://doi.org/10.25134/erjee.v10i1.5349

Aristeidou, M., & Herodotou, C. (2020). Online citizen science: A systematic review of effects on learning and scientific literacy. Citizen Science: Theory and Practice, 5(1), 1–12. https://doi.org/10.5334/cstp.224

Baas, J., Schotten, M., Plume, A., Côté, G., & Karimi, R. (2020). Scopus as a curated, high-quality bibliometric data source for academic research in quantitative science studies. Quantitative Science Studies, 1(1), 377–386. https://doi.org/10.1162/qss_a_00019

Broadus, R. N. (1987). Toward a definition of “bibliometrics.” Scientometrics, 12(5–6), 373–379. https://doi.org/10.1007/BF02016680

Coppi, M., Fialho, I., & Cid, M. (2023). Developing a scientific literacy assessment instrument for Portuguese 3rd cycle students. Education Sciences, 13(9), 1–25. https://doi.org/10.3390/educsci13090941

Djeki, E., Dégila, J., Bondiombouy, C., & Alhassan, M. H. (2022). E-learning bibliometric analysis from 2015 to 2020. Journal of Computers in Education, 9(4), 727–754. https://doi.org/10.1007/s40692-021-00218-4

Donthu, N., Kumar, S., Mukherjee, D., Pandey, N., & Lim, W. M. (2021). How to conduct a bibliometric analysis: An overview and guidelines. Journal of Business Research, 133(March), 285–296. https://doi.org/10.1016/j.jbusres.2021.04.070

Donthu, N., Kumar, S., Pandey, N., & Lim, W. M. (2021). Research constituents, intellectual structure, and collaboration patterns in journal of international marketing: An analytical retrospective. Journal of International Marketing, 29(2), 1–25. https://doi.org/10.1177/1069031X211004234

Eck, N. J. Van, & Waltman, L. (2014). Visualizing bibliometric networks. In Measuring Scholarly Impact (pp. 285–320). https://doi.org/10.1007/978-3-319-10377-8

Facione, P. A. (1991). Using the California critical thinking skills test in research, evaluation, and assessment. In California Academic Press (pp. 1–20). https://files.eric.ed.gov/fulltext/ED337498.pdf

Facione, P. A. (2020). Critical thinking: What it is and why it counts. In Insight assessment: Vol. XXVIII (Issue 1). https://www.insightassessment.com/wp-content/uploads/ia/pdf/whatwhy.pdf

Fisher, M. R., J., & Bandy, J. (2019). Assessing student learning. vanderbilt university center for teaching. https://cft.vanderbilt.edu/assessing-student-learning/

Goksu, I. (2021). Bibliometric mapping of mobile learning. Telematics and Informatics, 56(December 2019), 101491. https://doi.org/10.1016/j.tele.2020.101491

Haug, B. S., & Mork, S. M. (2021). Taking 21st century skills from vision to classroom: What teachers highlight as supportive professional development in the light of new demands from educational reforms. Teaching and Teacher Education, 100, 1–12. https://doi.org/10.1016/j.tate.2021.103286

Hernández-Torrano, D., & Ho, Y. S. (2021). A bibliometric analysis of publications in the web of science category of educational psychology in the last two decades. Psicologia Educativa, 27(2), 101–113. https://doi.org/10.5093/psed2021a19

Hidayah, A. N., Winingsih, P. H., & Amalia, A. F. (2020). Development of physics E-LKPD (electronic worksheets) using 3D pageflip based on problem based learning on balancing and rotation dynamics. Jurnal Ilmiah Pendidikan Fisika, 7(2), 36–43. https://jurnal.ustjogja.ac.id/index.php/COMPTON/article/download/9091/3693

Hsu, H. C. K., Wang, C. V., & Levesque-Bristol, C. (2019). Reexamining the impact of self-determination theory on learning outcomes in the online learning environment. Education and Information Technologies, 24(3), 2159–2174. https://doi.org/10.1007/s10639-019-09863-w

Khan, M. A., Pattnaik, D., Ashraf, R., Ali, I., Kumar, S., & Donthu, N. (2021). Value of special issues in the journal of business research: A bibliometric analysis. Journal of Business Research, 125(November 2020), 295–313. https://doi.org/10.1016/j.jbusres.2020.12.015

Kriswantoro, Kartowagiran, B., & Rohaeti, E. (2021). A critical thinking assessment model integrated with science process skills on chemistry for senior high school. European Journal of Educational Research, 10(1), 285–298. https://doi.org/10.12973/eu-jer.10.1.285

Lawson, A. E. (1978). The development and validation of a classroom test of formal reasoning. Journal of Research in Science Teaching, 15(1), 11–24. https://doi.org/10.1002/tea.3660150103

Lemke, M. R., Fischer, C., & Fritzer, G. (2004). A new model for assessment of mechanisms of behavior control and impulsivity. Pharmacopsychiatry, 36(5), 171. https://doi.org/10.1055/s-2003-825422

Miller, & Sundre, D. L. (2008). Achievement goal orientation toward general education versus overall coursework. The Journal of General Education, 57(3), 152–169. https://doi.org/10.1353/jge.0.0022

Nazhifah, N., Pasaribu, A., & Wiyono, K. (2022). Development of computer based test which is Integrated with Bengkulu local wisdom to measure the scientific literacy skills of junior high school students. Jurnal Penelitian & Pengembangan Pendidikan Fisika, 8(1), 45–56. https://doi.org/10.21009/1.08105

Neina, Q., & Qomariyah, U. (2021). The identification of merdeka belajar (freedom of learning) values through the development of portfolio assessment instruments in secondary schools. Proceedings of the 9th UNNES Virtual International Conference on English Language Teaching, Literature, and Translation, 1–9. https://doi.org/10.4108/eai.14-11-2020.2311108

Nuhfer, E., Cogan, C., Fleischer, S., Gaze, E., & Wirth, K. (2016). Random number simulations reveal wow random noise affects the measurements and graphical portrayals of self-assessed competency. Numeracy, 9(1), 1–27. https://doi.org/10.5038/1936-4660.9.1.4

Pham, H. H., Dong, T. K. T., Vuong, Q. H., Luong, D. H., Nguyen, T. T., Dinh, V. H., & Ho, M. T. (2021). A bibliometric review of research on international student mobilities in Asia with Scopus dataset between 1984 and 2019. Scientometrics, 126(6), 5201–5224. https://doi.org/10.1007/s11192-021-03965-4

Pratiwi, S. N., Cari, C., & Aminah, N. S. (2019). The 21st century science learning with student science literacy. Jurnal Materi Dan Pembelajaran Fisika (JMPF), 9, 34–42. https://doi.org/10.20961/jmpf.v9i1.31612

Quitadamo, I. J., Faiola, C. L., Johnson, J. E., & Kurtz, M. J. (2008). Community-based inquiry improves critical thinking in general education biology. CBE—Life Sciences Education, 7, 327–337. https://doi.org/10.1187/cbe.07

Rialp, A., Merigó, J. M., Cancino, C. A., & Urbano, D. (2019). Twenty-five years (1992–2016) of the international business review: A bibliometric overview. International Business Review, 28(6), 101587. https://doi.org/10.1016/j.ibusrev.2019.101587

Schildkamp, K., & Lai, M. K. (2013). Data-based decision making in education: Challenges and opportunities. https://doi.org/10.1007/978-94-007-4816-3_10

Schmid, M.,Brianza, E., Petko, D. 2020. Developing a short assessment instrument for Technological Pedagogical Content Knowledge (TPACK.xs) and comparing the factor structure of an integrative and a transformative model. Computer and Education,157,1-12. https://doi.org/10.1016/j.compedu.2020.103967

Si, H., Shi, J. G., Tang, D., Wen, S., Miao, W., & Duan, K. (2019). Application of the theory of planned behavior in environmental science: a comprehensive bibliometric analysis. International Journal of Environmental Research and Public Health, 16(15), 1–26. https://doi.org/10.3390/ijerph16152788

Spector, J. M., & Ma, S. (2019). Inquiry and critical thinking skills for the next generation: from artificial intelligence back to human intelligence. Smart Learning Environments, 6(1), 1–11. https://doi.org/10.1186/s40561-019-0088-z

Stanhope, L., Ziegler, L., Haque, T., Le, L., Vinces, M., Davis, G. K., Zieffler, A., Brodfuehrer, P., Preest, M., Belitsky, J. M., Umbanhowar, C., & Overvoorde, P. J. (2017). Development of a biological science quantitative reasoning exam (Biosquare). CBE Life Sciences Education, 16(4), 1–10. https://doi.org/10.1187/cbe.16-10-0301

Stein, B., & Haynes, A. (2011). Engaging faculty in the assessment and improvement of students’ critical thinking using the critical thinking assessment test. Change: The Magazine of Higher Learning, 43(2), 44–49. https://doi.org/10.1080/00091383.2011.550254

Stein, B., Haynes, A., Redding, M., Ennis, T., & Cecil, M. (2007). Assessing critical thinking in STEM and beyond. Innovations in E-Learning, Instruction Technology, Assessment, and Engineering Education, December 2014, 79–82. https://doi.org/10.1007/978-1-4020-6262-9

Stiggins, R. (2017). The perfect assessment system. ASDC. https://files.ascd.org/staticfiles/ascd/pdf/siteASCD/publications/books/PerfectAssessmentSystem_Stiggins.pdf

Suhaimi, N., Nur, S., & Mahmud, D. (2022). A Bibliometric analysis of climate change literacy between 2001 and 2021. Sustainability, 14(11940), 1–13. https://doi.org/10.3390/su141911940

Sundre, D. L. (2003). Assessment of quantitative reasoning to enhance educational quality (Issue Arpil). https://www.researchgate.net/publication/252163494_Assessment_of_Quantitative_Reasoning_to_Enhance_Educational_Quality/link/53ede95e0cf26b9b7dc63379/download?_tp=eyJjb250ZXh0Ijp7ImZpcnN0UGFnZSI6InB1YmxpY2F0aW9uIiwicGFnZSI6InB1YmxpY2F0aW9uIn19

Thu, H. L. T., Tran, T., Phuong, T. T. T., Tuyet, T. L. T., Huy, H. Le, & Thi, T. V. (2021). Two decades of stem education research in middle school: A bibliometrics analysis in scopus database (2000–2020). Education Sciences, 11(7). https://doi.org/10.3390/educsci11070353

Wirzal, M. D. H., Halim, N. S. A., Md Nordin, N. A. H., & Bustam, M. A. (2022). Metacognition in science learning: Bibliometric analysis of last two decades. Jurnal Penelitian Dan Pengkajian Ilmu Pendidikan: E-Saintika, 6(1), 43–60. https://doi.org/10.36312/esaintika.v6i1.665

Wirzal, M. D. H., Nordin, N. A. H. M., Bustam, M. A., & Joselevich, M. (2022). Bibliometric analysis of research on scientific literacy between 2018 and 2022: Science education subject. International Journal of Essential Competencies in Education, 1(2), 69–83. https://doi.org/10.36312/ijece.v1i2.1070

Wong, Y. J., Ho, M. H. R., Wang, S. Y., & Miller, I. S. K. (2017). Meta-analyses of the relationship between conformity to masculine norms and mental health-related outcomes. Journal of Counseling Psychology, 64(1), 80–93. https://doi.org/10.1037/cou0000176

Zhang, X. (2020). A bibliometric analysis of second language acquisition between 1997 and 2018. Studies in Second Language Acquisition, 42(1), 199–222. https://doi.org/10.1017/S0272263119000573

Zhu, J., & Liu, W. (2020). A tale of two databases: The use of Web of Science and Scopus in academic papers. Scientometrics, 123(1), 321–335. https://doi.org/10.1007/s11192-020-03387-8

Zupic, I., & Čater, T. (2015). Bibliometric methods in management and organization. Organizational Research Methods, 18(3), 429–472. https://doi.org/10.1177/1094428114562629

Downloads

Published

2024-03-29 — Updated on 2024-03-29

Issue

Section

Research Literacy and Trend