Konsistensi dan Dependabilitas Penilaian Microteaching pada Program Studi Pendidikan Kimia : Teori Generalisabilitas
Consistency and Dependability of Microteaching Assessment in the Chemistry Education Study Program: Generalisability Theory
Abstract
Subjektivitas dan kurang konsistennya penilai/rater dalam proses penyekoran merupakan kritik yang umum ditujukan pada penilaian dalam pembelajaran. Oleh karena itu, artikel ini menyajikan hasil reliabilitas penilaian pada matakuliah Microteaching menggunakan tori generalisabilitas dengan desain px (i:r). Data yang dikumpulkan merupakan nilai ujian mahasiswa baik ujian pada nilai pengamatan di kelas, ujian tengah semester, ujian akhir semester. Responden berjumlah 30 orang, dengan 5 jenis tes dan dinilai oleh 3 rater yang berbeda. Analisis data pada pada artikel ini menggunakan program R untuk menghitung relative error varians; mendapatkan koefisien generalizabilitas dan coefficient dependability dari hasil tes secara empirik. Hasil penelitian didapatkan nilai eror terbesar terdapat pada person dan rater 1 dengan nilai 28.6 % dan 18%, nilai koefisien generalizabilitas sebesar 0.93 dan nilai koefisien dependability sebesar 0.69. Sehingga berdasarkan teori D maka dilakukan modifikasi dengan menambahkan jumlah rater agar nilai koefisien dependability menjadi > 0.7.
Downloads
References
L. Lubna, “Isu-Isu Pendidikan Di Indonesia: Inovasi Kurikulum Dan Peningkatan Profesionalitas Guru,” Society, vol. 5, no. 2, pp. 15–25, Oct. 2014, https://doi.org/10.20414/society.v5i2.1455.
T. Andayani and F. Madani, “Peran Penilaian Pembelajaran Dalam Meningkatkan Prestasi Siswa di Pendidikan Dasar,” Jurnal Educatio FKIP UNMA, vol. 9, no. 2, pp. 924–930, Jun. 2023, https://doi.org/10.31949/educatio.v9i2.4402.
V. S. Damaianti, Y. Abidin, and R. Rahma, “Higher order thinking skills-based reading literacy assessment instrument: An Indonesian context,” Indonesian Journal of Applied Linguistics, vol. 10, no. 2, pp. 513–525, Oct. 2020, https://doi.org/10.17509/ijal.v10i2.28600.
J. L. Ferreira Neto, L. G. M. F. Duarte, and C. M. F. Penido, “Avaliação E Processos De Subjetivação Na Atenção Básica À Saúde: Avaliação E Subjetivação,” Psicol Estud, vol. 27, Feb. 2022, https://doi.org/10.4025/psicolestud.v27i0.48663.
S. Salsabila, “Sistem Pendukung Keputusan Penilaian Tenaga Kependidikan Dengan Metode Fuzzy Analytic Hierarchy Process,” Journal of Information Technology, vol. 4, no. 2, pp. 01–06, May 2023, https://doi.org/10.47292/joint.v4i2.73.
N. K. Park, M. Y. Chun, and J. Lee, “Revisiting Individual Creativity Assessment: Triangulation in Subjective and Objective Assessment Methods,” Creat Res J, vol. 28, no. 1, pp. 1–10, Jan. 2016, https://doi.org/10.1080/10400419.2016.1125259.
K. Grint, “What’s Wrong With Performance Appraisals? A Critique and A Suggestion,” Human Resource Management Journal, vol. 3, no. 3, pp. 61–77, Mar. 1993, https://doi.org/10.1111/j.1748-8583.1993.tb00316.x.
W. Priatna and R. Purnomo, “Implementasi Fuzzy Inference System Metode Sugeno Pada Aplikasi Penilaian Kinerja Dosen,” Techno.Com, vol. 19, no. 3, pp. 245–261, Aug. 2020, https://doi.org/10.33633/tc.v19i3.3638.
B. Bulut, H. Ulu, and A. Kan, “Multimodal Literacy Scale: A Study of Validity and Reliability,” Egitim Arastirmalari - Eurasian Journal of Educational Research, vol. 15, no. 61, pp. 45–60, 2015, https://doi.org/10.14689/ejer.2015.61.3.
A. Kan and O. Bulut, “Crossed random-effect modeling: Examining the effects of teacher experience and rubric use in performance assessments,” Eurasian Journal of Educational Research, no. 57, pp. 1–28, Dec. 2014, https://doi.org/10.14689/ejer.2014.57.4.
A. Jonsson and G. Svingby, “The use of scoring rubrics: Reliability, validity and educational consequences,” Educ Res Rev, vol. 2, no. 2, pp. 130–144, Jan. 2007, https://doi.org/10.1016/j.edurev.2007.05.002.
J. Stuhlmann, C. Daniel, A. Dellinger, R. Kenton, and T. Powers, “A Generalizability Study Of The Effects Of Training On Teachers’ Abilities To Rate Children’s Writing Using A Rubric,” Read Psychol, vol. 20, no. 2, pp. 107–127, Jun. 1999, https://doi.org/10.1080/027027199278439.
R. Smit, P. Bachmann, V. Blum, T. Birri, and K. Hess, “Effects of a rubric for mathematical reasoning on teaching and learning in primary school,” Instr Sci, vol. 45, no. 5, pp. 603–622, Oct. 2017, https://doi.org/10.1007/s11251-017-9416-2.
W. D. Shafer, G. Swanson, N. Bene, and G. Newberry, “Effects of Teacher Knowledge of Rubrics on Student Achievement in Four Content Areas,” Applied Measurement in Education, vol. 14, no. 2, pp. 151–170, Apr. 2001, https://doi.org/10.1207/S15324818AME1402_3.
T. Lumley, “Perceptions of Language-trained Raters and Occupational Experts in a Test of Occupational English Language Profficiency,” English for Specific Purposes, vol. 17, no. 4, pp. 347–367, Oct. 1998, https://doi.org/10.1016/S0889-4906(97)00016-1.
T. Eckes, “Rater types in writing performance assessments: A classification approach to rater variability,” Language Testing, vol. 25, no. 2, pp. 155–185, Apr. 2008, https://doi.org/10.1177/0265532207086780.
H. L. Andrade and Y. Du, “Student Perspectives on Rubric-Referenced Assessment Student Perspectives on Rubric-Referenced Assessment,” 2005. [Online]. Available: https://scholarsarchive.library.albany.edu/edpsych_fac_scholarhttps://scholarsarchive.library.albany.edu/edpsych_fac_scholar/2
R. L. Brennan, “Generalizability Theory and Classical Test Theory,” Applied Measurement in Education, vol. 24, no. 1, pp. 1–21, Dec. 2010, https://doi.org/10.1080/08957347.2011.532417.
N. M. Webb and R. J. Shavelson, “Generalizability Theory: Overview,” in Encyclopedia of Statistics in Behavioral Science, Wiley, 2005. https://doi.org/10.1002/0470013192.bsa703.
M. Sudaryanto, K. Saddhono, and Lina, “Applying Item Responses Theory For Measuring Student’s Ability In Academic Speaking,” Humanities & Social Sciences Reviews, vol. 8, no. 2, pp. 305–312, Mar. 2020, https://doi.org/10.18510/hssr.2020.8234.
M. H. Zubaidillah, “Prinsip dan Alat Evaluasi dalam Pendidikan,” Jul. 17, 2018. https://doi.org/10.31219/osf.io/4tgfm.
F. Razi, “Konsep Dasar Evaluasi Pembelajaran,” Mar. 12, 2021. https://doi.org/10.31219/osf.io/nmua2.
L. G. Otaya, H. Anwar, and R. T. Husain, “Estimating the Students’ Skill in Reciting and Writing Alqur’an at Faculty of Tarbiyah and Teacher Training IAIN Sultan Amai Gorontalo,” Nadwa, vol. 1, no. 1, p. 75, Aug. 2019, https://doi.org/10.21580/nw.2019.1.1.3590.
J. Sánchez-Meca, J. A. López-Pina, and J. A. López-López, “Una revisión de los estudios meta-analíticos de generalización de la fiabilidad,” Escritos de Psicología - Psychological Writings, vol. 2, no. 1, pp. 110–121, Dec. 2008, https://doi.org/10.24310/espsiescpsi.v2i1.13365.
P. E. Clayson, K. A. Carbine, S. Baldwin, J. A. Olsen, and M. J. Larson, “Using Generalizability Theory and the ERP Reliability Analysis (ERA) Toolbox for Assessing Test-Retest Reliability of ERP Scores Part 1: Algorithms, Framework, and Implementation,” Jul. 07, 2020. https://doi.org/10.31234/osf.io/kcven.
Brennan, R. L. (2001). Statistics for Social Science and Public Policy: Generalizability Theory. Springer Verlag
Copyright (c) 2024 Dewi Lestarani, Raden Rosnawati, Arvinda Ceniorita Lalang, Arvinda Ceniorita Lalang, Maria Ulfah
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
- Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Atribusi-Non Commercial-Share Alike (CC BY-NC-SA).
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Every publication (printed/electronic) are open access for educational purposes, research, and library. Other than the aims mentioned above, the editorial board is not responsible for copyright violation.