A Review of Rubrics in Education: Potential and Challenges
DOI:
https://doi.org/10.64420/ijitl.v2i1.197Keywords:
Rubric, Facilitate learning, Fair assessment, Validity, ReliabilityAbstract
Background: In recent years, the use of rubrics in educational assessment has gained significant attention due to their potential to enhance transparency, consistency, and learning outcomes. However, debates persist regarding their practical implementation, effectiveness, and limitations in various educational contexts. Objective: This study reviews the literature on rubrics, focusing on their role in learning facilitation, fair assessment, reliability, and validity. Method: The research follows a literature study approach, analyzing data from journals, scientific articles, and literature reviews. The analysis process includes data collection, classification, presentation, and conclusion drawing, with data triangulation used for validation. Results: Teachers should have the flexibility to decide whether to use rubrics. If adopted, rubrics must be well-designed and implemented by incorporating essential characteristics, avoiding common pitfalls, and continuously improving them to ensure fairness, reliability, and validity. Conclusion: Rubrics can be effective assessment tools when properly designed and implemented, but they require ongoing refinement to maximize their benefits. Contribution: This study provides insights into best practices for rubric use, highlighting key factors for effective implementation in educational settings.
References
Abram, P., Scarloss, B., Holthuis, N., Cohen, E., Lotan, R., & Schultz, S. E. (2002). The use of evaluation criteria to improve academic discussion in cooperative groups. Asia Pacific Journal of Education, 22(1), 16-27. https://doi.org/10.1080/0218879020220103
Adedoyin, C. A. (2013). Debate-proof grades: Experiences & challenges of using a grading rubric in a social welfare policy course. Journal of Teaching in Social Work, 33(2), 196-208. https://doi.org/10.1080/08841233.2013.774304
Andrade, H. G. (2001). The effects of instructional rubrics on learning to write. Current Issues in Education, 4(4), 1-22. https://scholarsarchive.library.albany.edu/etap_fac_scholar/6
Andrade, H. G. (2005). Teaching with rubrics: The good, the bad, and the ugly. College Teaching, 53(1), 27-30. http://www.jstor.org/stable/27559213
Andrade, H. L., Du, Y., & Wang, X. (2008). Putting rubrics to the test: The effect of a model, criteria generation, and rubric-referenced self-assessment on elementary school students' writing. Educational Measurement: Issues and Practice, 27(2), 3-13. https://doi.org/https://doi.org/10.1111/j.1745-3992.2008.00118.x
Andrade, H., & Du, Y. (2005). Student perspectives on rubric-referenced assessment. Practical Assessment, Research & Evaluation 10(3), 1-11. https://scholarsarchive.library.albany.edu/edpsych_fac_scholar/2
Baker, E. L., Abedi, J., Linn, R. L., & Niemi, D. (1996). Dimensionality and generalizability of domain-independent performance assessments. The Journal of Educational Research, 89(4), 197-205. http://www.jstor.org/stable/27542035
Bennett, C. (2016). Assessment rubrics: Thinking inside the boxes. Learning and Teaching, 9(1), 50-72. https://doi.org/10.3167/latiss.2016.090104
Bloxham, S., den-Outer, B., Hudson, J., & Price, M. (2016). Let’s stop the pretence of consistent marking: Exploring the multiple limitations of assessment criteria. Assessment & Evaluation in Higher Education, 41(3), 466-481. https://doi.org/10.1080/02602938.2015.1024607
Bolton, F. C. (2006). Rubrics and adult learners: Andragogy and assessment. Assessment Update, 18(3), 5-6. https://api.semanticscholar.org/CorpusID:113921569
Boud, D. (1989). The role of self‐assessment in student grading. Assessment & Evaluation in Higher Education, 14(1), 20-30. https://doi.org/10.1080/0260293890140103
Boud, D., & Soler, R. (2016). Sustainable assessment revisited. Assessment & Evaluation in Higher Education, 41(3), 400-413. https://doi.org/10.1080/02602938.2015.1018133
Brookhart, S. M. (2018). Appropriate criteria: Key to effective rubrics. Frontiers in Education, 3(22), 1-12. https://doi.org/10.3389/feduc.2018.00022
Bryant, C. L., Maarouf, S., Burcham, J., & Greer, D. (2016). The examination of a teacher candidate assessment rubric: A confirmatory factor analysis. Teaching and Teacher Education, 57, 79-96. https://doi.org/https://doi.org/10.1016/j.tate.2016.03.012
Chan, Z., & Ho, S. (2019). Good and bad practices in rubrics: The perspectives of students and educators. Assessment & Evaluation in Higher Education, 44(4), 533-545. https://doi.org/10.1080/02602938.2018.1522528
Chapman, V. G., & Inman, M. D. (2009). A conundrum: Rubrics or creativity/metacognitive development? Educational Horizons, 87(3), 198–202. https://files.eric.ed.gov/fulltext/EJ849020.pdf
Cohen, E. G., Lotan, R. A., Abram, P. L., Scarloss, B. A., & Schultz, S. E. (2002). Can Groups Learn? Teachers College Record, 104(6), 1045-1068. https://doi.org/10.1111/1467-9620.00196
Covill, A. E. (2012). College students' use of a writing rubric: Effect on quality of writing, self-efficacy, and writing practices.” Journal of Writing Assessment, 5(1), 1-16. https://escholarship.org/uc/item/5f79g6kz
Dunbar, N. E., Brooks, C. F., & Kubicka-Miller, T. (2006). Oral communication skills in higher education: Using a performance-based evaluation rubric to assess communication skills. Innovative Higher Education, 31(2), 115-128. https://doi.org/10.1007/s10755-006-9012-x
Falchikov, N. (1986). Product comparisons and process benefits of collaborative peer group and self-assessments. Assessment & Evaluation in Higher Education, 11(2), 146-166. https://doi.org/10.1080/0260293860110206
Galván-Sánchez, I., Verano-Tacoronte, D., González-Betancor, S. M., Fernández-Monroy, M., & Bolívar-Cruz, A. (2017). Assessing oral presentation skills in electrical engineering: Developing a valid and reliable rubric. International Journal of Electrical Engineering & Education, 54(1), 17-34. https://doi.org/10.1177/0020720916659501
Gray, J. S., Connolly, J. P., & Brown, M. A. (2019). Measuring intercultural knowledge and competence in college essays: Does a performance-based rubric have construct validity? Studies in Educational Evaluation, 62, 142-148. https://doi.org/https://doi.org/10.1016/j.stueduc.2019.05.007
Green, R., & Bowser, M. (2006). Observations from the Field. Journal of Library Administration, 45(1-2), 185-202. https://doi.org/10.1300/J111v45n01_10
Hafner, J., & Hafner, P. (2003). Quantitative analysis of the rubric as an assessment tool: an empirical study of student peer‐group rating. International Journal of Science Education, 25(12), 1509-1528. https://doi.org/10.1080/0950069022000038268
Hendry, G. D., Armstrong, S., & Bromberger, N. (2012). Implementing standards-based assessment effectively: incorporating discussion of exemplars into classroom teaching. Assessment & Evaluation in Higher Education, 37(2), 149-161. https://doi.org/10.1080/02602938.2010.515014
Higgins, R., Hartley, P., & Skelton, A. (2002). The conscientious consumer: Reconsidering the role of assessment feedback in student learning. Studies in Higher Education, 27(1), 53-64. https://doi.org/10.1080/03075070120099368
Huang, S. C. (2012). Like a bell responding to a striker: Instruction contingent on assessment. English Teaching: Practice and Critique, 11(4), 99–119. https://api.semanticscholar.org/CorpusID:56007690
Jeong, H. (2015). Rubrics in the classroom: Do teachers really follow them? Language Testing in Asia, 5(1), 6. https://doi.org/10.1186/s40468-015-0013-5
Johnson, C. S., & Gelfand, S. (2013). Self-assessment and writing quality. Academic Research International, 4(4), 571–580. http://www.savap.org.pk/journals/ARInt./Vol.4(4)/2013(4.4-60).pdf
Jonsson, A., & Svingby, G. (2007). The use of scoring rubrics: reliability, validity and educational consequences. Educational Research Review, 2(2), 130-144. https://doi.org/https://doi.org/10.1016/j.edurev.2007.05.002
Kane, M. T. (2001). Current concerns in validity theory. Journal of Educational Measurement, 38(4), 319-342. https://doi.org/10.1111/j.1745-3984.2001.tb01130.x
Kavanagh, S., & Luxton-Reilly, A. (2016). Rubrics used in peer assessment. Proceedings of the Australasian Computer Science Week Multiconference, Canberra, Australia. https://doi.org/10.1145/2843043.2843347
Klein, S. P., Stecher, B. M., Shavelson, R. J., McCaffrey, D., Ormseth, T., Bell, R. M., Comfort, K., & Othman, A. R. (1998). Analytic versus holistic scoring of science performance tasks. Applied Measurement in Education, 11(2), 121-137. https://doi.org/10.1207/s15324818ame1102_1
Knoch, U., Read, J., & von Randow, J. (2007). Re-training writing raters online: How does it compare with face-to-face training? Assessing Writing, 12(1), 26-43. https://doi.org/https://doi.org/10.1016/j.asw.2007.04.001
Kutlu, Ö., Bilican, S., & Yıldırım, Ö. (2010). A study on the primary school teachers’ attitudes towards rubrics with reference to different variables. Procedia - Social and Behavioral Sciences, 2(2), 5398-5402. https://doi.org/https://doi.org/10.1016/j.sbspro.2010.03.880
Lea, M. R., & Street, B. V. (1998). Student writing in higher education: An academic literacies approach. Studies in Higher Education, 23(2), 157-172. https://doi.org/10.1080/03075079812331380364
Li, J., & Lindsey, P. (2015). Understanding variations between student and teacher application of rubrics. Assessing Writing, 26, 67-79. https://doi.org/10.1016/j.asw.2015.07.003
Liew, C. P., Puteh, M., & Hamzah, S. H. (2020). Comparative study of engineering design project assessment rubrics to address the Washington Accord’s complexity attributes. ASEAN Journal of Engineering Education, 4(1), 71-94. https://doi.org/10.11113/ajee2020.4n1.21
Lumley, T., & McNamara, T. F. (1995). Rater characteristics and rater bias: implications for training. Language Testing, 12(1): 54-71. https://doi.org/10.1177/026553229501200104
Malouff, J. (2008). Bias in grading. College Teaching, 56(3), 191-192. https://doi.org/10.3200/CTCH.56.3.191-192
Mark, R., George, H., & Alan, C. (2006). Individualising students’ scores using blind and holistic peer assessment. Engineering Education, 1(1), 50-60. https://doi.org/10.11120/ened.2006.01010050
Messick, S. (1995). Validity of psychological assessment: Validation of inferences from persons' responses and performances as scientific inquiry into score meaning. American Psychologist, 50(9), 741-749. https://doi.org/10.1037/0003-066X.50.9.741
Moskal, B. M. (2000). Scoring rubrics: What, when and how? Practical Assessment, Research, and Evaluation, 7(3), 1-5. https://doi.org/10.7275/a5vq-7q66
Moskal, B. M., & Leydens, J. A. (2000). Scoring rubric development: Validity and reliability. Practical Assessment, Research, and Evaluation, 7(7), 1-6. https://doi.org/10.7275/q7rm-gg74
Murillo, F. J., & Hidalgo, N. (2017). Students’ conceptions about a fair assessment of their learning. Studies in Educational Evaluation, 53: 10-16. https://doi.org/10.1016/j.stueduc.2017.01.001
Oakleaf, M. (2008). Dangers and opportunities: A conceptual map of information literacy assessment approaches. portal: Libraries and the Academy, 8(3): 233-253. https://doi.org/10.1353/pla.0.0011
Orsmond, P., Merry, S., & Reiling, K. (1996). The importance of marking criteria in the use of peer assessment. Assessment & Evaluation in Higher Education, 21(3), 239-250. https://doi.org/10.1080/0260293960210304
Panadero, E., & Jonsson, A. (2020). A critical review of the arguments against the use of rubrics. Educational Research Review, 30: 100329. https://doi.org/https://doi.org/10.1016/j.edurev.2020.100329
Panadero, E., & Romero, M. (2014). To rubric or not to rubric? The effects of self-assessment on self-regulation, performance and self-efficacy. Assessment in Education: Principles, Policy & Practice, 21(2), 133-148. https://doi.org/10.1080/0969594X.2013.877872
Panadero, E., Romero, M., & Strijbos, J. W. (2013). The impact of a rubric and friendship on peer assessment: Effects on construct validity, performance, and perceptions of fairness and comfort. Studies in Educational Evaluation, 39(4), 195-203. https://doi.org/https://doi.org/10.1016/j.stueduc.2013.10.005
Parke, C. S. (2001). An approach that examines sources of misfit to improve performance assessment items and rubrics. Educational Assessment, 7(3), 201-225. https://doi.org/10.1207/S15326977EA0703_02
Peeters, M. J., Sahloff, E. G., & Stone, G. E. (2010). A standardized rubric to evaluate student presentations. Am J Pharm Educ 74(9), 171. https://doi.org/10.5688/aj7409171
Pepper, M. B., & Pathak, S. (2008). Classroom contribution: What do students perceive as fair assessment? Journal of Education for Business, 83(6), 360-368. https://doi.org/10.3200/JOEB.83.6.360-368
Petkov, D., & Petkova, O. (2006). Development of scoring rubrics for IS Projects as an assessment tool. Issues in Informing Science and Information Technology, 3, 499-510. https://doi.org/10.28945/910
Popham, W. J. (1997). What’s wrong - and what’s right - with rubrics. Educational Leadership, 55(2), 72. https://eric.ed.gov/?id=EJ552014
Rafilson, F. (1990). The case for validity generalization. Practical Assessment, Research, and Evaluation, 2(13), 1-3. https://doi.org/10.7275/b51w-5m96
Reddy, Y. M., & Andrade, H. (2010). A review of rubric use in higher education. Assessment & Evaluation in Higher Education, 35(4), 435-448. https://doi.org/10.1080/02602930902862859
Reynolds-Keefer, L. (2010). Rubric-referenced assessment in teacher preparation: An opportunity to learn by using. Practical Assessment, Research, and Evaluation, 15(8): 1-9. https://doi.org/10.7275/psk5-mf68
Rezaei, A. R., & Lovorn, M. (2010). Reliability and validity of rubrics for assessment through writing. Assessing Writing, 15(1), 18-39. https://doi.org/10.1016/j.asw.2010.01.003
Sadler, D. R. (1989). Formative assessment and the design of instructional systems. Instructional Science, 18(2), 119-144. https://doi.org/10.1007/BF00117714
Sadler, D. R. (2009). Indeterminacy in the use of preset criteria for assessment and grading. Assessment & Evaluation in Higher Education, 34(2), 159-179. https://doi.org/10.1080/02602930801956059
Sadler, D. R. (2014). The futility of attempting to codify academic achievement standards. Higher Education, 67(3), 273-288. https://doi.org/10.1007/s10734-013-9649-1
Sambell, K., McDowell, L., & Brown, S. (1997). But is it fair?: An exploratory study of student perceptions of the consequential validity of assessment. Studies in Educational Evaluation, 23(4), 349-371. https://doi.org/https://doi.org/10.1016/S0191-491X(97)86215-3
Scott, S., Webber, C. F., Lupart, J. L., Aitken, N., & Scott, D. E. (2014). Fair and equitable assessment practices for all students. Assessment in Education: Principles, Policy & Practice, 21(1), 52-70. https://doi.org/10.1080/0969594X.2013.776943
Simon, M., & Forgette-Giroux, R. (2001). A rubric for scoring postsecondary academic skills. Practical Assessment, Research, and Evaluation, 7(18), 1-4. https://doi.org/10.7275/bh4d-me80
Stellmack, M. A., Konheim-Kalkstein, Y. L., Manor, J. E., Massey, A. R., and Schmitz, J. A. P. (2009). An assessment of reliability and validity of a rubric for grading APA-style introductions. Teaching of Psychology, 36(2), 102-107. https://doi.org/10.1080/00986280902739776
Stemler, S. E. (2004). A comparison of consensus, consistency, and measurement approaches to estimating interrater reliability. Practical Assessment, Research, and Evaluation, 9(4), 1-11. https://doi.org/10.7275/96jp-xz07
Sundeen, T. H. (2014). Instructional rubrics: Effects of presentation options on writing quality. Assessing Writing, 21, 74-88. https://doi.org/https://doi.org/10.1016/j.asw.2014.03.003
Tierney, R., & Simon, M. (2004). What's still wrong with rubrics: Focusing on the consistency of performance criteria across scale levels. Practical Assessment, Research, and Evaluation, 9(2), 1-7. https://doi.org/10.7275/jtvt-wg68
Van Helvoort, A. A. J. (2012). How adult students in information studies use a scoring rubric for the development of their information literacy skills. The Journal of Academic Librarianship, 38(3), 165-171. https://doi.org/https://doi.org/10.1016/j.acalib.2012.03.016
Weems, G. H., & Onwuegbuzie, A. J. (2001). The impact of midpoint responses and reverse coding on survey data. Measurement and Evaluation in Counseling and Development, 34(3), 166-176. https://doi.org/10.1080/07481756.2002.12069033
Williams, L., & Rink, J. E. (2003). Teacher competency using observational scoring rubrics. Journal of Teaching in Physical Education, 22: 552-572. http://journals.humankinetics.com/jtpe-contents
Wilson, M. 2007. Why I won't be using rubrics to respond to students' writing. The English Journal, 96(4), 62-66. https://doi.org/10.2307/30047167
Zhang, B., Xiao, Y., & Luo, J. (2015). Rater reliability and score discrepancy under holistic and analytic scoring of second language writing. Language Testing in Asia, 5(1), 5. https://doi.org/10.1186/s40468-015-0014-4
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Jen Hua Ling

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Authors who publish with this journal agree to the following terms: (1) Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution-ShareAlike 4.0 International. that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal; (2) Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal; (3) Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work.