A Review of Rubrics in Education: Potential and Challenges

Authors

  • Jen Hua Ling University of Technology Sarawak

DOI:

https://doi.org/10.64420/ijitl.v2i1.197

Keywords:

Rubric, Facilitate learning, Fair assessment, Validity, Reliability

Abstract

Background: In recent years, the use of rubrics in educational assessment has gained significant attention due to their potential to enhance transparency, consistency, and learning outcomes. However, debates persist regarding their practical implementation, effectiveness, and limitations in various educational contexts. Objective: This study reviews the literature on rubrics, focusing on their role in learning facilitation, fair assessment, reliability, and validity. Method: The research follows a literature study approach, analyzing data from journals, scientific articles, and literature reviews. The analysis process includes data collection, classification, presentation, and conclusion drawing, with data triangulation used for validation. Results: Teachers should have the flexibility to decide whether to use rubrics. If adopted, rubrics must be well-designed and implemented by incorporating essential characteristics, avoiding common pitfalls, and continuously improving them to ensure fairness, reliability, and validity. Conclusion: Rubrics can be effective assessment tools when properly designed and implemented, but they require ongoing refinement to maximize their benefits. Contribution: This study provides insights into best practices for rubric use, highlighting key factors for effective implementation in educational settings.

References

Abram, P., Scarloss, B., Holthuis, N., Cohen, E., Lotan, R., & Schultz, S. E. (2002). The use of evaluation criteria to improve academic discussion in cooperative groups. Asia Pacific Journal of Education, 22(1), 16-27. https://doi.org/10.1080/0218879020220103

Adedoyin, C. A. (2013). Debate-proof grades: Experiences & challenges of using a grading rubric in a social welfare policy course. Journal of Teaching in Social Work, 33(2), 196-208. https://doi.org/10.1080/08841233.2013.774304

Andrade, H. G. (2001). The effects of instructional rubrics on learning to write. Current Issues in Education, 4(4), 1-22. https://scholarsarchive.library.albany.edu/etap_fac_scholar/6

Andrade, H. G. (2005). Teaching with rubrics: The good, the bad, and the ugly. College Teaching, 53(1), 27-30. http://www.jstor.org/stable/27559213

Andrade, H. L., Du, Y., & Wang, X. (2008). Putting rubrics to the test: The effect of a model, criteria generation, and rubric-referenced self-assessment on elementary school students' writing. Educational Measurement: Issues and Practice, 27(2), 3-13. https://doi.org/https://doi.org/10.1111/j.1745-3992.2008.00118.x

Andrade, H., & Du, Y. (2005). Student perspectives on rubric-referenced assessment. Practical Assessment, Research & Evaluation 10(3), 1-11. https://scholarsarchive.library.albany.edu/edpsych_fac_scholar/2

Baker, E. L., Abedi, J., Linn, R. L., & Niemi, D. (1996). Dimensionality and generalizability of domain-independent performance assessments. The Journal of Educational Research, 89(4), 197-205. http://www.jstor.org/stable/27542035

Bennett, C. (2016). Assessment rubrics: Thinking inside the boxes. Learning and Teaching, 9(1), 50-72. https://doi.org/10.3167/latiss.2016.090104

Bloxham, S., den-Outer, B., Hudson, J., & Price, M. (2016). Let’s stop the pretence of consistent marking: Exploring the multiple limitations of assessment criteria. Assessment & Evaluation in Higher Education, 41(3), 466-481. https://doi.org/10.1080/02602938.2015.1024607

Bolton, F. C. (2006). Rubrics and adult learners: Andragogy and assessment. Assessment Update, 18(3), 5-6. https://api.semanticscholar.org/CorpusID:113921569

Boud, D. (1989). The role of self‐assessment in student grading. Assessment & Evaluation in Higher Education, 14(1), 20-30. https://doi.org/10.1080/0260293890140103

Boud, D., & Soler, R. (2016). Sustainable assessment revisited. Assessment & Evaluation in Higher Education, 41(3), 400-413. https://doi.org/10.1080/02602938.2015.1018133

Brookhart, S. M. (2018). Appropriate criteria: Key to effective rubrics. Frontiers in Education, 3(22), 1-12. https://doi.org/10.3389/feduc.2018.00022

Bryant, C. L., Maarouf, S., Burcham, J., & Greer, D. (2016). The examination of a teacher candidate assessment rubric: A confirmatory factor analysis. Teaching and Teacher Education, 57, 79-96. https://doi.org/https://doi.org/10.1016/j.tate.2016.03.012

Chan, Z., & Ho, S. (2019). Good and bad practices in rubrics: The perspectives of students and educators. Assessment & Evaluation in Higher Education, 44(4), 533-545. https://doi.org/10.1080/02602938.2018.1522528

Chapman, V. G., & Inman, M. D. (2009). A conundrum: Rubrics or creativity/metacognitive development? Educational Horizons, 87(3), 198–202. https://files.eric.ed.gov/fulltext/EJ849020.pdf

Cohen, E. G., Lotan, R. A., Abram, P. L., Scarloss, B. A., & Schultz, S. E. (2002). Can Groups Learn? Teachers College Record, 104(6), 1045-1068. https://doi.org/10.1111/1467-9620.00196

Covill, A. E. (2012). College students' use of a writing rubric: Effect on quality of writing, self-efficacy, and writing practices.” Journal of Writing Assessment, 5(1), 1-16. https://escholarship.org/uc/item/5f79g6kz

Dunbar, N. E., Brooks, C. F., & Kubicka-Miller, T. (2006). Oral communication skills in higher education: Using a performance-based evaluation rubric to assess communication skills. Innovative Higher Education, 31(2), 115-128. https://doi.org/10.1007/s10755-006-9012-x

Falchikov, N. (1986). Product comparisons and process benefits of collaborative peer group and self-assessments. Assessment & Evaluation in Higher Education, 11(2), 146-166. https://doi.org/10.1080/0260293860110206

Galván-Sánchez, I., Verano-Tacoronte, D., González-Betancor, S. M., Fernández-Monroy, M., & Bolívar-Cruz, A. (2017). Assessing oral presentation skills in electrical engineering: Developing a valid and reliable rubric. International Journal of Electrical Engineering & Education, 54(1), 17-34. https://doi.org/10.1177/0020720916659501

Gray, J. S., Connolly, J. P., & Brown, M. A. (2019). Measuring intercultural knowledge and competence in college essays: Does a performance-based rubric have construct validity? Studies in Educational Evaluation, 62, 142-148. https://doi.org/https://doi.org/10.1016/j.stueduc.2019.05.007

Green, R., & Bowser, M. (2006). Observations from the Field. Journal of Library Administration, 45(1-2), 185-202. https://doi.org/10.1300/J111v45n01_10

Hafner, J., & Hafner, P. (2003). Quantitative analysis of the rubric as an assessment tool: an empirical study of student peer‐group rating. International Journal of Science Education, 25(12), 1509-1528. https://doi.org/10.1080/0950069022000038268

Hendry, G. D., Armstrong, S., & Bromberger, N. (2012). Implementing standards-based assessment effectively: incorporating discussion of exemplars into classroom teaching. Assessment & Evaluation in Higher Education, 37(2), 149-161. https://doi.org/10.1080/02602938.2010.515014

Higgins, R., Hartley, P., & Skelton, A. (2002). The conscientious consumer: Reconsidering the role of assessment feedback in student learning. Studies in Higher Education, 27(1), 53-64. https://doi.org/10.1080/03075070120099368

Huang, S. C. (2012). Like a bell responding to a striker: Instruction contingent on assessment. English Teaching: Practice and Critique, 11(4), 99–119. https://api.semanticscholar.org/CorpusID:56007690

Jeong, H. (2015). Rubrics in the classroom: Do teachers really follow them? Language Testing in Asia, 5(1), 6. https://doi.org/10.1186/s40468-015-0013-5

Johnson, C. S., & Gelfand, S. (2013). Self-assessment and writing quality. Academic Research International, 4(4), 571–580. http://www.savap.org.pk/journals/ARInt./Vol.4(4)/2013(4.4-60).pdf

Jonsson, A., & Svingby, G. (2007). The use of scoring rubrics: reliability, validity and educational consequences. Educational Research Review, 2(2), 130-144. https://doi.org/https://doi.org/10.1016/j.edurev.2007.05.002

Kane, M. T. (2001). Current concerns in validity theory. Journal of Educational Measurement, 38(4), 319-342. https://doi.org/10.1111/j.1745-3984.2001.tb01130.x

Kavanagh, S., & Luxton-Reilly, A. (2016). Rubrics used in peer assessment. Proceedings of the Australasian Computer Science Week Multiconference, Canberra, Australia. https://doi.org/10.1145/2843043.2843347

Klein, S. P., Stecher, B. M., Shavelson, R. J., McCaffrey, D., Ormseth, T., Bell, R. M., Comfort, K., & Othman, A. R. (1998). Analytic versus holistic scoring of science performance tasks. Applied Measurement in Education, 11(2), 121-137. https://doi.org/10.1207/s15324818ame1102_1

Knoch, U., Read, J., & von Randow, J. (2007). Re-training writing raters online: How does it compare with face-to-face training? Assessing Writing, 12(1), 26-43. https://doi.org/https://doi.org/10.1016/j.asw.2007.04.001

Kutlu, Ö., Bilican, S., & Yıldırım, Ö. (2010). A study on the primary school teachers’ attitudes towards rubrics with reference to different variables. Procedia - Social and Behavioral Sciences, 2(2), 5398-5402. https://doi.org/https://doi.org/10.1016/j.sbspro.2010.03.880

Lea, M. R., & Street, B. V. (1998). Student writing in higher education: An academic literacies approach. Studies in Higher Education, 23(2), 157-172. https://doi.org/10.1080/03075079812331380364

Li, J., & Lindsey, P. (2015). Understanding variations between student and teacher application of rubrics. Assessing Writing, 26, 67-79. https://doi.org/10.1016/j.asw.2015.07.003

Liew, C. P., Puteh, M., & Hamzah, S. H. (2020). Comparative study of engineering design project assessment rubrics to address the Washington Accord’s complexity attributes. ASEAN Journal of Engineering Education, 4(1), 71-94. https://doi.org/10.11113/ajee2020.4n1.21

Lumley, T., & McNamara, T. F. (1995). Rater characteristics and rater bias: implications for training. Language Testing, 12(1): 54-71. https://doi.org/10.1177/026553229501200104

Malouff, J. (2008). Bias in grading. College Teaching, 56(3), 191-192. https://doi.org/10.3200/CTCH.56.3.191-192

Mark, R., George, H., & Alan, C. (2006). Individualising students’ scores using blind and holistic peer assessment. Engineering Education, 1(1), 50-60. https://doi.org/10.11120/ened.2006.01010050

Messick, S. (1995). Validity of psychological assessment: Validation of inferences from persons' responses and performances as scientific inquiry into score meaning. American Psychologist, 50(9), 741-749. https://doi.org/10.1037/0003-066X.50.9.741

Moskal, B. M. (2000). Scoring rubrics: What, when and how? Practical Assessment, Research, and Evaluation, 7(3), 1-5. https://doi.org/10.7275/a5vq-7q66

Moskal, B. M., & Leydens, J. A. (2000). Scoring rubric development: Validity and reliability. Practical Assessment, Research, and Evaluation, 7(7), 1-6. https://doi.org/10.7275/q7rm-gg74

Murillo, F. J., & Hidalgo, N. (2017). Students’ conceptions about a fair assessment of their learning. Studies in Educational Evaluation, 53: 10-16. https://doi.org/10.1016/j.stueduc.2017.01.001

Oakleaf, M. (2008). Dangers and opportunities: A conceptual map of information literacy assessment approaches. portal: Libraries and the Academy, 8(3): 233-253. https://doi.org/10.1353/pla.0.0011

Orsmond, P., Merry, S., & Reiling, K. (1996). The importance of marking criteria in the use of peer assessment. Assessment & Evaluation in Higher Education, 21(3), 239-250. https://doi.org/10.1080/0260293960210304

Panadero, E., & Jonsson, A. (2020). A critical review of the arguments against the use of rubrics. Educational Research Review, 30: 100329. https://doi.org/https://doi.org/10.1016/j.edurev.2020.100329

Panadero, E., & Romero, M. (2014). To rubric or not to rubric? The effects of self-assessment on self-regulation, performance and self-efficacy. Assessment in Education: Principles, Policy & Practice, 21(2), 133-148. https://doi.org/10.1080/0969594X.2013.877872

Panadero, E., Romero, M., & Strijbos, J. W. (2013). The impact of a rubric and friendship on peer assessment: Effects on construct validity, performance, and perceptions of fairness and comfort. Studies in Educational Evaluation, 39(4), 195-203. https://doi.org/https://doi.org/10.1016/j.stueduc.2013.10.005

Parke, C. S. (2001). An approach that examines sources of misfit to improve performance assessment items and rubrics. Educational Assessment, 7(3), 201-225. https://doi.org/10.1207/S15326977EA0703_02

Peeters, M. J., Sahloff, E. G., & Stone, G. E. (2010). A standardized rubric to evaluate student presentations. Am J Pharm Educ 74(9), 171. https://doi.org/10.5688/aj7409171

Pepper, M. B., & Pathak, S. (2008). Classroom contribution: What do students perceive as fair assessment? Journal of Education for Business, 83(6), 360-368. https://doi.org/10.3200/JOEB.83.6.360-368

Petkov, D., & Petkova, O. (2006). Development of scoring rubrics for IS Projects as an assessment tool. Issues in Informing Science and Information Technology, 3, 499-510. https://doi.org/10.28945/910

Popham, W. J. (1997). What’s wrong - and what’s right - with rubrics. Educational Leadership, 55(2), 72. https://eric.ed.gov/?id=EJ552014

Rafilson, F. (1990). The case for validity generalization. Practical Assessment, Research, and Evaluation, 2(13), 1-3. https://doi.org/10.7275/b51w-5m96

Reddy, Y. M., & Andrade, H. (2010). A review of rubric use in higher education. Assessment & Evaluation in Higher Education, 35(4), 435-448. https://doi.org/10.1080/02602930902862859

Reynolds-Keefer, L. (2010). Rubric-referenced assessment in teacher preparation: An opportunity to learn by using. Practical Assessment, Research, and Evaluation, 15(8): 1-9. https://doi.org/10.7275/psk5-mf68

Rezaei, A. R., & Lovorn, M. (2010). Reliability and validity of rubrics for assessment through writing. Assessing Writing, 15(1), 18-39. https://doi.org/10.1016/j.asw.2010.01.003

Sadler, D. R. (1989). Formative assessment and the design of instructional systems. Instructional Science, 18(2), 119-144. https://doi.org/10.1007/BF00117714

Sadler, D. R. (2009). Indeterminacy in the use of preset criteria for assessment and grading. Assessment & Evaluation in Higher Education, 34(2), 159-179. https://doi.org/10.1080/02602930801956059

Sadler, D. R. (2014). The futility of attempting to codify academic achievement standards. Higher Education, 67(3), 273-288. https://doi.org/10.1007/s10734-013-9649-1

Sambell, K., McDowell, L., & Brown, S. (1997). But is it fair?: An exploratory study of student perceptions of the consequential validity of assessment. Studies in Educational Evaluation, 23(4), 349-371. https://doi.org/https://doi.org/10.1016/S0191-491X(97)86215-3

Scott, S., Webber, C. F., Lupart, J. L., Aitken, N., & Scott, D. E. (2014). Fair and equitable assessment practices for all students. Assessment in Education: Principles, Policy & Practice, 21(1), 52-70. https://doi.org/10.1080/0969594X.2013.776943

Simon, M., & Forgette-Giroux, R. (2001). A rubric for scoring postsecondary academic skills. Practical Assessment, Research, and Evaluation, 7(18), 1-4. https://doi.org/10.7275/bh4d-me80

Stellmack, M. A., Konheim-Kalkstein, Y. L., Manor, J. E., Massey, A. R., and Schmitz, J. A. P. (2009). An assessment of reliability and validity of a rubric for grading APA-style introductions. Teaching of Psychology, 36(2), 102-107. https://doi.org/10.1080/00986280902739776

Stemler, S. E. (2004). A comparison of consensus, consistency, and measurement approaches to estimating interrater reliability. Practical Assessment, Research, and Evaluation, 9(4), 1-11. https://doi.org/10.7275/96jp-xz07

Sundeen, T. H. (2014). Instructional rubrics: Effects of presentation options on writing quality. Assessing Writing, 21, 74-88. https://doi.org/https://doi.org/10.1016/j.asw.2014.03.003

Tierney, R., & Simon, M. (2004). What's still wrong with rubrics: Focusing on the consistency of performance criteria across scale levels. Practical Assessment, Research, and Evaluation, 9(2), 1-7. https://doi.org/10.7275/jtvt-wg68

Van Helvoort, A. A. J. (2012). How adult students in information studies use a scoring rubric for the development of their information literacy skills. The Journal of Academic Librarianship, 38(3), 165-171. https://doi.org/https://doi.org/10.1016/j.acalib.2012.03.016

Weems, G. H., & Onwuegbuzie, A. J. (2001). The impact of midpoint responses and reverse coding on survey data. Measurement and Evaluation in Counseling and Development, 34(3), 166-176. https://doi.org/10.1080/07481756.2002.12069033

Williams, L., & Rink, J. E. (2003). Teacher competency using observational scoring rubrics. Journal of Teaching in Physical Education, 22: 552-572. http://journals.humankinetics.com/jtpe-contents

Wilson, M. 2007. Why I won't be using rubrics to respond to students' writing. The English Journal, 96(4), 62-66. https://doi.org/10.2307/30047167

Zhang, B., Xiao, Y., & Luo, J. (2015). Rater reliability and score discrepancy under holistic and analytic scoring of second language writing. Language Testing in Asia, 5(1), 5. https://doi.org/10.1186/s40468-015-0014-4

Downloads

Published

2025-02-27

How to Cite

Ling, J. H. (2025). A Review of Rubrics in Education: Potential and Challenges. Indonesian Journal of Innovative Teaching and Learning, 2(1), 1–14. https://doi.org/10.64420/ijitl.v2i1.197