Assessing mathematical problem solving using comparative judgement

Article English OPEN
Jones, Ian ; Swan, Malcolm ; Pollitt, Alistair (2014)

There is an increasing demand from employers and universities for school leavers to be able to apply their mathematical knowledge to problem solving in varied and unfamiliar contexts. These aspects are however neglected in most examinations of mathematics and, consequentially, in classroom teaching. One barrier to the inclusion of mathematical problem solving in assessment is that the skills involved are difficult to define and assess objectively. We present two studies that test a method called comparative judgement (CJ) that might be well suited to assessing mathematical problem solving. CJ is an alternative to traditional scoring that is based on collective expert judgements of students’ work rather than item-by-item scoring schemes. In Study 1 we used CJ to assess traditional mathematics tests and found it performed validly and reliably. In Study 2 we used CJ to assess mathematical problem-solving tasks and again found it performed validly and reliably. We discuss the implications of the results for further research and the implications of CJ for the design of mathematical problem-solving tasks.
  • References (44)
    44 references, page 1 of 5

    ACT. (2006). Ready for College and Ready for Work: Same or Different? Iowa: American College Tests, INC.

    AQA (2010). GCSE Higher Tier Mathematics Paper 1 (Specification A). Monday 7 June 2010. Manchester: Assessment and Qualifications Alliance.

    Black, P. (2008). Strategic decisions: Ambitions, feasibility and context. Educational Designer, 1(1).

    Black, P., & Wiliam, D. (2007). Large-scale assessment systems: Design principles drawn from international comparisons. Measurement: Interdisciplinary Research and Perspectives, 5, 1-53.

    Black, P., Burkhardt, H., Daro, P., Jones, I., Lappan, G., Pead, D., & Stephens, M. (2012). High-stakes examinations to support policy. Educational Designer, 2(5).

    Bond, T. G., & Fox, C. M. (2007). Applying the Rasch Model: Fundamental Measurement in the Human Sciences. Abingdon: Routledge.

    Bramley, T., Bell, J., & Pollitt, A. (1998). Assessing changes in standards over time using Thurstone paired comparisons. Education Research and Perspectives, 25, 1-24.

    CBI. (2006). Working with the Three Rs: Employers' Priorities for Functional Skills in Mathematics and English. London: DfES.

    Davies, D., Collier, C., & Howe, A. (2012). Assessing scientific and technological enquiry skills at age 11 using the e-scape system. International Journal of Technology and Design Education, 22, 247-263.

    Derrick, K. (2012). Developing the e-scape software system. International Journal of Technology and Design Education, 22, 171-185.

  • Metrics
    No metrics available
Share - Bookmark