Formative Use of Assessment to Foster Self‑Regulated Learning: the Alignment of Teachers’ Conceptions and Classroom Assessment Practices Journal of Formative Design in Learning (AECT)
Assessment in higher education (HE) is often focused on concluding modules with one or more tests that students need to pass. As a result, both students and teachers are primarily concerned with the summative function of assessment: information from tests is used to make pass/fail decisions about students. In recent years, increasing attention has been paid to the formative function of assessment and focus has shifted towards how assessment can stimulate learning. However, this also leads to a search for balance between both functions of assessment. Programmatic assessment (PA) is an assessment concept in which their intertwining is embraced to strike a new balance. A growing number of higher education programmes has implemented PA. Although there is consensus about the theoretical principles that form the basis for the design of PA, programmes make various specific design choices based on these principles, fitting with their own context. This paper provides insight into the design choices that programmes make when implementing PA and into the considerations that play a role in making these design choices. Such an overview is important for research purposes because it creates a framework for investigating the effects of different design choices within PA.
In programmatic assessment (PA), an arrangement of different assessment methods is deliberately designed across the entire curriculum, combined and planned to support both robust decision-making and student learning. In health sciences education, evidence about the merits and pitfalls of PA is emerging. Although there is consensus about the theoretical principles of PA, programs make diverse design choices based on these principles to implement PA in practice, fitting their own contexts. We therefore need a better understanding of how the PA principles are implemented across contexts—within and beyond health sciences education. In this study, interviews were conducted with teachers/curriculum designers representing nine different programs in diverse professional domains. Research questions focused on: (1) design choices made, (2) whether these design choices adhere to PA principles, (3) student and teacher experiences in practice, and (4) context-specific differences between the programs. A wide range of design choices were reported, largely adhering to PA principles but differing across cases due to contextual alignment. Design choices reported by almost all programs include a backbone of learning outcomes, data-points connected to this backbone in a longitudinal design allowing uptake of feedback, intermediate reflective meetings, and decision-making based on a multitude of data-points made by a committee and involving multi-stage procedures. Contextual design choices were made aligning the design to the professional domain and practical feasibility. Further research is needed in particular with regard to intermediate-stakes decisions.
LINK
Wil je als docententeam meer zicht krijgen op het gehele toetsprogramma? Wil je kritisch kijken naar mogelijke verbeterpunten? Of ben je bezig met herontwerp? Met KIT2.0 kijk je als opleidingsteam vanuit de principes van programmatisch toetsen naar de inrichting van de opleiding.Doel Met KIT2.0 willen we opleidingsteams helpen om kritisch naar het curriculum en het toetsprogramma te kijken. Dit doen we aan de hand van vijf kwaliteitscriteria: fitness for purpose, validiteit, leerfunctie, beslisfunctie en condities. Resultaten Op de website van KIT2.0 vind je informatie en filmpjes met verdere uitleg. Via de website kun je ook (gratis) inloggen en zelf aan slag met KIT2.0. Op de website www.husite.nl/toetsing vind je informatie en praktijkvoorbeelden over programmatisch toetsen Blog over interview met Liesbeth Baartman over KIT2.0. Korte uitleg van dr. Liesbeth Baartman (2017) programmatisch toetsen. Toetsbijeenkomst Hogeschool van Rotterdam. Keynote van dr. Liesbeth Baartman (2017) met een inleiding over toetsprogramma’s. Fontys Toetscongres. Baartman, L.K.J., Kloppenburg, R., & Prins, F.J. (2017). Kwaliteit van toetsprogramma’s. In H. van Berkel, A. Bax, & D. Joosten-ten-Brinke (Red.). Toetsen in het Hoger Onderwijs, pp.38-49. Bohn Stafleu van Loghum Van der Vleuten, C.P.M., Schuwirth, L.T.W., Driessen, E., Dijkstra, J., Tigelaar, D., Baartman, L.K.J., & Van Tartwijk, J. (2012). A model for programmatic assessment fit for purposes. Medical Teacher, 34, 205-214. Dronkers, J., de Kwant, E., Kruitwagen, C., & Baartman, L. (2017). Kwantitatieve analyse van een toetsprogramma. Examens, 3, augustus. Looptijd 01 september 2018 - 01 september 2020 Aanpak KIT2.0 is gebaseerd op wetenschappelijk onderzoek naar programmatisch toetsen. De oorsprong van KIT2.0 ligt in de promotieonderzoeken van dr. Liesbeth Baartman en dr. Raymond Kloppenburg (waaruit KIT1.0 voortkwam). KIT2.0 is ontwikkeld op basis van nieuwste inzichten in de wetenschappelijk literatuur over programmatisch toetsen én 10 jaar praktijkervaringen. KIT2.0 is ontwikkeld in valideringsrondes met opleidingen en wetenschappers. Meedoen? Wil je als opleiding meedoen aan het onderzoek naar KIT2.0? Neem dan contact op met Liesbeth Baartman. We werken aan de evaluatie en verbetering van KIT2.0 op basis van praktijkervaringen.