covid
Buscar en
Psicología Educativa - Educational Psychology
Toda la web
Inicio Psicología Educativa - Educational Psychology Cognitively Diagnostic Assessments and the Cognitive Diagnosis Model Framework
Información de la revista
Vol. 20. Núm. 2.
Páginas 89-97 (diciembre 2014)
Compartir
Compartir
Descargar PDF
Más opciones de artículo
Vol. 20. Núm. 2.
Páginas 89-97 (diciembre 2014)
Open Access
Cognitively Diagnostic Assessments and the Cognitive Diagnosis Model Framework
El marco de la evaluación y los modelos de diagnostico cognitivo
Visitas
6919
Jimmy de la Torre1,
Autor para correspondencia
j.delatorre@rutgers.edu

Corresponding author.
, Nathan Minchen
Rutgers, The State University of New Jersey, U.S.A
Este artículo ha recibido

Under a Creative Commons license
Información del artículo
Abstract

This paper aims to identify the utility of and the need for cognitively diagnostic assessments (CDAs) in conjunction with cognitive diagnosis models (CDMs), and to outline various considerations involved in their development and use. We begin by contrasting the CDA/CDM framework against existing assessment frameworks, which are typically based on item response theory or classical test theory, and show that CDAs used in the CDM context can provide valuable diagnostic information that could enhance classroom instruction and learning. We then detail how the components of a CDA fit into the assessment triangle framework, as well as the evidence-centered design framework. Attribute identification and item development in the context of CDA are discussed, and examples from relevant research are provided. Details of CDMs, which are the statistical models that underpin the practical implementations of CDAs, are also discussed.

Keywords:
Cognitively diagnostic assessments
Cognitive diagnosis model
Evidence-centered design
Assessment triangle
Resumen

El presente artículo trata de identificar la utilidad y la necesidad de las Evaluaciones para el Diagnóstico Cognitivo (EDC) junto a los Modelos de Diagnóstico Cognitivo (MDC) y plantea algunas de las consideraciones implicadas en su desarrollo y uso. Se comienza comparando el marco EDC/MDC con otros marcos existentes basados típicamente en la teoría de respuesta al ítem o en la teoría clásica de los tests, mostrando que las EDC utilizadas en el contexto MDC pueden proporcionar información diagnóstica muy valiosa para mejorar el proceso de enseñanza-aprendizaje en el aula. Seguidamente se utiliza el marco del triángulo de la evaluación y del diseño centrado en la evidencia para presentar los componentes de una evaluación de este tipo. Se analiza la identificación de atributos y el desarrollo de ítems en el contexto de una EDC y se proporcionan ejemplos tomados de una investigación relevante. También se analizan cuestiones relativas a los MDC, que son los modelos estadísticos que vertebran la implementación práctica de las EDC.

Palabras clave:
Evaluación para el diagnóstico cognitivo
Modelo de diagnóstico cognitivo
Diseño centrado en la evidencia
Triángulo de evaluación
El Texto completo está disponible en PDF
References
[Bachman and Palmer, 2010]
L. Bachman, A. Palmer.
Language assessment in practice: Developing language assessments and justifying their use in the real world.
Oxford University Press, (2010),
[Chudowsky and Pellegrino, 2003]
N. Chudowsky, J.W. Pellegrino.
Large-scale assessments that support learning: What will it take?.
Theory into Practice, 42 (2003), pp. 75-83
[de la Torre, 2009]
J. de la Torre.
A cognitive diagnosis model for cognitively based multiple-choice options.
Applied Psychological Measurement, 33 (2009), pp. 163-183
[de la Torre, 2012]
J. de la Torre.
Application of the DINA Model Framework to Enhance Assessment and Learning.
Self-directed learning oriented assessments in the Asia-Pacific, pp. 92-110
[de La Torre, 2011]
J. de La Torre.
The generalized DINA model framework.
Psychometrika, 76 (2011), pp. 179-199
[de la Torre and Karelitz, 2009]
J. de la Torre, T.M. Karelitz.
Impact of diagnosticity on the adequacy of models for cognitive diagnosis under a linear attribute structure: A simulation study.
Journal of Educational Measurement, 46 (2009), pp. 450-469
[de la Torre and Lee, 2010]
de la Torre, J., & Lee, Y.S. (2010, April). Item-level comparison of saturated and reduced cognitive diagnosis models. Paper presented at the Annual Meeting of the National Council on Measurement in Education, Denver, CO.
[DiBello et al., 2007]
L.V. DiBello, L.A. Roussos, W. Stout.
Review of cognitively diagnostic assessment and a summary of psychometric models.
Handbook of statistics, 26 (2007), pp. 979-1030
[Haertel, 1989]
E.H. Haertel.
Using restricted latent class models to map the skill structure of achievement items.
Journal of Educational Measurement, 26 (1989), pp. 301-321
[Hartz, 2002]
Hartz, S.M. (2002). A Bayesian framework for the Unified Model for assessing cognitive abilities: blending theory with practicality.(Unpublished doctoral dissertation).
[Henson and Douglas, 2005]
R. Henson, J. Douglas.
Test construction for cognitive diagnosis.
Applied Psychological Measurement, 29 (2005), pp. 262-277
[Junker and Sijtsma, 2001]
B.W. Junker, K. Sijtsma.
Cognitive assessment models with few assumptions, and connections with nonparametric item response theory.
Applied Psychological Measurement, 25 (2001), pp. 258-272
[Kane, 2013]
M. Kane.
Validating the interpretations and uses of test scores.
Journal of Educational Measurement, 50 (2013), pp. 1-73
[Leighton et al., 2004]
J.P. Leighton, M.J. Gierl, S. Hunka.
The attribute hierarchy method for cognitive assessment: A variation on Tatsuoka's rule-space approach.
Journal of Educational Measurement, 41 (2004), pp. 205-237
[Maris, 1999]
E. Maris.
Estimating multiple classification latent class models.
Psychometrika, 64 (1999), pp. 187-212
[Mislevy, 2011]
Mislevy, R.J. (2011). Evidence-centered design for simulation-based assessment (CRESST Report 800). Los Angeles, CA: The National Center for Research on Evaluation, Standards, and Student Testing, UCLA.
[Mislevy and Haertel, 2006]
R.J. Mislevy, G.D. Haertel.
Implications of evidence-centered design for educational testing.
Educational Measurement: Issues and Practice, 25 (2006), pp. 6-20
[Mislevy et al., 2003]
R.J. Mislevy, L.S. Steinberg, R.G. Almond.
On the structure of educational assessments.
Measurement: Interdisciplinary research and perspectives, 1 (2003), pp. 3-62
[Pellegrino, 2014]
J.W. Pellegrino.
Assessment as a positive influence on 21st century teaching and learning: A systems approach to progress.
Psicología Educativa, 20 (2014), pp. 65-77
[Pellegrino et al., 2001]
Knowing what students know: The science and design of educational assessment (National Research Council's Committee on the Foundations of Assessment),
[Shepard, 2000]
L.A. Shepard.
The role of assessment in a learning culture.
Educational Researcher, 29 (2000), pp. 4-14
[Tatsuoka, 1995]
K.K. Tatsuoka.
Architecture of knowledge structures and cognitive diagnosis: A statistical pattern recognition and classification approach.
Cognitively diagnostic assessment, pp. 327-359
[Tatsuoka, 1983]
K.K. Tatsuoka.
Rule-space: An approach for dealing with misconceptions based on item response theory.
Journal of Educational Measurement, 20 (1983), pp. 345-354
[Tatsuoka et al., 2014]
Tatsuoka et al., Tatsuoka, C., Clements, D.H., Sarama, J., Izsak, A., Orril, C.H., de la Torre, J., & Khasanova, E. (in press). Developing workable attributes for psychometric models based on the Q-matrix. Journal of Research on Mathematics Education.
[Templin and Henson, 2006]
J.L. Templin, R.A. Henson.
Measurement of psychological disorders using cognitive diagnosis models.
Psychological methods, 11 (2006), pp. 287-305
[Tjoe and de la Torre, 2012]
H. Tjoe, J. de la Torre.
Proportional reasoning problems: Current state and a possible future direction.
Proceedings of the 12th International Congress on Mathematics Education, pp. 6741-6750
[Tjoe and de la Torre, 2013a]
H. Tjoe, J. de la Torre.
Designing cognitively-based proportional reasoning problems as an application of modern psychological measurement models.
Journal of Mathematics Education, 6 (2013), pp. 17-22
[Tjoe and de la Torre, 2013b]
H. Tjoe, J. de la Torre.
The identification and validation process of proportional reasoning attributes: An application of a cognitive diagnosis modeling framework.
Mathematics Education Research Journal, 26 (2013), pp. 237-255
[Tjoe and de la Torre, 2014]
H. Tjoe, J. de la Torre.
On recognizing proportionality: Does the ability to solve missing value proportional problems presuppose the conception of proportional reasoning?.
The Journal of Mathematical Behavior, 33 (2014), pp. 1-7
[Department of Education, 2013]
U.S. Department of Education, Institute of Education Sciences, National Center for Education Statistics (2013). National Assessment of Educational Progress (NAEP), 2013 Mathematics Assessment. Retrieved from http://nces.ed.gov/nationsreportcard/itemmaps/index.asp.
[Zieky, 2014]
M.J. Zieky.
An introduction to the use of evidence-centered design in test development.
Psicología Educativa, 20 (2014), pp. 79-87

Correspondence concerning this article should be addressed to Jimmy de la Torre. Department of Educational Psychology Rutgers. The State University of New Jersey. 10 Seminary Place. New Brunswick, NJ 08901 USA.

Copyright © 2014. Colegio Oficial de Psicologos de Madrid
Descargar PDF
Opciones de artículo