Atnaujinkite slapukų nuostatas

Validity in Educational and Psychological Assessment [Kietas viršelis]

  • Formatas: Hardback, 280 pages, aukštis x plotis: 242x170 mm, weight: 620 g
  • Išleidimo metai: 22-Apr-2014
  • Leidėjas: Sage Publications Ltd
  • ISBN-10: 1446253228
  • ISBN-13: 9781446253229
  • Formatas: Hardback, 280 pages, aukštis x plotis: 242x170 mm, weight: 620 g
  • Išleidimo metai: 22-Apr-2014
  • Leidėjas: Sage Publications Ltd
  • ISBN-10: 1446253228
  • ISBN-13: 9781446253229
Lecturers, request your electronic inspection copy to review it for your course. 

Validity is the hallmark of quality for educational and psychological measurement. But what does quality mean in this context? And to what, exactly, does the concept of validity apply? These apparently innocuous questions parachute the unwary inquirer into a minefield of tricky ideas. This book guides you through this minefield, investigating how the concept of validity has evolved from the nineteenth century to the present day.

Communicating complicated concepts straightforwardly, the authors answer questions like:

  • What does 'validity' mean?
  • What does it mean to 'validate'?
  • How many different kinds of validity are there?
  • When does validation begin and end?
  • Is reliability a part of validity, or distinct from it?

This book will be of interest to anyone with a professional or academic interest in evaluating the quality of educational or psychological assessments, measurements and diagnoses.



Offering an in-depth study of validity, which forms an important part of education assessment, this book addresses complex concepts in a straightforward manner, while guiding the reader through the evolution of validity from the 19th century to the present day.

Recenzijos

Countless books have addressed validity, but this is the first volume to provide a comprehensive treatment of the evolution of validity theory in the last century and a framework for evaluating educational and psychological testing in the 21st century. -- Wayne J. Camara In this groundbreaking book, Newton and Shaw show how the "consensus" view of validitythat validity is not a property of tests, but of inferences made on the basis of the evidence they elicitwas at best rather shallow. Notably, the "consensus" view leaves unresolved important tensions between those who see validation as a never-ending process (e.g., Cronbach, Messick) and those who understand the needs of those who produce assessments to be able to say that they have undertaken necessary due diligence to attest to the quality of assessments (e.g., Ebel, Kane). As well as providing an excellent, scholarly review of the history of the idea of validity, Newton and Shaw show how a modified version of Messicks facet model of validity can produce a rigorously grounded, and yet practical, approach to assuring the quality of educational and psychological assessments. Every serious scholar of assessment should read this book. -- Dylan Wiliam At last - a scholarly and well-argued book that means validity, the key concept in any assessment, is no longer an essentially American debate. Its scope, logic and clarity will quickly make it the standard international text. -- Gordon Stobart With concerns over the validity of educational and psychological tests and assessments growing, this book by Paul Newton and Stuart Shaw is a welcome, and much needed, contribution to the measurement literature.   Anyone concerned with the technical quality and value propositions surrounding large-scale testing and assessment ought to read this book.  Newton and Shaw take the reader through the history of "validity" as researchers across time and across disciplines use the term.  In doing so, they advance the debate by offering a unique and novel framework for assessing issues of test validity.  Newton and Shaw leave us with some new conceptstheoretical plausibility and practical viabilityfor judging the evidence in support of a tests validity.   This book presents todays and tomorrows perspectives on validity as the core idea in educational and psychological assessment.  It is a must read for those working in the broad field of psychometrics. -- Howard T. Everson [ This book] was enlightening in terms of helping me reflect on a range of issues around testing and validity. It provides a good way of thinking about these issues in a way that takes into account a range of views from different stakeholders about why a particular testing policy exists. The framework enables different agendas to be made explicit and tensions between different purposes of a particular policy laid bare. It provides a way of critically evaluating policy and making decisions about implementing testing and assessment policies. I wish I had read this book before writing a sub chapter called Why Assess? - covering the same ground on issues of dyslexia assessment would have been easier using Newton and Shaws revised framework...I would recommend the book to experienced practitioners who want to reconsider the validity of assessments that they undertake. It is perhaps useful for trainers as a book that enables trainee psychologists to consider the social context in which psychometric measures are used and to consider validity in terms which take them beyond the technical aspects of testing and assessment. It is also a useful book for policy makers to consider, especially around the social issues related to assessment and testing policy implementation.  -- Dr Garry Squires, programme director for the Doctorate in Educational Psychology, University of Manchester An ideal text for post-graduate courses on educational or psychological assessment, and selections would also make strong contributions to reading lists for under-graduate courses. There is plenty in it for more seasoned scholars of assessment theory as well ... because the historical survey and analysis of the literature is so insightful -- Michael Johnston In an accessible, statistics-free treatment, Newton and Shaw first distinguish assessment validity from related concepts (such as research validity), before tracing its evolution through the primarily US-led assessment research tradition over the past century. As doctoral students we hope to tap into issues at the cutting edge of the field, and Newton and Shaws book enables this, providing the background and theory needed to contribute to contemporary assessment research. -- Ricky Jeffrey

About the authors xi
Acknowledgements xiii
List of tables and figures xv
Foreword xvii
Preface xix
1 Validity and validation 1(26)
What do we mean by validity?
2(5)
Validity across disciplines
2(1)
Validity for research
3(2)
Validity for measurement: attributes and decisions
5(2)
Kinds of validity for measurement
7(3)
Conventions used in the book
10(4)
Educational and psychological measurement
10(1)
Attributes or constructs?
11(1)
Particular kinds of attribute
12(1)
Reliability and validity
13(1)
An outline of the history of validity
14(13)
The genesis of validity (mid-1800s-1951)
16(3)
A gestational period (mid-1800s-1920)
17(1)
A period of crystallization (1921-1951)
18(1)
The fragmentation of validity (1952-1974)
19(2)
The (re)unification of validity (1974-1999)
21(1)
The deconstruction of validity (2000-2012)
22(2)
Twenty-first-century evaluation
24(3)
2 The genesis of validity: mid-1800s-1951 27(36)
A gestational period (pre-1921)
27(3)
A period of crystallization (post-1921)
30(8)
Existing accounts of the early years
32(4)
Explaining the caricature
36(2)
From quality of measurement to degree of correlation
36(1)
From degree of correlation to coefficient of validity
37(1)
Validating tests for different purposes
38(16)
Validity and school achievement tests
39(7)
Sampling theory versus test construction practice
41(1)
Measurement versus evaluation
42(1)
Sampling theory versus test construction theory
43(2)
The ongoing tension between empirical and logical
45(1)
Validity and general intelligence tests
46(3)
A logical approach to test development
46(1)
An empirical approach to validation
47(2)
Beyond correlation with criterion measures
49(1)
Validity and special aptitude tests
49(3)
An empirical approach to validation
50(1)
An empirical approach to test development
50(1)
The criterion problem
51(1)
Validity and personality tests
52(2)
Paul Meehl
53(1)
Lee Cronbach
53(1)
Alternative explanations
54(1)
Validity and validation by the end of the 1940s
54(9)
Divergent views
55(2)
A more interesting and nuanced story
57(6)
3 The fragmentation of validity: 1952-1974 63(36)
Professional standards: edition 1
64(7)
Different approaches to validation
65(4)
Content validity
66(1)
Predictive validity
66(1)
Concurrent validity
67(1)
Construct validity
68(1)
Different kinds of validity
69(2)
The invention of construct validity
71(3)
Professional standards: editions 2 and 3
74(3)
Operationism
77(3)
The entrenchment of fragmented thinking
80(8)
Seeds of discontent
84(4)
Cronbach on validation
88(3)
Validity and validation by the mid-1970s
91(8)
4 The (re)unification of validity: 1975-1999 99(36)
The Messick years: triumph and tribulation
100(16)
The Messick years prefigured
100(2)
The new science of validity
102(14)
Deeply entrenched fragmentation
102(1)
Messick's mission
103(1)
Why 'content validity' is insufficient
103(3)
Why 'criterion validity' is insufficient
106(2)
The general relevance of nomological networks
108(1)
The ethical imperative underpinning construct validity
109(1)
All validity is construct validity
110(4)
The new practice of validation
114(2)
The progressive matrix
116(13)
The logic of the matrix
117(1)
The illogic of the matrix
118(3)
The confusing role of social consequences
121(15)
Consequences and the scope of validity theory
122(1)
Earlier years
122(1)
Transitional years
123(2)
Later years
125(2)
The transition
127(2)
Professional standards: editions 4 and 5
129(2)
Validity and validation by the end of the 1990s
131(4)
5 The deconstruction of validity: 2000-2012 135(48)
The desire to simplify validation practice
136(10)
The methodology in principle
137(4)
The methodology in practice
141(5)
The desire to simplify validity theory
146(20)
The nature and significance of construct validity
146(20)
The construction of construct validity theory
146(2)
Deconstruction and reconstructions of construct validity
148(19)
Borsboom (part 1): tests are valid, not interpretations
148(1)
Lissitz and Samuelsen: no need for nomological networks
149(1)
Embretson, Pellegrino and Gorin: the cognitive approach
150(3)
Kane: not all attributes are hypothetical
153(2)
Borsboom (part 2): validity is ontological, not epistemological
155(10)
Michell and Maraun: doubts about measurement
165(1)
Moss: situated validation
165(1)
Construct validity in the balance
166(6)
The nature of validation research
167(2)
The nature of constructs/attributes
168(1)
The nomological network
169(1)
The label 'construct validity'
170(1)
In the balance
171(1)
The scope of validity
172(7)
Emergence of the debate
172(2)
The arguments for
174(1)
The arguments against
175(1)
The various camps
176(2)
The ongoing debate
178(1)
Validity and validation into the 21st century
179(4)
6 Twenty-first-century evaluation 183(44)
A framework for the evaluation of testing policy
184(5)
Defending the new matrix
189(3)
Evaluation of technical quality: cells 1 to 3
192(2)
Evaluation of social value: cells 4a to 4c and the Overall Judgement
194(4)
Illustrating the framework
198(26)
Cell 1
198(4)
Specifying the attribute
198(1)
Theoretical plausibility
199(1)
Policy owner values and purposes
200(1)
Reconciling values and the 'real world'
201(1)
Cell 2
202(8)
Decision-making versus intervention
203(1)
Specifying the outcome
204(1)
Theoretical plausibility
205(2)
Different ways of constructing a theory of the decision
207(3)
Cell 3
210(2)
Specifying the impact
210(1)
Theoretical plausibility
211(1)
Cells 4a to 4c
212(6)
Cell 4a expenses
213(1)
Cell 4b primary expenses, pay-offs, impacts and side-effects
213(3)
Cell 4c secondary expenses, pay-offs, impacts and side-effects
216(2)
The Overall Judgement
218(6)
Acceptable to the policy owner
221(1)
Acceptable to the academy
221(1)
Acceptable to the public
222(1)
Acceptable under the law
223(1)
The context of the Overall Judgement
224(1)
Conclusion
224(3)
References 227(20)
Index 247
Paul E. Newton is Professor of Educational Assessment at the Institute of Education, University of London. His research focuses primarily upon issues related to the evaluation of large-scale educational assessment systems, and he is particularly interested in theories of validity for educational and psychological measurement, past and present. He has published on a range of assessment topics, including validity, comparability, assessment purposes, national curriculum test reliability, and the public understanding of measurement inaccuracy.

Having obtained a PhD in developmental psychology, Paul moved into educational assessment and has spent most of his career as a researcher within a range of assessment agencies including the Associated Examining Board, the National Foundation for Educational Research, the Qualifications and Curriculum Authority, Ofqual and Cambridge Assessment.

Paul is a member of the Editorial Board of the international journal Assessment in Education: Policy, Principles and Practice and has served on a variety of national and international committees, including the Executive Committee of the International Association for Educational Assessment. He was a member of the Assessment Reform Group, until its retirement in 2010, and is a Fellow of the Association for Educational Assessment Europe.

Stuart D. Shaw is particularly interested in demonstrating how educational, vocational and second language English assessments meet the demands of validity, reliability and fairness. Stuart has a wide range of pub­lications in English second-language assessment and educational research journals. His assessment books include: Examining Writing: Research and practice in assessing second language writing (Shaw and Weir, 2007); The IELTS Writing Assessment Revision Project: Towards a revised rating scale (Shaw and Falvey, 2008); Validity in Educational and Psychological Assessment (Newton and Shaw, 2014); and Language Rich: Insights from multilingual schools (Shaw, Imam and Hughes, 2015). Stuart is a Fellow of the Association for Educational Assessment in Europe (AEA-E) and a Fellow of the Chartered Institute of Educational Assessors (CIEA).