The EFPA Test-Review Model: When Good Intentions Meet a Methodological Thought Disorder
Abstract
:1. Introduction
“The main goal of the EFPA Test Review Model is to provide a description and a detailed and rigorous assessment of the psychological assessment tests, scales and questionnaires used in the fields of Work, Education, Health and other contexts. This information will be made available to test users and professionals in order to improve tests and testing and help them to make the right assessment decisions.”
“By methodological thought disorder, I do not mean simply ignorance or error, for there is nothing intrinsically pathological about either of those states …Hence, the thinking of one who falsely believes he is Napoleon is adjudged pathological because the delusion was formed and persists in the face of objectively overwhelming contrary evidence. I take thought disorder to be the sustained failure to see things as they are under conditions where the relevant facts are evident. Hence, methodological thought disorder is the sustained failure to cognize relatively obvious methodological facts …I am interested, however, not so much in methodological ignorance and error amongst psychologists per se, as in the fact of systemic support for these states in circumstances where the facts are easily accessible. Behind psychological research exists an ideological support structure. By this I mean a discipline-wide, shared system of beliefs which, while it may not be universal, maintains both the dominant methodological practices and the content of the dominant methodological educational programmes. This ideological support structure is manifest in three ways: in the contents of textbooks; in the contents of methodology courses; and in the research programmes of psychologists. In the case of measurement in psychology this ideological support structure works to prevent psychologists from recognizing otherwise accessible methodological facts relevant to their research. This is not then a psychopathology of any individual psychologist. The pathology is in the social movement itself, i.e., within modern psychology.”[2] (p. 374, Section 5.1)
2. A Critical Evaluation of the EFPA Test Review Guidelines
2.1. The EFPA Guidelines Which Are Sensible and of Pragmatic Utility
- General description
- Classification
- Measurement and scoring
- Computer-generated reports
- Supply conditions and costs
- Quality of computer-generated reports
- Final evaluation
2.2. Those Guidelines Which Represent a Target for Legal Challenge
“It follows that for attributes investigated in science, there are three structural possibilities: (1) classificatory attributes (with heterogeneous differences between categories); (2) heterogeneous orders (with heterogeneous differences between degrees); and (3) quantitative attributes (with thoroughly homogeneous differences between magnitudes). Measurement is possible only with attributes of kind (3) and, as far as we know, psychological attributes are exclusively of kinds (1) or (2). However, contrary to the known facts, psychometricians, for their own special reasons insist that test scores provide measurements.”
“We need more, not less, by way of modern causal mechanisms”. Pennington (2014) asked and answered the question of what is required to provide a scientific explanation in the following way: “What does it mean to explain something? Basically, it means that we identify the cause of that thing in terms of relevant mechanisms” [15] (p. 3, emphasis added). Psychologists claim to have mechanism information, but as Tryon [16] and Tryon, Hoffman, and McKay [17] explain, these claims are mainly false and illusory. For example, the very popular biopsychosocial model is just a list of relevant factors; it explains nothing more about psychology and behavior than a glass–metal–petroleum model would explain about how automobiles work. Listing variables, or ingredients, does not constitute explanation. We place variable names in boxes and draw arrows among the boxes, thereby imputing causality that is never explained. Drawing arrows does not constitute explanation. We say that squared correlations explain variance when they only account for variance. Accounting is not explaining. We use brain scans to identify brain lobes that are associated with psychological functions, but we cannot explain how those brain lobes do anything psychological any better than phrenologists could. Associations are not explanations. We identify mediators by correlational methods and discuss them as causal mechanisms. Correlation cannot establish causation. These “explanations” are “illusions of understanding”.[13] (p. 505)
2.3. How Have We Reached This Sub-Optimal State of Affairs
2.3.1. The Aspiration to Emulate Measurement within the Physical Sciences
2.3.2. The Commercial Imperative
2.3.3. The ”Nelson” Syndrome
“The phrase to turn a blind eye is attributed to an incident in the life of Admiral Horatio Nelson. Nelson was blinded in one eye early in his Royal Navy career. During the Battle of Copenhagen in 1801 the cautious Admiral Sir Hyde Parker, in overall command of the British forces, sent a signal to Nelson’s forces ordering them to discontinue the action. Naval orders were transmitted via a system of signal flags at that time. When this order was brought to the more aggressive Nelson’s attention, he lifted his telescope up to his blind eye, saying, “I really do not see the signal,” and most of his forces continued to press home the attack.”
2.4. Does Any of This Really Matter?
“On the other hand, in much, perhaps even most, research in psychology, there are no practically relevant units anyway. For example, in social psychology, what is an attitude unit, a prejudice unit, or a self-affirmation unit? In clinical psychology, what is a depression, anxiety, or psychopathy unit? In education, what is a knowledge unit? Similar questions can be asked with respect to measurement in many areas.”
“Now, if a person’s correct response to an item depended solely on ability, with no random ‘error’ component involved, one would only learn the ordinal fact that that person’s ability at least matches the difficulty level of the item. Item response modellers derive all quantitative information (as distinct from merely ordinal) from the distributional properties of the random ‘error’ component. If the model is true, the shape of the ‘error’ distribution reflects the quantitative structure of the attribute, but if the attribute is not quantitative, the supposed shape of ‘error’ only projects the image of a fictitious quantitivity. Here, as elsewhere, psychometricians derive what they want most (measures) from what they know least (the shape of ‘error’) by presuming to already know it.”
“Item response modellers derive all quantitative information (as distinct from merely ordinal) from the distributional properties of the random ‘error’ component.”
“The many and diverse choices that made certain companies great were consistent with just three seemingly elementary rules:
Better before cheaper-in other words, compete on differentiators other than price. Revenue before cost-that is, prioritize increasing revenue over reducing costs. There are no other rules-so change anything you must to follow Rules 1 and 2.The rules don’t dictate specific behaviors; nor are they even general strategies. They’re foundational concepts on which companies have built greatness over many years.”
3. The Characteristic Features of the Next Generation of Assessments
“No technology of which we are aware—computers, telecommunications, televisions, and so on—has shown the kind of ideational stagnation that has characterized the testing industry. Why? Because in other industries, those who do not innovate do not survive. In the testing industry, the opposite appears to be the case. Like Rocky I, Rocky II, Rocky III, and so on, the testing industry provides minor cosmetic successive variants of the same product where only the numbers after the names substantially change. These variants survive because psychologists buy the tests and then loyally defend them (see preceding nine commentaries, this issue).The existing tests and use of tests have value, but they are not the best they can be. When a commentator says that it will never be possible to improve much on the current admissions policies of Yale and its direct competitors [47] (p. 572, this issue), that is analogous to what some said about the Model T automobile and the UNIVAC computer. Comments such as this one prove our point [48] better than we ever could.”
- They are designed from the ground-up using evidence-bases drawn from a wide vista of psychology, neuroscience theory, and experiment-evidence.
- There are no self-report/ability-test-type questionnaire items or static text-rich/scenario-type items (such as video or text-based Situational Judgment Tests); just information acquired from dynamic game or task performance, behavior, and/or linguistic analysis of text-based internet activity.
- Assessments sometimes are invariably comprised of thousands of very specific observations, clustered by expert-system rule or empirically into broader categories.
- Scoring is via expert-system rules and theory-relevant algorithms and/or varieties of machine-learning optimized prediction models.
- There are no psychometric “scales” as such, just attribute “type”, ordered-class, or quantitative (e.g., time) magnitude assessment which might be constituted from many diverse but theoretically/experimentally related behavioral information sources.
- Scoring information for any commercial assessment constitutes the Intellectual Property (IP) of an assessment, and is not made public.
- Reliability of any assessment outcome, type, ordered-class, or quantitative magnitude is assessed using appropriate retest methodologies only.
- Structural “validation” psychometrics (factor analysis, SEM modelling etc.) cannot be effectively used due to the mix of assessment-variable properties and expert-system rule-based scoring procedures which produce the final attribute magnitudes, orders, and classes. Basically, concepts from psychometric test theory simply do not apply to the kinds of attribute constituents, classifications, or “score” magnitudes.
4. The Proposed Next-Generation Test Review Frameworks
4.1. The Scientific Framework
4.2. The Pragmatic Framework
“The argument about ordering vs. quantification has been made about all social science research. It does not hold water. There is plenty of evidence that many psychological scales are essentially interval scales or close to it”.
“Baseball and horse racing provide two popular analogues. Baseball batters with the greatest number of home runs, runs batted in (RBIs) and highest batting average in a year are triple crown winners. In U.S. horse racing, the triple crown goes to the 3-year-old thoroughbred that wins the Kentucky Derby, the Preakness Stakes and the Belmont Stakes in the same year.”
“In the end, triple crown companies share the following attributes:
Clarity of vision Disciplined resource allocation Excellence in execution” p. 39
- Non-systematic random error associated with the internal integrity of the test itself.
- Systematic and meaningful attribute variation of the attribute over periods of time within and extending across the retest duration.
- Memory of previous responses artificially causing consistency in 2nd occasion response patterns.
“Validity is not complex, faceted, or dependent on nomological networks and social consequences of testing. It is a very basic concept and was correctly formulated, for instance, by Kelley [64] (p. 14) when he stated that a test is valid if it measures what it purports to measure …A test is valid for measuring an attribute if and only if (a) the attribute exists and (b) variations in the attribute causally produce variations in the outcomes of the measurement procedure.”
“This is clear because validity is a property, whereas validation is an activity. In particular, validation is the kind of activity researchers undertake to find out whether a test has the property of validity. Validity is a concept like truth: It represents an ideal or desirable situation. Validation is more like theory testing: the muddling around in the data to find out which way to go. Validity is about ontology; validation is about epistemology. The two should not be confused.”
4.3. A Matter of Scientific Integrity
“I am interested, however, not so much in methodological ignorance and error amongst psychologists per se, as in the fact of systemic support for these states in circumstances where the facts are easily accessible. Behind psychological research exists an ideological support structure. By this I mean a discipline-wide, shared system of beliefs which, while it may not be universal, maintains both the dominant methodological practices and the content of the dominant methodological educational programmes. This ideological support structure is manifest in three ways: in the contents of textbooks; in the contents of methodology courses; and in the research programmes of psychologists. In the case of measurement in psychology this ideological support structure works to prevent psychologists from recognizing otherwise accessible methodological facts relevant to their research. This is not then a psychopathology of any individual psychologist. The pathology is in the social movement itself, i.e., within modern psychology.”
“Why did you proceed with guidelines post-1997 which included methods of assessment evaluation predicated upon an untenable and untested assumption?”
“Computational psychometrics (CP) is defined as a blend of data-driven computer science methods (machine learning and data mining, in particular), stochastic theory, and theory-driven psychometrics in order to measure latent abilities in real-time”.
“In the South Seas there is a cargo cult of people. During the war they saw airplanes land with lots of good materials, and they want the same thing to happen now. So they’ve arranged to make things like runways, to put fires along the sides of the runways, to make a wooden hut for a man to sit in, with two wooden pieces on his head like headphones and bars of bamboo sticking out like antennas—he’s the controller -and they wait for the airplanes to land. They’re doing everything right. The form is perfect. It looks exactly the way it looked before. But it doesn’t work. No airplanes land. So I call these things cargo cult science, because they follow all the apparent precepts and forms of scientific investigation, but they’re missing something essential, because the planes don’t land.Now it behooves me, of course, to tell you what they’re missing… It is a kind of scientific integrity, a principle of scientific thought that corresponds to a kind of utter honesty—a kind of leaning over backwards. For example, if you’re doing an experiment, you should report everything that you think might make it invalid—not only what you think is right about it: other causes that could possibly explain your results; and things you thought of that you’ve eliminated by some other experiment, and how they worked—to make sure the other fellow can tell they have been eliminated… In summary, the idea is to try to give all of the information to help others to judge the value of your contribution; not just the information that leads to judgment in one particular direction or another.”[72] (p. 1405)
5. Concluding Remarks
5.1. The Legal Challenge to EFPA Guidelines
5.2. The “Next Generation” Assessment Challenge
Conflicts of Interest
References
- EFPA Review Model for the Description and Evaluation of Psychological and Educational Tests, version 4.2.6. 2013. Available online: http://www.efpa.eu/download/650d0d4ecd407a51139ca44ee704fda4 (accessed on 2 November 2017).
- Michell, J. Quantitative science and the definition of measurement in Psychology. Br. J. Psychol. 1997, 88, 355–383. [Google Scholar] [CrossRef]
- Maraun, M.D. Measurement as a normative practice: Implications of Wittgenstein’s philosophy for measurement in Psychology. Theory Psychol. 1998, 8, 435–461. [Google Scholar] [CrossRef]
- Buntins, M.; Buntins, K.; Eggert, F. Clarifying the concept of validity: From measurement to everyday language. Theory Psychol. 2017, 27, 703–710. [Google Scholar] [CrossRef]
- Barrett, P.T.; Rolland, J.P. The Meta-Analytic Correlation between the Big Five Personality Constructs of Emotional Stability and Conscientiousness: Something is Not Quite Right in the Woodshed. 2009. Available online: http://www.pbarrett.net/stratpapers/metacorr.pdf (accessed on 2 November 2017).
- Pace, V.L.; Brannick, M.T. How similar are personality scales of the “same” construct? A meta-analytic investigation. Personal. Individ. Differ. 2010, 49, 669–676. [Google Scholar] [CrossRef]
- Michell, J. Alfred Binet and the concept of heterogeneous orders. Front. Quant. Psychol. Meas. 2012, 3, 261. [Google Scholar] [CrossRef] [PubMed]
- Wood, R. Fitting the Rasch model—A heady tale. Br. J. Math. Stat. Psychol. 1978, 31, 27–32. [Google Scholar] [CrossRef]
- Michell, J. Item Response Models, pathological science, and the shape of error. Theory Psychol. 2004, 14, 121–129. [Google Scholar] [CrossRef]
- Maraun, M.D. Latent Variable Modeling: Myths and Confusions. E-book. 2007. Available online: http://www.sfu.ca/~maraun/myths-and-confusions.html (accessed on 2 November 2017).
- Michell, J. The psychometricians’ fallacy: Too clever by half? Br. J. Math. Stat. Psychol. 2009, 62, 41–55. [Google Scholar] [CrossRef] [PubMed]
- Michell, J. The constantly recurring argument: Inferring quantity from order. Theory Psychol. 2012, 22, 255–271. [Google Scholar] [CrossRef]
- Tryon, W.W. Underreliance on mechanistic models: Comment on Ferguson. Am. Psychol. 2016, 71, 505–506. [Google Scholar] [CrossRef] [PubMed]
- Ferguson, C.J. Everybody knows psychology is not a real science: Public perceptions of psychology and how we can improve our relationship with policymakers 2015, the scientific community, and the general public. Am. Psychol. 2015, 70, 527–542. [Google Scholar] [CrossRef] [PubMed]
- Pennington, B.F. Explaining Abnormal Behavior: A Cognitive Neuroscience Perspective; Guilford Press: New York, NY, USA, 2014. [Google Scholar]
- Tryon, W.W. Cognitive Neuroscience and Psychotherapy: Network Principles for a Unified Theory; Academic Press: New York, NY, USA, 2014. [Google Scholar]
- Tryon, W.W.; Hoffman, J.; McKay, D. Neural networks as explanatory frameworks of psychopathology and its treatment. In Mechanisms of Syndromes and Treatment for Psychological Problems; McKay, D., Abramowitz, J.S., Storch, E., Eds.; Wiley: Chichester, UK, 2016. [Google Scholar]
- Freedman, D.A.; Berk, R.A. Statistical assumptions as empirical commitments. In Law, Punishment, and Social Control: Essays in Honor of Sheldon Messinger, 2nd ed.; Blomberg, T.G., Cohen, S., Eds.; Aldine de Gruyter: Piscataway, NJ, USA, 2003; pp. 235–254. [Google Scholar]
- Byrne, B.M.; Shavelson, R.J.; Muthén, B.O. Testing for equivalence of factor covariance and mean structures: The issue of partial measurement invariance. Psychol. Bull. 1989, 105, 456–466. [Google Scholar] [CrossRef]
- Meredith, W. Measurement invariance, factor analysis and factorial invariance. Psychometrika 1993, 58, 525–543. [Google Scholar] [CrossRef]
- Barrett, P.T. Measurement Invariance and Latent Variable Theory in Cross-Cultural Psychology. Presentation Given at SIOP 2009; New Orleans, LA, USA, 2009. Available online: http://www.pbarrett.net/SIOP/SIOP2009.html (accessed on 2 November 2017).
- Borsboom, D.; Mellenbergh, G.J. True scores, latent variables, and constructs: A comment on Schmidt and Hunter. Intelligence 2002, 30, 505–514. [Google Scholar] [CrossRef]
- Trendler, G. Conjoint Measurement undone. Theory Psychol. 2017. submitted. [Google Scholar]
- Borbsoom, D.; Cramer, A.O.J.; Kievit, R.A.; Scholten, A.Z.; Franic, S. The end of construct validity. In The Concept of Validity: Revisions, New Directions, and Applications; Lissitz, R.W., Ed.; Information Age Publishing: Charlotte, NC, USA, 2009; Chapter 7; pp. 135–170. [Google Scholar]
- Michell, J. Invalidity in Validity. In The Concept of Validity: Revisions, New Directions, and Applications; Lissitz, R.W., Ed.; Information Age Publishing: Charlotte, NC, USA, 2009; Chapter 6; pp. 111–133. [Google Scholar]
- Michell, J. Constructs, inferences, and mental measurement. New Ideas Psychol. 2013, 31, 13–21. [Google Scholar] [CrossRef]
- Maraun, M.D.; Gabriel, S.M. Illegitimate concept equating in the partial fusion of construct validation theory and latent variable modeling. New Ideas Psychol. 2013, 31, 32–42. [Google Scholar] [CrossRef]
- Slaney, K.L.; Garcia, D.A. Constructing psychological objects: The rhetoric of constructs. J. Theor. Philos. Psychol. 2015, 35, 244–259. [Google Scholar] [CrossRef]
- Slaney, K. Validating Psychological Constructs: Historical, Philosophical, and Practical Dimensions; Palgrave Macmillan: London, UK, 2017. [Google Scholar]
- Michell, J. Measurement in Psychology: Critical History of a Methodological Concept; Cambridge University Press: Cambridge, UK, 1999; ISBN 0-521-62120-8. [Google Scholar]
- Vautier, S.; Veldhuis, M.; Lacot, E.; Matton, N. The ambiguous utility of psychometrics for the interpretative foundation of socially relevant avatars. Theory Psychol. 2012, 22, 810–822. [Google Scholar] [CrossRef] [Green Version]
- Lacot, E.; Afzali, M.H.; Vautier, S. Test validation without measurement: Disentangling scientific explanation of item responses and justification of focused assessment policies based on test data. Eur. J. Psychol. Assess. 2016, 32, 204–214. [Google Scholar] [CrossRef]
- Postman, N. Social science as theology. ETC Rev. Gen. Semant. 1984, 41, 22–32. [Google Scholar]
- Zenker, F.; Witte, E. From discovery to justification: Outline of an ideal research program for empirical psychology. Front. Psychol. Quant. Psychol. Meas. 2017, 8, 1847. [Google Scholar] [CrossRef]
- Briner, R.B.; Rousseau, D.M. Evidence-based I-O psychology: Not there yet. Ind. Organ. Psychol. Perspect. Sci. Pract. 2011, 4, 3–22. [Google Scholar] [CrossRef]
- Wikipedia. Turning a Blind Eye. Available online: https://en.wikipedia.org/wiki/Turning_a_blind_eye (accessed on 2 November 2017).
- NIST Reference on Constants, Units, and Uncertainty. Available online: https://physics.nist.gov/cuu/Units/units.html. (accessed on 2 November 2017).
- Quotation. Available online: https://www.quotes.net/mquote/733006. (accessed on 2 November 2017).
- Trafimow, D. Using the coefficient of confidence to make the philosophical switch from a posteriori to a priori inferential statistics. Educ. Psychol. Meas. 2017, 77, 831–854. [Google Scholar] [CrossRef]
- Barrett, P.T. Rasch scaling, yet another incomprehensible test theory or something far more dangerous? In Proceedings of the British Psychological Society’s Division of Occupational Psychology Conference: Assessment in the Millennium: Beyond Psychometrics, Birkbeck University of London, London, UK, 20 November 1998; Available online: http://www.pbarrett.net/presentations/BPS-rasch_98.pdf (accessed on 2 November 2017).
- Fisher, W.P., Jr. Objectivity in measurement: A philosophical history of Rasch’s separability theorem. In Objective Measurement: Theory into Practice; Wilson, M., Ed.; Ablex Publishing Corp.: Norwood, NJ, USA, 1992; Chapter 3; pp. 29–58. [Google Scholar]
- Haig, B. An abductive theory of scientific method. Psychol. Methods 2005, 10, 371–388. [Google Scholar] [CrossRef] [PubMed]
- Sherry, D. Thermoscopes, thermometers, and the foundations of measurement. Stud. Hist. Philos. Sci. Part A 2011, 42, 509–524. [Google Scholar] [CrossRef]
- Mari, L. A quest for the definition of measurement. Measurement 2013, 46, 2889–2895. [Google Scholar] [CrossRef]
- Raynor, M.E.; Ahmed, M. Three Rules for making a company really great. Harvard Bus. Rev. 2013, 91, 108–117. [Google Scholar]
- Sternberg, R.J.; Williams, W. You proved our point better than we did: A reply to our critics. Am. Psychol. 1998, 53, 576–577. [Google Scholar] [CrossRef]
- Darlington, R.B. Range restriction and the Graduate Record Examination. Am. Psychol. 1998, 53, 572–573. [Google Scholar] [CrossRef]
- Sternberg, R.J.; Williams, W.M. Does the Graduate Record Examination predict meaningful success in the graduate training of psychologists? A case study. Am. Psychol. 1997, 52, 630–641. [Google Scholar] [CrossRef] [PubMed]
- Grigorenko, E.L.; Sternberg, R.J. Dynamic Testing. Psychol. Bull. 1998, 124, 75–111. [Google Scholar] [CrossRef]
- Irvine, S.H. Computerised Test Generation for Cross-National Military Recruitment: A Handbook; IOS Press: Amsterdam, The Netherland, 2014. [Google Scholar]
- Trendler, G. Measurement Theory, Psychology and the Revolution That Cannot Happen. Theory Psychol. 2009, 19, 579–599. [Google Scholar] [CrossRef]
- Trendler, G. Measurement in psychology: A case of ignoramus et ignorabimus? A rejoinder. Theory Psychol. 2013, 23, 591–615. [Google Scholar] [CrossRef]
- Carpenter, P.A.; Just, M.A.; Shell, P. What one intelligence test measures: A theoretical account of the processing in the Raven Progressive Matrices Test. Psychol. Rev. 1990, 97, 404–431. [Google Scholar] [CrossRef] [PubMed]
- Braithwaite, D.W.; Pyke, A.A.; Siegler, R.S. A computational model of fraction arithmetic. Psychol. Rev. 2017, 124, 603–625. [Google Scholar] [CrossRef] [PubMed]
- Lehrl, S.; Fischer, B. The Basic Parameters of Human Information Processing - their role in the determination of intelligence. Personal. Individ. Differ. 1988, 9, 883–896. [Google Scholar] [CrossRef]
- Lehrl, S.; Fischer, B. A basic information psychological parameter for reconstruction of concepts of intelligence (BIP). Eur. J. Personal. 1990, 4, 259–286. [Google Scholar] [CrossRef]
- Roberts, R.D.; Pallier, G.; Stankov, L. The Basic Information Processing (BIP) Unit, mental speed and human cognitive abilities: Should the BIP R.I.P? Intelligence 1996, 23, 133–155. [Google Scholar] [CrossRef]
- Barrett, P.T. Using Psychometric Test Scores: When Outcomes Are Critical. 2011. Available online: http://www.pbarrett.net/stratpapers/critical_outcome_tests.pdf (accessed on 2 November 2017).
- Researchgate. Beyond Questionable Research Methods: The Role of Intellectual Honesty in Research Credibility. Available online: https://www.researchgate.net/publication/313038713_Beyond_Questionable_Research_Methods_The_Role_of_Intellectual_Honesty_in_Research_Credibility?focusedCommentId=5974be99b53d2ff30bd63f42 (accessed on 29 December 2017).
- Schmidt, F.L. Beyond questionable research methods: The role of omitted relevant research in the credibility of research. Arch. Sci. Psychol. 2017, 5, 32–41. [Google Scholar] [CrossRef]
- Raynor, M.E.; Gurumurthy, R.; Ahmed, M. Growths Triple Crown: When it comes to exceptional performance, the best companies do not make trade-offs: They break them. Deloitte Rev. 2011, 9, 25–39. [Google Scholar]
- Freedman, D.A. Statistical models and shoe leather. Sociol. Methodol. 1991, 21, 291–313. [Google Scholar] [CrossRef]
- Borsboom, D.; Mellenbergh, G.J.; Van Heerden, J. The concept of validity. Psychol. Rev. 2004, 111, 1061–1071. [Google Scholar] [CrossRef] [PubMed]
- Kelley, T.L. Interpretation of Educational Measurements; Macmillan: New York, NY, USA, 1927. [Google Scholar]
- US Supreme Court Judgement No. 15-797. Moore vs. Texas. Certiorari to the Court of Criminal Appeals of Texas, Argued 29 November 2016—Decided 28 March 2017. Available online: https://www.supremecourt.gov/opinions/16pdf/15–797_n7io.pdf (accessed on 2 November 2017).
- Cipresso, P.; Matic, A.; Giakoumis, D.; Ostrovsky, Y. Advances in Computational Psychometrics. Comput. Math. Methods Med. 2015, 2015, 418683. [Google Scholar] [CrossRef] [PubMed]
- Von Davier, A.A. Computational psychometrics in support of collaborative educational assessments. J. Educ. Meas. 2017, 54, 3–11. [Google Scholar] [CrossRef]
- Polyak, S.T.; von Davier, A.A.; Peterschmidt, K. Computational Psychometrics for the Measurement of Collaborative Problem Solving Skills. 2017. Available online: https://actnext.org/research-and-projects/computational-psychometrics-for-the-measurement-of-collaborative-problem-solving-skills/ (accessed on 2 November 2017).
- Webster, C.D.; Harris, G.T.; Rice, M.E.; Cormier, C.; Quinsey, V.L. The Violence Prediction Scheme: Assessing Dangerousness in High Risk Men; Centre of Criminology, University of Toronto: Toronto, ON, Canada, 1994. [Google Scholar]
- Feynman, R.P. Cargo Cult Science: Some remarks on science, pseudoscience, and learning how not to fool yourself. Eng. Sci. 1974, 37, 10–13. [Google Scholar]
- Feynman, R.P. Surely You’re Joking, Mr. Feynman!: Adventures of a Curious Character; W. W. Norton & Co.: New York, NY, USA, 1985. [Google Scholar]
- Wilson, J.R. Doctoral Colloquium Keynote Address: Conduct, Misconduct, and Cargo Cult Science. In Proceedings of the 1997 Winter Simulation Conference; Andradottir, S., Healy, K.J., Withers, D.H., Nelson, B.L., Eds.; IEEE Computer Society: Washington, DC, USA, 1997; Available online: https://www.researchgate.net/publication/2819932_Doctoral_Colloquium_Keynote_Address_Conduct_Misconduct_And_Cargo_Cult_Science (accessed on 2 November 2017).
- Ziskin, J. Coping with Psychiatric and Legal Testimony; Law and Psychology Press: Beverley Hills, CA, USA, 1970. [Google Scholar]
- Ziskin, J. Coping with Psychiatric and Psychological Testimony: Practical Guidelines, Cross-Examination and Case Illustrations, 5th ed.; Law and Psychology Press: Beverley Hills, CA, USA, 1995. [Google Scholar]
© 2017 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Barrett, P. The EFPA Test-Review Model: When Good Intentions Meet a Methodological Thought Disorder. Behav. Sci. 2018, 8, 5. https://doi.org/10.3390/bs8010005
Barrett P. The EFPA Test-Review Model: When Good Intentions Meet a Methodological Thought Disorder. Behavioral Sciences. 2018; 8(1):5. https://doi.org/10.3390/bs8010005
Chicago/Turabian StyleBarrett, Paul. 2018. "The EFPA Test-Review Model: When Good Intentions Meet a Methodological Thought Disorder" Behavioral Sciences 8, no. 1: 5. https://doi.org/10.3390/bs8010005
APA StyleBarrett, P. (2018). The EFPA Test-Review Model: When Good Intentions Meet a Methodological Thought Disorder. Behavioral Sciences, 8(1), 5. https://doi.org/10.3390/bs8010005