Richard P. Phelps

Richard P. Phelps' Home Page

Richard P. Phelps's Web Page
-----------------------------------------------------------------------------------------------------------------------------------------------
Education: PhD: U.Pennsylvania (Wharton); Master's: Harvard, Indiana; Bachelor's: Washington (St. Louis)
Postings:US-Romania, US-Chile Fulbright Commissions, The Association of Boarding Schools (TABS); ACT; National/Pearson Evaluation Systems, OECD, Westat, AIR; U.S. General Accounting Office (GAO). Indiana Education Department; Harvard Business School, Professeur de Mathématiques, Séminaire St. Augustin, Baskouré, Burkina Faso

Fulbright Press Release: Richard Phelps, Ph.D Receives Fulbright Specialist Award to Romania at University of Bucharest

April 2023: The Malfunction of US Education Policy: Elite Misinformation, Disinformation, and Selfishness Roman & Littlefield, 188 pages.

Reviews of The Malfunction of US Education Policy (.docx file)

Quotations from The Malfunction of US Education Policy (.docx file)


Appendices for book, The Malfunction of U.S. Education Policy: Elite Misinformation, Disinformation, and Selfishness




Memberships:
Founder, Editor: Nonpartisan Education Group;
Vice-President, Institute for Objective Policy Assessment (IOPA);
Chair, Committee on Informing Assessment Policy and Practice, 2019--2024 (National Council on Measurement in Education);
Groupe International de Recherche en Ingénierie des Evaluations en Formation, 2018+;
External Education Expert (on Standardized Testing) for the Learning Scientists;
National Association of Nonpartisan Reformers (NANR);
International Test Commission (ITC);
National Association of Scholars (NAS);
Association for Psychological Science (APS)

Awards, Fellowships:
Fulbright Scholar: US Fulbright Specialist Program, 2012--2018; 2022–2025
Fellow: Psychophysics Laboratory;
Research Fellow: Educational Testing Service;
New Scholars Grant, American Education Finance Association;
Doctoral Scholars Award, National Center for Education Statistics;
Morganthau Fellowship, Wharton School of Commerce and Finance (U. Pennsylvania);
Kennedy Fellow, Harvard Kennedy School of Government;
Gustav Harris Scholarship (Harvard University)

Contact: richard (at) nonpartisaneducation (dot) org

Richard P Phelps on ResearchGate
Follow me on Academia.edu

----------------------------------------------------------------------------------------------------------------------------------------------------

Links and citations are listed in these categories, respectively, below:
  • Fordham Institute's Pretend Research # How PARCC's False Rigor Stunts the Academic Growth of All Students # The Revenge of K-12: How Common Core and the new SAT lower college standards in the U.S. # Test Frequency, Stakes, and Feedback in Student Achievement: A Meta-Regression # The Effect of Testing on Student Achievement, 1910-2010 # Correcting Fallacies # Standardized Testing Primer # Defending Standardized Testing # Kill the Messenger
  • International # Research Ethics # Testing # Education Finance # Methodological/Evaluations # Source of Lake Wobegon
  • Not on the Web: (International) # (Testing) # (Education Finance) # (Methodological/Evaluations)

  • Critically Speaking Science Podcasts "we separate fact from fantasy for topics both current and controversial"
  • Richard Phelps: Is our education system failing us? Critically Speaking (Apple podcast)
  • Should Everyone Go to College? Critically Speaking (Apple podcast)
  • Dr. Richard Phelps: No More College Admission Test? Critically Speaking (Apple podcast)

  • Common Core Collaborators: Six Organizational Portraits
  • Historical, financial and media analyses of the organization that spawned the Common Core Initiative, the two copyright holders, two of the paid proselytizers, and the delivery vehicle, where the reputed CC "architect" now runs things (for a cool annual salary of well over a $million).
    The Organization Named Achieve: Cradle of Common Core Cronyism (26 pp)
    The Council of Chief State School Officers and National Governors Association: Whom do they serve? (23 pp)
    Real Clear Propaganda: Bellwether's Education News Bias (28 pp)
    The Thomas B. Fordham Foundation and Institute: Influence for Hire (30 pp)
    Does College Board Deserve Public Subsidies? (45 pp)

  • Guest Post in the Learning Scientists Blog: What causes test-score inflation? Comparing two theories.

  • Essays published in Education News:
  • One Size Fits All National Tests Not Deeper or More Rigorous
    Republicans Muster Narrow Expertise on NCLB Rewrite
    Are US Education Journalists Objecctive, Or Players in the Game?
    On Test Security, the GAO Could Do More
    Campbell's Law is Like the Soup, Ubiquitous, Innocuous
    What Happened at the OECD?
    OECD Encourages World to Adopt Failed US Ed Programs
    The School-Test Publisher Complex
    The (Secret) Document That Drives Standardized Testing
    The Rot Festers: Another National Research Council Report on Testing

  • Interview with Richard Phelps on Education News, June 11, 2015, "Are education journalists objective or players in the game?"

  • Nonpartisan Education Review

  • Review of the OECD's Synergies for better learning: an international perspective on evaluation and assessment
  • The Rot Spreads Worldwide: The OECD: Taken In and Taking Sides
  • The Rot Festers: Another National Research Council Report on Testing

  • Censorship Has Many Fathers: A progeny of excuses for suppressing information that one dislikes

  • Dismissive Reviews and Citation Cartels in Education Policy Research, A List


  • Presentations:
  • US Education Journalists Source Research Expertise, NCME Informing Assessment Policy 22023 Winter Webinar: Media Coverage of Educational Testing, January 19, 2023
  • Dismissive Reviews, Citation Cartels, and the Replication Crisis (.ppt), University of Bucharest, Interdisciplinary School of Doctoral Studies 17 May, 2023 (promo poster)
  • Over- and Under-used Criticisms of Standardized Tests (.ppt), University of Bucharest, 29 May, 2023 (promo poster)
  • Why Standardized Tests? (.ppt), Romania (online webinar), 26 May, 2023 (promo poster)
  • Standardized Testing: The Interplay of Achievement & Aptitude Testing (.ppt), University of Bucharest, 25 May, 2023
  • US Education Journalists Source Research Expertise (.ppt), 2023 NCME Winter Webinars
  • The Successful Degradation of Evidence on Educational Testing in the United States (.ppt), 2021 ITC Virtual Conference
  • It's a Myth: High Stakes Cause Test Score Inflation (.ppt), 2018 ITC Conference, Montréal
  • From aptitude to achievement: How's that going? for Symposium: University admission testing in Chile: Current controversies and future directions, (.pptx), 2018 ITC Conference, Montréal
  • Why Standardized Tests? (.pptx), University of Bucharest, May 2018
  • Comparing achievement and aptitude tests for university admission, (.pptx), Departmento de Evaluacion Medición y Registro, Universidad de Chile, Santiago, December 2017
  • PSU : El Desafío del Cambio, (.pptx), Centro de Investigación Avanzada en Educación, Universidad de Chile, Santiago, diciembre 2016
  • Boarding School: Benefits and Drawbacks, Part 1, (.pptx), Symposium on Best Practices in Educational Justice, Boarding Schools for Marginalized Youth, Kassel Universitat, Hannover, Germany, November 2016
  • Boarding School: Benefits and Drawbacks, Part 2, (.pptx), Symposium on Best Practices in Educational Justice, Boarding Schools for Marginalized Youth, Kassel Universitat, Hannover, Germany, November 2016
  • Designing an Assessment System, (.pptx), International Research-to-Practice Conference, Nazarbayev Intellectual Schools AEO, Astana, Kazakhstan, October 2016
  • Fortalezas y debilidades de las pruebas estandarizadas como mecanismos inclusivos de admisión a la educación superior, (.pptx), Pontificia Universidad Católica de Chile, Santiago, Programa Talento y Inclusión, diciembre 2015
  • The Source of Lake Wobegon, (.ppt)
  • No Child Left Behind, Common Core, and the Lost Benefits of Effective Testing, (.pptx), Governor's Council on Common Core Review, Arkansas State Capitol, May 2015
  • Testing in the classroom: Using tests to promote learning (.pptx), Universidad Finis Terrae, Santiago, Chile, January 2014
  • L'effet de testing sur les résultats scolaires : 1910-2010, (.ppt), Association Mondiale des Sciences d'Éducation, 17th Congrès, Reims, France, June 2012
  • The effect of testing on student achievement: 1910-2010, (.ppt), International Test Commission, 8th Conference, Amsterdam, July, 2012
  • L'effet de tests standardisés sur les résultats scolaires des éléves : Méta-analyses et résumé de la recherche, (.pdf), pour le Colloque sur l'évaluation des performances complexes, XVI'me congrès mondial des sciences de l'éducation, Monterrey, Mexique, juin 2010
  • Worse than plagiarism: Firstness claims and dismissive reviews, (.ppt), International Test Commission, 7th Conference, Hong Kong, July 2010; Center for Academic Integrity, 2009 Conference, St. Louis
  • Economic Perspectives on Testing, (.ppt), for the Research in Educational Measurement Program, U. Massachusetts, Amherst, 2003
  • ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------


    Education Writers Association Casts Its Narrowing Gaze on Boston, May 1-3, April 2016

    Fordham Institute's Pretend Research, Pioneer Institute report [download], February 2016

    The Fordham Institute has long been at work on a study of the relative quality of tests produced by the two Common Core-aligned and federally funded consortia (PARCC and SBAC), ACT (Aspire), and the Commonwealth of Massachusetts (MCAS). What Fordham has produced is only in the most superficial way an actual analysis. In fact, it reads more like propaganda and lacks the basic elements of objective research.

    It takes only a little digging under the surface to reveal pervasive conflicts of interest, a one-sided sourcing of evidence, and a research design so slanted it cannot stand against any scrutiny. In developing their supposedly analytic comparisons of PARCC, SBAC, Aspire and MCAS, the authors do not employ standard test evaluation criteria, organizations, or reviewers. Instead, they employ criteria developed by the Common Core co-copyright holder, organizations paid handsomely in the past by Common Core's funders, and predictable reviewers who have worked for them before. The authors also fill the report with the typical vocabulary and syntax of Common Core advertising: positive-sounding adjectives and adverbs are attached to everything Common Core, and negative-sounding adjectives and adverbs are attached to the alternatives.

    No reader should take the report seriously; those who produced the report did not. Fordham Institute used to do serious work; in the area of assessments and standards, sadly, that is no longer the case.


    Fordham Report Predictable, Conflicted, November 2015

    Setting Academic Performance Standards: MCAS vs. PARCC, November 2015

    How PARCC's False Rigor Stunts the Academic Growth of All Students, Pioneer Institute report [download] (written with Mark McQuillan and Sandra Stotsky), October 2015.

    ...compares PARCC, which is based on Common Core's K-12 English and math standards, to pre-Common Core MCAS reading and writing tests. Among its many findings, the study demonstrates that PARCC fails to meet the accountability provisions set forth in the state's Memorandum of Understanding with the U.S. Department of Education in order to qualify for Race to the Top funding in 2010.

    The pre-Common Core MCAS tests were chosen in part because MCAS is now based on Common Core standards, not the state's pre-2011 standards. Pre-Common Core MCAS tests were also chosen because of the dumbing down of 10th grade MCAS tests. What was at the very least a failure by the state to maintain the academic rigor of 10th grade MCAS tests is one reason why the authors recommend that MCAS 2.0 be developed by an entity independent of the state Department of Elementary and Secondary Education.

    The authors also dispute claims that PARCC could simultaneously determine whether students are academically eligible for a high school diploma and ensure college readiness. High school is radically different than college, and the academic demands of various post-secondary programs vary dramatically. Our international competitors use different tests for high school graduation and entrance to post-secondary education.

    Supporters also claim PARCC does a better job of testing "higher-order thinking." In fact, its new types of test items are not research-based and not very good. They are often difficult to navigate and what passes for testing higher-order thinking are simply multi-step problems.

    At the root of PARCC's weaknesses are the Common Core standards to which they are tied. The authors call on the Commonwealth to phase out Common Core and PARCC, and to base a revised MCAS on Massachusetts' pre-Common Core curriculum frameworks, updated by pertinent new research.

    The Revenge of K-12: How Common Core and the new SAT lower college standards in the U.S., Pioneer Institute report [download] (written with R. James Milgram), September 2014.

    It is now clear that the original promise to anchor K-12 education to higher education and backmap the Common Core Mathematics Standards (CCMS) from the upper grades down to the primary grades was empty rhetoric. Higher education has scarcely been involved at all, with the exception of the institutions that agreed to place high school students who pass a Common Core-based high school examination directly into credit-bearing freshman coursework (without remediation) in return for their states receiving "Race to the Top" grant funds.

    Because the CCMS are standards for all public school students in this country, regardless of achievement level, they are low standards, topping out at about the level of a weak Algebra II course. And because this level is to determine "college readiness" as they define it (which is not remotely what our public four year college and universities currently assume it to be), it is apt to mean fewer high school students taking advanced mathematics and science coursework before they go to college, more college freshmen with even less knowledge of mathematics than currently, and more college credit-bearing courses set at an international level of seventh or eighth grade.

    However, the greatest harm to higher education may accrue from the alignment of the SAT to Common Core's high school standards, converting the SAT from an adaptable test predictive of college work to an inflexible retrospective test aligned to and locking in a low level of mathematics. This means that future SAT scores will be less informative to college admission counselors than they now are, and that the SAT will lose its role in locating students with high STEM potential in high schools with weak mathematics and science instruction.


    ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

    Test Frequency, Stakes, and Feedback in Student Achievement: A Meta-Regression Evaluation Review, August 2019)

    Background: Test frequency, stakes associated with educational tests, and feedback from test results have been identified in the research literature as relevant factors in student achievement. Objectives: Summarize the separate and joint contribution to student achievement of these three treatments and their interactions via multivariable meta-analytic techniques using a database of English-language studies spanning a century (1910–2010), comprising 149 studies and 509 effect size estimates. Research design: Analysis employed robust variance estimation. Considered as potential moderators were hundreds of study features comprising various test designs and test administration, demographic, and source document characteristics. Subjects: Subjects were students at all levels, from early childhood to adult, mostly from the United States but also eight other countries. Results: We find a summary effect size of 0.84 for the three treatments collectively. Further analysis suggests benefits accrue to the incremental addition of combinations of testing and feedback or stakes and feedback. Moderator analysis shows higher effect sizes associated with the following study characteristics: more recent year of publication, summative (rather than formative) testing, constructed (rather than selected) item response formats, alignment of subject matter between pre- and posttests, and recognition/recall (rather than core subjects, art, or physical education). Conversely, lower effect sizes are associated with postsecondary students (rather than early childhood–upper secondary), special education population, larger study population, random assignment (rather than another sampling method), use of shadow test as outcome measure, designation of individuals (rather than groups) as units of analysis, and academic (rather than corporate or government) research.


    ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

    The Effect of Testing on Student Achievement, 1910-2010 International Journal of Testing, January 2012)

    This article summarizes research on the effect of testing on student achievement as found in English-language sources, comprising several hundred studies conducted between 1910 and 2010. Among quantitative studies, mean effect sizes range from a moderate d ≈ 0.55 to a fairly large d ≈ 0.88, depending on the way effects are aggregated or effect sizes are adjusted for study artifacts. Testing with feedback produces the strongest positive effect on achievement. Adding stakes or frequency also strongly and positively affects achievement. Survey studies produce effect sizes above 1.0. Ninety-three percent of qualitative studies analyzed also reported positive effects.

    Acknowledgements: Thanks to three anonymous reviewers and editor Steve Sireci for their excellent suggestions. Thanks also to individuals who aided the research, including: Jeff Allen; Kurt Burkum; Richard Coley; Lori Dockery; John Hattie; George Johanson; Kathy Lynch; Karen McQuillan; Katie Paul; John Poggio; Terris Raiford; Jacqueline Snider; Gang Wang; and Karen Zimmerman. Furthermore, credit goes to those who conducted research reviews on relevant subtopics (in chronological order): Panlasigui (1928); Ross (1942); Kirkland (1971); Proger & Mann (1973); Jones (1974); Bjork (1975); Peckham & Roe (1977); Wildemuth (1977); Jackson & Battiste (1978); Kulik, Kulik, Bangert-Drowns, & Schwalb (1983-1991); Natriello & Dornbusch (1984); Dawson & Dawson (1985); Levine (1985); Resnick & Resnick (1985); Guskey & Gates (1986); Hembree (1987); Crooks (1988); Dempster (1991); Cameron & Pierce (1994); Adams & Chapman (2002); Locke & Latham (2002); Roediger & Karpicke (2006); J. Lee (2007); and Basol & Johanson (2009).


    ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

    Correcting Fallacies about Educational and Psychological Testing (American Psychological Association, December 2008)

    Standardized testing bears the twin burden of controversy and complexity and is difficult for many to understand either dispassionately or technically. In response to this reality, Richard P. Phelps and a team of well-noted measurement specialists describe the current state of public debate about testing across fields, explain and refute the primary criticisms of testing, acknowledge the limitations and undesirable consequences of testing, provide suggestions for improving testing practices, and present a vigorous defense of testing as well as a practical vision for its promise and future.
    Those who are charged with translating the science of testing into public information and policy, including administrators, social scientists, test publishers, professors, and journalists who specialize in education and psychology, will find a wealth of valuable information here with which to balance the debate.

    Table of Contents:
    Foreword, Thomas Oakland
    Introduction and Overview
    Logical Fallacies Used to Dismiss the Evidence on Intelligence Testing, Linda S. Gottfredson
    Psychological Diagnostic Testing: Addressing Challenges in Clinical Applications of Testing, Janet F. Carlson & Kurt F. Geisinger
    Educational Achievement Testing: Critiques and Rebuttals, Richard P. Phelps
    College Admission Testing: Myths and Realities in an Age of Admissions Hype, Wayne J. Camara
    Criticisms of Employment Testing: A Commentary, Ernest H. O'Boyle & Michael A. McDaniel
    Mission: Protect the Public: Licensure and Certification Testing in the 21st Century, Stephen G. Sireci & Ron K. Hambleton
    Mistaken Impressions of Large-Scale Cognitive Diagnostic Testing, Jacqueline P. Leighton
    Summary & Discussion, Richard P. Phelps & Linda S. Gottfredson


    ISBN: 9781433803925 or 1433803925
    LB3051.C6386 2009
    Dewey: 371.262 C824

    #18 on Best Sellers in Education list for 2009, Library Journal

    Reviews for Correcting Fallacies about Educational and Psychological Testing
    ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

    Standardized Testing Primer (Peter Lang, 2007)

    The Standardized Testing Primer provides non-specialists with a thorough overview of this controversial and complicated topic. It eschews the statistical details of scaling, scoring, and measurement that are widely available in textbooks and at testing organization Web sites, and instead describes standardized testing's social and political roles and its practical uses, who tests, when, where, and why. Topics include: an historical background of testing's practical uses in psychology, education, and the workplace; the varied structures of educational testing programs and systems across countries; the mechanics of test development and quality assurance; and current trends in test development and administration. A glossary and bibliography are also provided. The Standardized Testing Primer is an ideal text for teaching this subject to undergraduate and graduate students.


    Table of Contents:
    Introduction
    Aptitude or Achievement: Two Separate Historical Paths
    The Effects of Testing
    Mechanics of Test Development and Quality Assurance
    Conclusion
    References & Resources

    Find at Peter Lang Publishing, at Amazon, or at Barnes and Noble
    Search: FetchBook.info

    ISBN: 9780820497419 082049741X
    OCLC: 155715191 paper
    LOC: LB3051.P543
    Dewey: 371.26'2-dc22

    Reviews for Standardized Testing Primer ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

    Defending Standardized Testing (Lawrence Erlbaum, 2005)

    The education reform movement of the past two decades has focused on raising academic standards. Some standards advocates attach a testing mechanism to gauge the extent to which high standards are actually accomplished, whereas some critics accuse the push for standards and testing of impeding reform and perpetuating inequality. At the same time, the testing profession has produced advances in the format, accuracy, dependability, and utility of tests. Never before has obtaining such an abundance of accurate and useful information about student learning been possible. Meanwhile, the American public remains steadfast in support of testing to measure student performance and monitor the performance of educational systems.

    Many educational testing experts who acknowledge the benefits of testing also believe that those benefits have been insufficiently articulated. Although much has been written on standardized testing policy, most of the published material has been written by opponents. The contributing authors of this volume are both accomplished researchers and practitioners who are respected and admired worldwide. They bring to the project an abundance of experience working with standardized tests.

    The goal of Defending Standardized Testing is to: describe current standardized testing policies and strategies; explain many of the common criticisms of standardized testing; document the public support for, and the realized benefits of, standardized testing; acknowledge the limitations of, and suggest improvements to, testing practices; provide guidance for structuring testing programs in light of public preference and the No Child Left Behind Act; and present a defense of standardized testing and a practical vision for its promise and future.

    Table of Contents:
    J.J. Fremer, Foreword: The Rest of the Story
    R.P. Phelps, Persistently Positive: Forty Years of Public Opinion on Standardized Testing.
    G.J. Cizek, High-Stakes Testing: Contexts, Characteristics, Critiques, and Consequences.
    R.P. Phelps, The Rich, Robust Research Literature on Testing's Achievement Benefits.
    D. Goodman, R.K. Hambleton, Some Misconceptions About Large-Scale Assessments.
    S.G. Sireci, The Five Most Frequently Unasked Questions About Standardized Testing.
    G.K. Cunningham, Must High-Stakes Mean Low Quality?
    C. Buckendahl, R. Hunt, The Relationship Between the "Rules" and "Law" of Testing.
    L. Crocker, Teaching FOR the Test: How and Why Test Preparation is Appropriate.
    B.S. Plake, Doesn't Everybody Know That 70% Is Passing?
    K.F. Gelsinger, The Testing Industry, Ethnic Minorities and Those with Disabilities.
    D. Rogosa, API Awards and the Orange County Register Margin of Error Folly.
    M.L. Bourque, Leave no Standardized Test Behind.

    ISBN: 0805849114 cloth # 0805849122 paper

    Reviews for Defending Standardized Testing

    Available at your friendly neighborhood virtual bookstore, or through Psychology Press: # cloth # paper or eBooks

    Amazon # Barnes&Noble # AbeBooks
    Search: FetchBook.info

    ----------------------------------------------------------------------------------------------------------------------------------------------------------

    Kill the Messenger (Transaction, 1st Printing, 2003 (cloth); 2nd Printing, 2005 (paper); Psychology Press, 3rd Printing, 2017 (paper)

    Description: In response to public demand, new federal legislation requires testing of most students in the United States in reading and mathematics, for grades three through eight. In much of the country, this new order promotes an Increase in the amount of standardized testing. Many educators, parents, and policymakers who have paid little attention to testing policy issues in the past will now do so. They deserve to have better information on the topic than has generally been available, and Kill the Messenger is intended to fill this gap. Kill the Messenger is perhaps the most thorough and authoritative work in defense of educational testing ever written. Phelps points out that much research conducted by education insiders on the topic is based on ideological preference or profound self-interest. It is not surprising that they arrive at emphatically anti-testing conclusions. He notes that external and high stakes testing in particular attracts a cornucopia of invective. Much, if not most, of this hostile research is passed on to the public by journalists as if it were neutral, objective, and independent. Kill the Messenger describes the current debate, the players, their interests, and their positions. It explains and refutes many of the common criticisms of testing. It describes testing opponents strategies, through case studies of Texas and the SAT. It acknowledges testing's limitations, and suggests how it can be improved. It defends testing by comparing it with its alternatives. And finally, it outlines the consequences of losing the war on standardized testing.

    ISBN Nos. 0765801787 (cloth) # 1412805120 (2nd printing, paper) # 1412805124 (3rd printing, paper)
    Library of Congress call number: LB3051 .P54 2003
    Dewey Decimal System number: 371.26/2

    Association of American Publishers' Summary of Kill the Messenger

    Reviews for Kill the Messenger: The War on Standardized Testing

    Selected as a Profiler's Pick by YBP, Inc. and Academia Magazine from among their CORE 1000 recommended books for academic libraries, 2003/2004. "Reviews the debate over standardized testing requirements and argues in favor of testing."

    Available at your friendly neighborhood virtual bookstore: (or through Transaction Publishers)

    Abe Books # Amazon # Barnes&Noble

    SEARCH: FetchBook.info

    -------------------------------------------------------------------------------------------------------------

    The Source of Lake Wobegon

    John J. Cannell's late 1980's "Lake Wobegon" reports suggested widespread deliberate educator manipulation of norm-referenced standardized test (NRT) administrations and results, resulting in artificial test score gains. The Cannell studies have been referenced in education research since, but as evidence that high stakes (and not cheating or lax security) cause test score inflation. This article examines that research and Cannell's data for evidence that high stakes cause test score inflation. No such evidence is found. Indeed, the evidence indicates that, if anything, the absence of high stakes is associated with artificial test score gains. The variable most highly correlated with test score inflation is general performance on achievement tests, with traditionally low-performing states exhibiting more test score inflation, on low-stakes norm-referenced tests, than traditionally high-performing states, regardless of whether or not a state also maintains a high-stakes testing program. The unsupported high-stakes-cause-test-score-inflation hypothesis seems to derive from the surreptitious substitution of an antiquated definition of the term "high stakes" and a few studies afflicted with left-out-variable bias. The source of test-score inflation is lax test security, regardless the stakes of the assessment.

    Dan Koretz's Big Con [book review]

    It's a Myth: High Stakes Cause Test Score Inflation [PowerPoint presentation]

    The "teaching to the test" family of fallacies

    Teaching to the test: A very large red herring

    Guest Post at the Learning Scientists Blog: What Causes Test-score Inflation? Comparing Two Theories

    Measuring Up: What Educational Testing Really Tells Us, by Daniel Koretz [book review]

    The Source of Lake Wobegon [Updated]

    The Source of Lake Wobegon [Updated] (pdf file)

    The Source of Lake Wobegon [slide show]

    The Rocky Score-line of Lake Wobegon

    The Source of Lake Wobegon [original]

    The Source of Lake Wobegon [original, pdf]

    -------------------------------------------------------------------------------------------------------------

    RESEARCH ETHICS & DISSEMINATION:

    Phelps, R. P. (June 26, 2024). What a database of more than a thousand dismissive literature reviews can tell us, Retraction Watch

    Phelps, R. P. (2024). Dismissive Literature Reviews Reduce Understanding — So Why do Academics Keep Making Them?, Impact of Social Sciences blog, London School of Economics & Political Science, June 17, 2024.

    Phelps, R. P. (2024, February 9). Worse than Plagiarism: False Firstness Claims and Dismissive Literature Reviews. Minding the Campus.

    Phelps, R. P. (2023, June 16). Against Federally Funded Education Research Centers. Minding the Campus.

    Phelps, R. P. (2023, May 31). Have the Gates Foundation and Its Allies Purchased US Education Journalism?, Minding the Campus.

    Phelps, R. P. (2023, May 10). Are English Departments Really Dying? A closer look at the numbers reveals a more subtle story. The James G. Martin Center for Academic Renewal.

    Phelps, R. P. (January 19, 2023). US Education Journalists Source Research Expertise, NCME 2023 Winter Webinar Series, Informing Assessment Policy, Session 1: Media Coverage of Educational Testing

    Phelps, R. P. (November 15, 2022). Challenging the Academic Publisher Oligopoly, Technological and political changes may liberate scientific research. , The James G. Martin Center for Academic Renewal

    Phelps, R. P. (October 13, 2022). Subtractive Scholarship, Academe Blog

    Phelps, R. P. (May 17, 2022). How citation cartels give 'strategic scholars' an advantage: A simple model, Retraction Watch

    Phelps, R. P. (Fall, 2021). The Politicization of Education Research and the AERA, Academic Questions, National Association of Scholars

    Phelps, R. P. (July 16, 2021). Peer Review: A Tarnished Gold Standard, James G. Martin Center for Academic Renewal

    Phelps, R. P. (Summer 2021). In Defense of Sandra Stotdky, Academic Questions

    Phelps, R. P. (Mar. 23, 2021). Dismissive reviews: A cancer on the body of knowledge, Retraction Watch

    Phelps, R. P. (September 2020). Looking Back on DC Education Reform 10 Years After, Part 1: The Grand Tour Nonpartisan Education Review

    Phelps, R. P. (September 2020) .Looking Back on DC Education Reform 10 Years After, Part 2: Test Cheats Nonpartisan Education Review

    Phelps, R. P. (Summer 2020) Down the Memory Hole: Evidence on Educational Testing, Academic Questions

    Phelps, R. P. (2019, April 12). Strategic Partnerships - The end of independent viewpoints in education policy, Truth in American Education.

    Phelps, R. P. (2019, July 13). Education Next, the Fordham Institute, and Common Core, Missouri Education Watchdog.

    Phelps, R. P. (2019, July 13). Back to school means education news stories …for a while., Missouri Education Watchdog.

    Phelps, R. P. (2019, April 26). Citation cartels and US education journalism, The James G. Martin Center for Academic Renewal.

    Phelps, R. P. (2019, April 12). US Education's Dominant Research Method: Cherry Picking Evidence, Truth in American Education.

    Phelps, R. P. (2019, April 4). There's A Deeper Systemic Problem in the College Admissions Scandal No One Is Talking About, Independent Voter News.

    Phelps, R. P. (2019, March 27). The elitist strain in US education journalism, Missouri Education Watchdog.

    Phelps, R. P. (2018). To save the research literature, get rid of the literature review, Impact of Social Sciences blog, London School of Economics & Political Science, June 12, 2018.

    Phelps, R. (April 11, 2018). Nation’s Report Card: Common Core Delivering Education Stagnation, Independent Voters Network (IVN).

    Phelps, R. (April 3, 2018). There are Only Two Sides to US Education Policy (Thanks to the Parties), Independent Voters Network (IVN).

    Phelps, R. P. (December 21, 2017). The What Works Clearinghouse Isn't Working, Real Clear Education.

    Phelps, R. (September 6, 2017). The Gates Effect: Common Core Has Powerful Ally Keeping Unpopular Program Alive, Independent Voters Network (IVN).

    Phelps, R. P. (2015). "The Cork in the Bottle" The education establishment doesn't need to censor and suppress most of the evidence and information they do not like, the think tank elite does it for them. The think tank elite is the cork in the bottle that keeps the American public misinformed.

    Phelps, R. P. (2015). "Educational Morass Forever" It requires long seasoning in the bog for any person to develop a taste for which education information is accurate, which is myth, and which is just plain dishonest.

    Phelps, R. P. (2015). Kamenetz, A. (2015). The Test: Why our schools are obsessed with standardized testing, but you don't have to be. New York: Public Affairs. [book review]

    Phelps, R. P. (2014). The Gauntlet: Think tanks and federally funded centers misrepresent and suppress other education research

    Phelps, R. P. (2014). Synergies for better learning: an international perspective on evaluation and assessment, by the OECD [report review by R.P. Phelps], in Assessment in Education: Principles, Policy & Practice, v.21, i.4, 2014

    Myron Lieberman Tributes, Journal of School Choice, v.8, i.2, 2014: # Introduction: Myron Lieberman, Knowledgeable, Forthright, Independent, Rare # Educational Morass Forever # The Cork in the Bottle # A Myron Lieberman Bibliogrphy

    Phelps, R. P., & R. James Milgram. (2014, September). The Revenge of K-12: How Common Core and the new SAT lower college standards in the U.S., Pioneer Institute report.

    Phelps, R. P. (2013). The rot spreads worldwide: The OECD: Taken in and taking sides. New Educational Foundations, Winter, 2013.

    Phelps, R. P. (2012, October 2). The school-test publishers complex Education News

    Phelps, R. P. (2012, September 19). The (secret) document that drives standardized testing Education News

    Phelps, R. P. (2012, May 30). Dismissive reviews: Academe's memory hole, Academic Questions, v.25, n.2.

    Phelps, R. P. (2013). The rot festers: Another National Research Council report on testing. New Educational Foundations, Summer, 2012.

    Phelps, R. P. (2011, Autumn). Teach to the Test? Wilson Quarterly

    The Rocky Score-Line of Lake Wobegon

    The source of Lake Wobegon [updated] #

    The source of Lake Wobegon [updated, pdf]

    The National Research Council's Testing Expertise

    Measuring Up: What Educational Testing Really Tells Us, by Daniel Koretz [book review]

    Educational testing policy: Stuck between two political parties

    High Stakes: Testing for Tracking, Promotion, and Graduation, book review, Educational and Psychological Measurement, Vol.60, No.6, December 2000.

    Phelps, Richard P. (1999) "Education Establishment Bias: A Look at the National Research Council's Critique of Test Utility Studies." The Industrial-Organizational Psychologist, 36(4). Society of Industrial-Organizational Psychologists, April.

    Test-basher benefit-cost analysis," Network News & Views

    "The dissolution of education knowledge," Educational Horizons

    Censorship has many fathers



    INTERNATIONAL:

    Phelps, R. P. (2017). Boarding School: Benefits and Drawbacks, International Research-to-Practice Conference Nazarbayev Intellectual Schools, Astana, Kazakhstan, October, 2016.

    Phelps, R. P. (2014). Synergies for better learning: An international perspective on evaluation and assessment, by the OECD [report review by R.P. Phelps], in Assessment in Education: Principles, Policy & Practice, v.21, i.4, 2014.

    Phelps, R. P. (2014). Evaluaciones educacionales de gran escala en Chile: Son necesarias? CIPER, February 6, 2014.

    Phelps, R. P. (2013). The rot spreads worldwide: The OECD: Taken in and taking sides. New Educational Foundations, Winter, 2013.

    Phelps, R. P., Zenisky, A., Hambleton, R. K., & Sireci, S. G. (2010, March). Report to Ofqual: On the reporting of measurement uncertainty and reliability for U.S. educational and licensure tests

    Aubrey H. Wang, Ashaki B. Coleman, Richard J. Coley, & Richard P. Phelps Preparing Teachers Around the World Educational Testing Service, 2003

    Higher Education: An International Perspective (c)2003, Richard P. Phelps
    ...also available in paper and microfiche from ERIC, no. ED474484, Clearinghouse no. HE035711

    Organisation for Economic Co-operation and Development, Centre for Educational Research and Innovation (2000), Investing in Education: Analysis of the 1999 World Education Indicators, Paris: Author, October.

    Trends in large-scale, external testing outside the United States, Educational Measurement: Issues and Practice (EMIP), V.19, N.1, Spring 2000.

    Are U.S. students the most heavily tested on earth? Educational Measurement: Issues and Practice, V.15, N.3, Fall 1996

    U.S. Department of Education. National Center for Education Statistics. Education in States and Nations, (2nd ed.), NCES 96-160, by Richard P. Phelps, Thomas M. Smith, and Nabeel Alsalam.Washington, DC: 1996.

    U.S. Department of Education. National Center for Education Statistics. International Education Expenditure Comparability Study: Final Report, (Vol. 2). Working Paper #17, by Richard P. Phelps and Shelley Burns. Washington, DC: 1997.


    TESTING:

    Phelps, R. P. (Septemberl 25, 2024). The Great AP Score Recalibration James G Martin Center for Academic Renewal

    Phelps, R. P. (April 15, 2024). US Colleges and Admission Testing: Requiredd, Optional, or Blind? Minding the Campus.

    Phelps, R. P. (April 22, 2022). The US Test Mess James G Martin Center for Academic Renewal

    Phelps, R. P. (December 3, 2021). Are College Exit Exams a Valid Measure of Learning? It's Complicated James G Martin Center for Academic Renewal

    Phelps, R. P. (September 2020). Looking Back on DC Education Reform 10 Years After, Part 1: The Grand Tour Nonpartisan Education Review

    Phelps, R. P. (September 2020). Looking Back on DC Education Reform 10 Years After, Part 2: Test Cheats Nonpartisan Education Review

    Phelps, R. P. (2017). Designing an Assessment System, International Research-to-Practice Conference Nazarbayev Intellectual Schools, Astana, Kazakhstan, October, 2016.

    The "teaching to the test" family of fallacies, Revista Iberoamericana de Evaluacion Educativa, 10(1), (2017)

    Phelps, R. P. (2015). Presentation before the Governor's Council on Common Core Review, Little Rock, Arkansas, May 2015 [slide show]

    Phelps, R. P. (2015). "The Cork in the Bottle" The education establishment doesn't need to censor and suppress most of the evidence and information they do not like, the think tank elite does it for them. The think tank elite is the cork in the bottle that keeps the American public misinformed.

    Phelps, R. P. (2015). Kamenetz, A. (2015). The Test: Why our schools are obsessed with standardized testing, but you don't have to be. New York: Public Affairs. [book review]

    Phelps, R. P. (2014). Synergies for better learning: an international perspective on evaluation and assessment, by the OECD [report review by R.P. Phelps], in Assessment in Education: Principles, Policy & Practice, v.21, i.4, 2014

    Phelps, R. P. (2014). Evaluaciones educacionales de gran escala en Chile: Son necesarias? CIPER, February 6, 2014.

    Phelps, R. P. (2012, October 2). The school-test publishers complex Education News

    Phelps, R. P. (2012, September 19). The (secret) document that drives standardized testing Education News

    Phelps, R. P. (2013). The rot festers: Another National Research Council report on testing. New Educational Foundations, Summer, 2012.

    Phelps, R. P. (2011, Autumn). Teach to the Test? Wilson Quarterly

    Phelps, R. P. (2011). The Effect of Testing on Achievement: Meta-Analyses and Research Summary, 1910-2010: Source List, Effect Sizes, and References for Quantitative Studies. Nonpartisan Education Review/Resources.

    Phelps, R. P. (2011). The Effect of Testing on Achievement: Meta-Analyses and Research Summary, 1910-2010: Source List, Effect Sizes, and References for Survey Studies. Nonpartisan Education Review/Resources.

    Phelps, R.P. (2011). The Effect of Testing on Achievement: Meta-Analyses and Research Summary, 1910-2010: Source List, Effect Sizes, and References for Qualitative Studies. Nonpartisan Education Review/Resources.

    Phelps, R. P., Zenisky, A., Hambleton, R. K., & Sireci, S. G. (2010, March). Report to Ofqual: On the reporting of measurement uncertainty and reliability for U.S. educational and licensure tests

    The Rocky Score-Line of Lake Wobegon

    The source of Lake Wobegon [updated] #

    The source of Lake Wobegon [updated, pdf]

    The National Research Council's Testing Expertise

    Measuring Up: What Educational Testing Really Tells Us, by Daniel Koretz [book review]

    Characteristics of an effective testing system

    The source of Lake Wobegon #

    The source of Lake Wobegon [pdf] # See Bas Braams' excellent essay on this topic.

    Educational testing policy: Stuck between two political parties

    Standardized testing, regrettable food, urine therapy, and trichotillomania (Adults Only!)

    Organizations and Individuals to Contact for Information on Testing

    Economic Perspectives on Standardized Testing (powerpoint file)

    High Stakes: Testing for Tracking, Promotion, and Graduation, book review, Educational and Psychological Measurement, Vol.60, No.6, December 2000.

    Test-basher benefit-cost analysis," Network News & Views

    Phelps, Richard P. (1999) "Education Establishment Bias: A Look at the National Research Council's Critique of Test Utility Studies." The Industrial-Organizational Psychologist, 36(4). Society of Industrial-Organizational Psychologists, April.

    The Demand for Standardized Student Testing, Educational Measurement: Issues and Practice, V.17, N.3, Fall 1998.

    Trends in large-scale, external testing outside the United States, Educational Measurement: Issues and Practice (EMIP), V.19, N.1, Spring 2000.

    Are U.S. students the most heavily tested on earth? Educational Measurement: Issues and Practice, V.15, N.3, Fall 1996

    U.S. General Accounting Office, Student Testing: Current Extent and Expenditures, With Cost Estimates for a National Examination. PEMD-93-8 January 13, 1993


    EDUCATION FINANCE:

    Phelps, R. P. (2006, September 1). Thoroughly Inadequate: The 'School Funding Adequacy' Evasion Nevada Policy Research Institute.

    U.S. General Accounting Office, Student Testing: Current Extent and Expenditures, With Cost Estimates for a National Examination. PEMD-93-8 January 13, 1993

    "A Primer on Privatization," July 2001


    STATISTICAL/METHODOLOGY/EVALUATION:

    Phelps, R. P. (2013). The rot spreads worldwide: The OECD: Taken in and taking sides. New Educational Foundations, Winter, 2013.

    "book review: Raising the Grade: How High School Reform Can Save Our Youth and Our Nation by Bob Wise," Educational Horizons,

    "Dropping the ball on dropouts," Educational Horizons,

    "The dissolution of education knowledge," Educational Horizons,

    Standardized testing, regrettable food, urine therapy, and trichotillomania (Adults Only!)

    "A review of the Manhattan Institute's Graduation Rates in the United States," Practical, Assessment, Research, and Evaluation, v.10, n.15

    U.S. Department of Education. National Center for Education Statistics. Features of Occupational Programs at the Secondary and Postsecondary Education Levels. NCES 2001-018, by Richard P. Phelps, Basmat Parsad, Elizabeth Farris, and Lisa Hudson. Project Officer: Bernard Greene. Washington, DC: 2001.

    U.S. Department of Education. National Center for Education Statistics. State Indicators in Education 1997, NCES 97-376, by Richard P. Phelps, Andrew Cullen, Jack C. Easton, and Clayton M. Best. Project Officer, Claire Geddes. Washington, DC: 1997.

    Censorship has many fathers


    Not on the Web (INTERNATIONAL):

    "Benchmarking to the World's Best in Mathematics: Quality Control in Curriculum and Instruction Among the Top-Performing Countries on the TIMSS," Evaluation Review, August 2001

    Review of Survey Items on Continuing Education and Training, NCES-OECD, Feb. 1998

    "Student Achievement and Adult Literacy," chapter 6, Education at a Glance: OECD Indicators, Centre for Educational Research and Innovation, OECD, 1997

    Description and Status of U.S.-based International Benchmarking Efforts, NCES, 1997

    "International Comparisons of Public Expenditures on Education," in: The Condition of Education, 1998, _1997, _1996, _1995, _1994, NCES

    Survey of Countries' Practices Benchmarking Math Standards, NCES, 1996

    "Education in States and Nations" in Measuring What Students Learn, OECD, 1995

    "Education Finance Indicators: What Can We Learn From Comparing States With Nations?" in Developments in School Finance, NCES, January 1995.

    An Overview of the Korean Vocational Education System, PES, July 1994

    "American Public Opinion on Trade, 1950-1990," Business Economics, (v.28 n.3), July 1993

    "A Welcome Rain Falls in West Africa," The Christian Science Monitor, June 23, and on Monitor Radio, June 26-28, 1987.


    Not on the Web (TESTING):

    Phelps, R.P. (2012). The effect of testing on student achievement, 1910-2010. International Journal of Testing, 12(1), pp.21-43.

    "Estimating the Cost of Systemwide Student Testing in the United States," Journal of Education Finance (JEF), Winter 2000

    "Extent and Character of Systemwide Testing in the U.S." Educational Assessment, V.4, N.2, 1997

    "The Fractured Marketplace for Standardized Testing," book review, Economics of Education Review, (v.13 n.4) December, 1994

    "The Economics of Standardized Testing," delivered at the annual meeting of the American Education Finance Association, 1994

    "Benefit-Cost Analyses of Testing Programs," delivered at the annual meeting of the American Education Finance Association, 1994

    "National Testing, Pro and Con," The Education Digest, November, 1993.


    Not on the Web (EDUCATION FINANCE):

    National Assessment of Vocational Education: Funding and Accountability, PES

    "Estimating the Cost of Systemwide Student Testing in the United States," Journal of Education Finance (JEF), Winter 2000

    The Effect of University Host Community Size on State Growth," Economics of Education Review, v.17,n.2, 1998.

    Education System Benefits of U.S. Metric Conversion," Evaluation Review, Feb. 1996.

    "Education Finance Indicators: What Can We Learn From Comparing States With Nations?" in Developments in School Finance, NCES, January 1995.

    "The Fractured Marketplace for Standardized Testing," book review, Economics of Education Review, (v.13 n.4) December, 1994

    "The Economics of Standardized Testing," delivered at the annual meeting of the American Education Finance Association, 1994

    "Benefit-Cost Analyses of Testing Programs," delivered at the annual meeting of the American Education Finance Association, 1994

    "Education Finance in States and Nations," delivered at the annual meeting of the American Education Finance Association, 1994

    "Sensitivity Analysis of International Education Finance Indicators," delivered at the annual meeting of the American Education Finance Association, 1994


    Not on the Web (STATISTICAL/METHODOLOGY/EVALUATION):

    Longitudinal Data Analysis -- A Guide for NCES Data Users, Chapter 3, NCES

    Doctoral Scientists and Engineers in the United States, 1997, NSF, November 2000

    The Effect of University Host Community Size on State Growth," Economics of Education Review, v.17,n.2, 1998.

    Review of Survey Items on Continuing Education and Training, NCES-OECD, Feb. 1998

    Review of Proposed Indicators for Network B-Student Outcomes, NCES-OECD, July 1997

    "Student Achievement and Adult Literacy," chapter 6, Education at a Glance: OECD Indicators, Centre for Educational Research and Innovation, OECD, 1997

    "International Comparisons of Public Expenditures on Education," in: The Condition of Education, 1998, _1997, _1996, _1995, _1994, NCES

    Education System Benefits of U.S. Metric Conversion," Evaluation Review, Feb. 1996.