History of Standardized Testing
History of Standardized Testing
21
Chapter 1
22
Accountability and Standardized Testing
23
Chapter 1
out of what many saw as the province of the states, Congress demoted it to an
Office of Education in 1868. The Office of Education spent time being shuffled
between the Department of the Interior and the Federal Security Agency before
settling in the Department of Health, Education, and Welfare. It was eventually
given cabinet-level authority as the Department of Education in 1980 (U.S.
Department of Education, The Federal Role in Education).
The main purpose of establishing the Department of Education in 1867 as
described in the Act was to have an agency that gathered information on the
condition and progress of our educational system:
Be it enacted by the Senate and House of Representatives of the
United States of America in Congress assembled, That there shall
be established, at the city of Washington, a department of ed-
ucation, for the purpose of collecting such statistics and facts
as shall show the condition and progress of education in the
several States and Territories, and of diffusing such informa-
tion respecting the organization and management of schools
and school systems, and methods of teaching, as shall aid the
people of the United States in the establishment and main-
tenance of efficient school systems, and otherwise promote
the cause of education throughout the country. (An Act to
establish a Department of Education, 1867)
Upon establishment of this department, a number of people began to ad-
vocate for the implementation of a national standardized exam. But a strong
adherence to states’ rights and logistical barriers to the implementation of a na-
tional exam kept such efforts at bay. In fact, it would be just over one hundred
years after establishing a Federal Department of Education, in 1969, that the first
national exam, the NAEP, was administered. Nonetheless, the mid-1800s and
early 1900s marked a rapid expansion and development of educational testing
and measurement in the United States—much of it through the efforts of our
universities to influence curriculum at the secondary level in order to ensure stu-
dents were prepared for university-level work. While these early efforts were not
referred to in terms of accountability, but rather as a process of articulation, this
process helped lay the groundwork for the systems of accountability at play today.
As early as 1833 Harvard and other colleges began to administer written
exams as proof of achievement—the first in math (Black 192). By 1851, Har-
vard faculty recognized they could no longer assume students would arrive with
a uniform set of skills, and in response instituted one of the first standardized,
written entrance exams, focusing primarily on Latin grammar and math (Han-
son 193) and, by the mid-1860s, including Greek composition, history, and ge-
24
Accountability and Standardized Testing
ography. During this same time period, the number of children in government-
funded schools began to swell, and public schools began to follow the example
set by colleges in terms of measuring achievement. With increasing demand
from universities for these schools to produce college-ready students, as well as
the organization of boards of education in the states, standardized testing began
to find solid footing in the United States.
The written standardized exam administered to all Boston school children
in 1845 is thought to be the first large-scale achievement test of its kind, and a
full account of this exam, including test questions, sample responses, and results,
was collected for the 1925 edition of Then and Now in Education, 1845:1923
(Caldwell and Courtis). Prior to 1845, the Boston public schools followed the
standard practice of requiring oral exams administered by a traveling panel of
examiners. But by 1845 there were 7,000 students in nineteen different schools,
and this approach to measurement was no longer feasible. Instead, Boston in-
stituted a written exam thought to be more objective, reliable, and economical
than the oral exams (Mathison 3). The language sections of these tests focused
on definitions and prescriptive grammar. Early examiners describe the condition
and progress of education in Boston schools at this time in their report:
The first feeling occasioned by looking over these returns is
that of entire incredulity. It is very difficult to believe that,
in the Boston Schools, there should be so many children in
the first classes, unable to answer such questions; that there
should be so many who try to answer, and answer imper-
fectly; that there should be so many absurd answers, so many
errors in spelling, in grammar, and in punctuation. If by any
accident these documents should be destroyed, we could
hardly hope that your faith in our accuracy would induce you
to believe the truth if we told it. But the papers are all before
you, each signed by the scholar who wrote it. . . . The most
striking results are shown in the attempts to give definitions
to words. There were twenty-eight words selected from the
reading book, which the classes have probably read through
during the year, and some probably more than once. Some of
these words are the very titles or headings of reading lessons;
some of them occur several times in the book, and yet, of the
516 children who had these questions before them, one hour,
not a single one defined correctly every word; only 47 defined
half of them; and 29 could not define correctly a single one of
the whole 28 words. (Then and Now 171, 175)
25
Chapter 1
While most of these very early tests did not resemble those with which we
are familiar today, it was not long before the basic structure of standardized tests
of written communication were in place—a structure to which we still largely
adhere. Standardization of writing tests took a significant leap forward in 1860
with the introduction of scaled tests of writing achievement. George Fisher, an
English schoolteacher, provided us with the first written account of educators
using anchor papers on a scale of 1–5 designed to measure writing achievement
of large numbers of students. Fisher used these tests to assess handwriting, spell-
ing, grammar, and composition (Bryant and Bryant 420). While it is not clear
if the standard scale books themselves still exist, Fisher’s description of them can
be found in a copy of a paper he presented to the Statistical Section F, British
Association, Cambridge, October 1, 1862:
On the Numerical Mode of Estimating and Recording
Educational Qualifications As Pursued in the Greenwich
Hospital Schools
It has been observed that “no mode of teaching can be prop-
erly appreciated so long as we are without recognized princi-
ples of examination, and accuracy in recording the results; for
without such means neither failures nor improvements will
add to our common stock of experience in such matters; and
we hand down to posterity no statistical information of such
value as will mark the progress of Education. . . .
Such a plan of numerical estimation has been carried out
in the Greenwich Hospital Schools. A book, called the “Stan-
dard Scale-Book,” has been there kept since the first general
introduction of the plan containing the numerical value of
each degree of proficiency in the various subjects of examina-
tion. If it be required, for instance, to determine the numerical
equivalent to any specimen of writing, a comparison is made
with various standard specimens of writing contained in this
book, which are arrayed and numerically valued according to
the degree of merit. The best executed being represented by the
number 1, and the worst by the number 5. . . . So long as such
standard specimens are preserved in the School, constant and
permanent values for proficiency in writing can be maintained;
and since facsimiles can now be multiplied with very little
expense, it appears obvious that the same principle might be
generally adopted, provided well-considered standards were
agreed upon and recognized. . . .
26
Accountability and Standardized Testing
27
Chapter 1
to assessing writing, we are still using a system originally developed in the 1860s
in England and then later refined in the early 1900s in the United States.
The Hillegas-Thorndike Scale, and the goals of Hillegas and Thorndike them-
selves, were widely debated in composition teaching and research publications
of the time, including numerous references from 1912–1925 in NCTE’s English
Journal. While many found the scale useful in very controlled contexts, most
found it impractical due to the variation among genres, styles, grade levels, and
other matters familiar to us today. As one critic pointed out, “You can not measure
light, and warmth, and redness on the same rod” (Thomas 3) and, similarly, you
can not measure all student writing achievement using the same rod (Thomas 3).
Even in the twenty-first century, with technology unimagined in the early twen-
tieth century, we are still using the same rod to measure student writing achieve-
ment. Rather than use technology to bring wide-scale innovation to this process,
we have been content to focus on bringing economy of scale to the process.
One other major development requires mention in our brief history. In
1900, the College Entrance and Examination Board (CEEB) was established
by a group of private high schools and elite colleges in order to standardize the
admissions process and drive a more uniform curriculum at the private New
England high schools from which the colleges drew most of their students. The
CEEB later became College Board, a nonprofit testing agency most of us are
familiar with as the administrator of the SAT. By the mid-1950s College Board
was administering the Advanced Placement Program and soon developed the
PSAT to measure students’ critical reading and math skills in preparation for
college entrance exams like the SAT and ACT. In 1959 ACT was formed as an
alternative testing option to the SAT. Both of these organizations have grown
immensely over the years, reaching ever farther into the educational landscape.
28
Accountability and Standardized Testing
Many have and continue to use this event as proof of declining educational
standards, particularly in math and science, making ample room for the argu-
ment that education is a matter of national security and the common good, and
thus requires federal intervention. But Sputnik may be the most successful and
persistent manufactured myth about the state of America’s educational system
to date. The crisis generated by this manufactured myth allowed for the political
capital needed to pass the National Defense of Education Act4 in 1958, opening
the door to a national test of achievement—the NAEP—a giant leap toward the
accountability movement that is now in full swing. Furthermore, this is a crisis
that has remained a persuasive touchstone for educational reform movements
for almost sixty years.
For example, Christopher Tienken and Donald Orlich remind us:
President Bill Clinton’s Secretary of Education Richard Riley
(1995) used Sputnik to justify further federal involvement in
education as part of the America 2000 legislation: “When the
Russians woke us up by flying Sputnik over our heads late at
night—a few of you may remember that experience—Con-
gress passed the 1958 National Defense of Education Act,
which sent millions of Americans to college and educated a
generation of scientists who helped us to win the Cold War.”
Ronald Reagan used Sputnik as a propaganda tool in 1982 to
support his plan to give tax credits for parents to send their
students to private schools. (25)
And in his 2011 State of the Union Address, President Obama declared:
“This is our generation’s Sputnik moment. . . . But if we want to win the fu-
29
Chapter 1
30
Accountability and Standardized Testing
But it wasn’t just the perceived loss of the space race that finally led to a na-
tional exam. Equality of Educational Opportunity for All, often referred to as the
Coleman Report, may have had an equally important effect. The Coleman Re-
port was commissioned by the U.S. Office of Education and published in 1966.
It “marks the first time there is made available a comprehensive collection of data
gathered on consistent specifications throughout the whole nation” (Coleman
1). Approximately 645,000 students from 4,000 public schools in grades 3, 6,
9, and 12 participated in this research, which focused on the extent to which
equality of education was a reality for America’s school children.
This was a landmark study leading to a flurry of activity but, as many ar-
gue, little in the way of educational progress. As a brief aside, we can link The
Colman Report with claims such as those of Berliner that the real education
crisis is a crisis of poverty, not a crisis of overall educational achievement. In a
retrospective on The Coleman Report, Adam Gorman and Daniel Long of the
Wisconsin Center for Education Research found that forty years later the major
findings of the report hold up well, most notably that per-pupil spending is less
important than level of teacher training, the black-white achievement gap per-
sists, and “Student achievement still varies substantially within schools . . . and
31
Chapter 1
this variation is still tied to students’ social and economic backgrounds” (19). In
fact, when discussing the 2015 reauthorization of the Elementary and Second-
ary Education Act, Secretary of Education Arne Duncan prioritized equity for
low-income and minority students because “Education Department data show
that 6.6 million students from low-income families are being shortchanged
when it comes to state and local education funding” (U.S. Department of Edu-
cation, “Secretary Duncan”). For example, the education department estimates
that in Pennsylvania, the highest-poverty districts spend 33 percent less than
the lowest-poverty districts, while in Vermont, Illinois, Missouri, and Virginia,
the highest-poverty districts spend 18–17 percent less than the lowest-poverty
districts. And in Nevada, the highest-minority districts spend 30 percent less
than the lowest-minority districts, while in Nebraska and Arizona, the highest-
minority districts spend 17–15 percent less than the lowest-minority districts
(U.S. Department of Education, “Secretary Duncan”).
Importantly, it is difficult to attract, retain, and develop high-quality teachers
in high-poverty schools (Clotfelter, Ladd, and Vigdor 2005; Grissom 2011). A
2014 report by the Alliance for Excellent Education estimates that 13 percent of
our teachers move or leave the teaching profession each year: “This high turnover
rate disproportionately affects high-poverty schools and seriously compromises
the nation’s capacity to ensure that all students have access to skilled teaching”
(Haynes). This is especially problematic when we consider that, as Ben Ost says,
“one of the most consistent findings in the literature on teacher quality is that
teachers improve with experience” (1).
Most studies of teacher turnover in high-poverty schools have attributed
turnover to characteristics of the students and the teachers, rather than the or-
ganizational structure of the schools themselves—organizational structures that
can be improved with increased funding. Emerging research on teacher turn-
over in high-poverty schools suggests “when these teachers leave, it is frequently
because the working conditions in their schools impede their chance to teach
and their students’ chance to learn” (Simon and Johnson 4). Organizational
factors that are associated with higher rates of turnover include administrative
support, teacher input in decision-making, salary, and aspects of school culture
(Simon and Johnson 12). We will return to a discussion of some of these factors
in our last chapter, but for the moment let’s turn back to our history of stan-
dardized testing.
The ability tests collected as part of The Coleman Report were administered
by ETS and the language section focused on items such as sentence completion
and identifying analogies—items that could easily and efficiently be measured.
This is not surprising given the number of students involved in this study and
research appearing as early as the 1940s claiming a high correlation between
32
Accountability and Standardized Testing
33
Chapter 1
34
Accountability and Standardized Testing
that a test has yet to be developed that can reliably measure change in writing
achievement over time due to the rapidly changing writing demands placed on
students and workers.
Despite the misgivings of the National Assessment Governing Board itself,
and constant revision and critique of attempts to assess writing over time, many
were not deterred by these concerns and instead began to argue for the use of
such tests within higher education.
35
Chapter 1
36
Accountability and Standardized Testing
Good as its predecessors, the National Defense of Education Act and A Nation
at Risk, the U.S. Department of Education’s A Test of Leadership: Charting the
Future of U.S. Higher Education urges a “robust culture of accountability” (20):
Interestingly, and very much in line with the rhetoric and practice of No
Child Left Behind, the authors of this report note in their introductory sum-
mary that, “According to the most recent National Assessment of Adult Literacy
. . . the percentage of college graduates deemed proficient in prose literacy has
actually declined from 40 to 31 percent in the past decade” (3). And yet, in its
recommendations, the commission “urge[s] these institutions to develop new
pedagogies, curricula and technologies to improve learning, particularly in the
areas of science and mathematics” (5, emphasis ours), choosing not to place an
emphasis on writing in U.S. schools.
A Test of Leadership names specific standardized tests, such as the Collegiate
Learning Assessment (CLA), for use in our colleges as a means of rigorous ac-
countability. The CLA was developed under the auspices of the Council for Aid
to Education (CAE), a nonprofit organization initially established in 1952 to
encourage corporate support of education. The CAE currently conducts policy
research on higher education as well as focuses on improving quality and access
in higher education, primarily through the CLA, and now CLA+ (a revision of
CLA). CAE describes CLA+ as a way for national and international institutions
to “benchmark value-added growth in student learning at their college or in-
stitution compared to other institutions.” CAE uses “performance-based tasks
. . . to evaluate the critical-thinking and written-communication skills of college
students. It measures analysis and problem solving, scientific and quantitative
reasoning, critical reading and evaluation, and critiquing argument, in addition
37
Chapter 1
38
Accountability and Standardized Testing
improve his or her reasoning and writing skills” (4). In fact, as Richard Haswell
makes clear, “every one of their twenty-seven subgroups recorded gain” (488),
but the authors of Academically Adrift claim that this gain was “modest” or “lim-
ited” based on their set standard of statistical significance. Equally concerning,
as Haswell explains, “Not one piece of past research showing undergraduate im-
provement in writing and critical thinking—and there are hundreds—appears
in the authors’ discussion or their bibliography, although both are a swim with
think-tank books and blue-ribbon papers opining the opposite” (488).
Examined from another angle, Lane and Oswald make the case that:
This 45% finding is, indeed, shocking—but for a completely
different reason. Considering that each significance test was
based on a sample size of 1 (i.e., each student’s change in
the CLA measure), it is hard to imagine that as many as 55
percent of students would show statistically significant gains.
Indeed, one would expect to find an order of magnitude
fewer significant improvements, based on the mean difference
between the pre- and post-tests the authors reported in their
study. The reason Arum and Roska found that so many (not
so few) students improved significantly is that they computed
the wrong significance test.
This particular problem is further highlighted in a paper published by the
CLA itself titled The Collegiate Learning Assessment: Facts and Fantasies, in which
they make clear that “The CLA focuses on the institution (rather than the stu-
dent) as the unit of analysis . . . [and] The CLA itself does not identify the rea-
sons why a school’s students do better or worse than expected” (Klein, et al. 3).
But for those of us not statistically inclined, there are other glaring prob-
lems with claims that this standardized test of writing can be used to measure
change in student ability over time. In fact, these problems seem to echo the
very same ones that caused the National Assessment of Educational Progress to
question the validity and reliability of their long-term trend assessments in writ-
ing and, ultimately, to declare them not reliable or valid enough upon which to
make claims about change in writing achievement over time. The first problem
is whether or not the writing tasks and the measurement tools used at two dif-
ferent intervals were controlled to a level that would allow for valid and reliable
comparison of change over time. It is important to emphasize that they seem to
echo these problems because the authors of Academically Adrift will not release
the actual pre- and post-writing prompts used in their research so that those who
specialize in writing assessment and test development can measure the validity
and reliability of their claims. This unwillingness to engage in full peer review,
39
Chapter 1
especially to a degree that would allow others to determine the validity and re-
liability of their results through means such as replicability, certainly calls their
research and motives into question.
The second problem concerns the writing tasks themselves. CLA and the
authors of Academically Adrift emphasize numerous times that their perfor-
mance-based assessments of writing are authentic and based on general skills as
opposed to specific content knowledge gained through exposure to the primary
texts in one’s major or discipline. They point to the following performance-based
assessment as representative of a task requiring only general skills:
The “DynaTech” performance task asks students to generate a
memo advising an employer about the desirability of purchas-
ing a type of airplane that has recently crashed. Students are
informed: “You are the assistant to Pat Williams, the president
of DynaTech, a company that makes precision electronic
instruments and navigational equipment. Sally Evans, a mem-
ber of DynaTech’s sales force, recommended that DynaTech
buy a small private plane (a SwiftAir 235) that she and other
members of the sales force could use to visit customers. Pat
was about to approve the purchase when there was an acci-
dent involving a SwiftAir 235.” Students are provided with
the following set of documents for this activity: newspaper
articles about the accident, a federal accident report on in-
flight breakups in single engine planes, Pat Williams’ e-mail
to her assistant and Sally Evans’ e-mail to Pat Williams, charts
on SwiftAir’s performance characteristics, an article from Am-
ateur Pilot magazine comparing the SwiftAir 235 to similar
planes, and pictures and descriptions of SwiftAir models 180
and 235. Students are then instructed to “prepare a memo
that addresses several questions, including what data support
or refute the claim that the type of wing on the SwiftAir 235
leads to more in-flight breakups, what other factors might
have contributed to the accident and should be taken in
account, and your overall recommendation about whether
or not DynaTech should purchase the plane. (Academically
Adrift, 21–22)
Of course, there is the obvious problem of the timed nature of this task, as
no one of any repute would tackle such a serious writing task in ninety min-
utes. Perhaps more perplexing is that it is difficult at best to understand how
a prompt requiring knowledge of a discipline-specific genre, a formal business
40
Accountability and Standardized Testing
41
Chapter 1
its attendant calls for systems of accountability, the CCSS are being propelled
by a fear that the United States is falling dangerously behind other countries in
global tests of academic achievement. As the October 7, 2013, issue of Time pro-
claimed: “What’s driving the core standards conversation now is the ambition to
succeed in a global economy and the anxiety that American students are failing
to do so” (Meacham 44). This crisis rhetoric can be found in the Council on
Foreign Relations Task Force’s report US Education Reform and National Security
that argues a failing U.S. education system threatens our national security in five
specific ways: “threats to economic growth and competitiveness, U.S. physical
safety, intellectual property, U.S. global awareness, and U.S. unity and cohe-
sion” (qtd. in Klein and Rice 7). Further, while critiques of the CCSS abound,
overall their adoption has been swift and ongoing as textbooks are realigned,
tests developed, school district rubrics restructured, and teachers trained. In fact,
as mentioned in our introduction, when a small number of governors began
to publicly denounce CCSS after previously adopting the standards, the group
Higher Ed for Higher Standards was formed and includes over 200 presidents,
chancellors, state officials, and organizations such as the American Association
of Colleges and Universities (AAC&U). Much like Harvard in the 1800s, this
group is working to establish processes of articulation, this time via CCSS. Per-
haps not surprisingly, this coalition is part of the Collaborative for Student Suc-
cess, funded in large part by the Bill and Melinda Gates Foundation (Mangan),
the primary investor in the CCSS itself.
The conflicts of interest in terms of how the CCSS are being funded and
implemented forebode systems of accountability and measurement that will rest
heavily on writing instruction at the college level. Thomas Newkirk begins to
unravel these conflicts in “Speaking Back to the Common Core”:
The Common Core State Standards are joined at the hip to
standardized tests, not surprising because both the College
Board and the ACT have had such a big role in their creation.
It was clear from their conception that they would play a large
part in teaching evaluation, a requirement for applications for
Race to the Top funds and exemptions from No Child Left
Behind. (4)
For example, David Coleman, who became the president of College Board
in 2012, and thus overseer of the SAT, is not only one of the major initiators
of the CCSS, but one of the people who convinced Bill and Melinda Gates to
fund them. Bill Gates did more than simply fund their development; he “was
de facto organizer, providing money and structure for states to work together
on common standards in a way that avoided the usual collision between states’
42
Accountability and Standardized Testing
rights and national interests that had undercut every previous effort” (Layton).
Coleman went on to write much of the standards for math and literacy. Most
recently, in many well-publicized events, he announced that the SAT will be re-
designed to align with the CCSS. One of the changes includes making the essay
part of the exam optional. The entanglements don’t end here. As reported in the
November 3, 2013, issue of the Chronicle of Higher Education, the Bill and Me-
linda Gates Foundation hired Richard Arum, one of the authors of Academically
Adrift, as a senior fellow on educational quality.
The influence of private foundations reaches far beyond investment in the
development of the standards. For example, the National Writing Project is now
significantly funded in part by the Bill and Melinda Gates Foundation, and
this funding reaches down into local sites specifically in an increased effort to
gain compliance with the CCSS. In 2010 The National Writing Project received
a $550,000 grant from the Bill & Melinda Gates Foundation and teams of
teachers were expected to “create a model for classroom teachers in writing in-
struction across the curriculum that will support students to achieve the out-
comes of the Common Core Standards” (“To Create”). In 2011 the Bill and
Melinda Gates Foundation awarded $3,095,593 in grant money to local sites of
the National Writing Project to “create curricula models for classroom teachers
in writing instruction that will support students to achieve the outcomes of the
newly state-adopted Common Core Standards” (“Denver Writing Project”). In
2014, the Bill and Melinda Gates Foundation funded the Assignments Matter
program. These grants were designed to “introduce large numbers of teachers to
the Literacy Design Collaborative (LDC) and its tools for making and sharing
writing assignments. Specifically, we will introduce teachers to the LDC task
bank and jurying rubric, tools meant to support teachers in creating clear and
meaningful writing prompts” (“Assignments Matter”).
While the official website for the Common Core State Standards empha-
sizes the flexibility teachers have in developing curriculum, the Literacy Design
Collaborative belies what may appear to be support for teacher agency. In 2013,
the Bill and Melinda Gates Foundation directed $12,000,000 to “incubate an
anchor Literacy Design Collaborative (LDC) organization to further expand
reach and impact [of the Common Core State Standards]” (Literacy Design
Collaborative, Inc.). On their official website, the LDC purports to put “educa-
tors in the lead” but only in so much as they operate within the relatively narrow
parameters of rubrics designed and approved by the Collaborative. For example:
[LDC] has created a process to validate the CCRS align-
ment of LDC-created content. The SCALE-created “jurying”
process looks at how richly the tasks and modules engage
43
Chapter 1
44
Accountability and Standardized Testing
WestEd affirms research evidencing that standardized tests alone are not the best
means for determining college admissions and placement (Bracco et. al). This is
important given the research we previously detailed on the use of standardized
tests for this purpose. The report discusses the range of measures in Core to Col-
lege states that are being considered for college placement. Perhaps all we can
take away from the WestEd studies of Core to College is that the effectiveness of
Common Core State Standards in creating greater alignment and collaboration
among K–12 and higher education is quite mixed. The mixed results of the Core
to College initiative make it difficult to determine ongoing effects of this type
of work. The Core to College initiative formally ended in 2014, although some
states are certainly continuing this work and it will be important to see if it will
lead to lasting and impactful K–12 and college collaborations. While we might
be optimistic about the rich opportunities K–12 and college collaborations can
yield, given how these efforts are being funded and how often they are used to
establish ever greater systems of accountability and control over our K–12 class-
rooms, we must be cautious and critical optimists as we move forward.
All of this raises questions about who is driving U.S. higher education
these days. Of course, higher education in the United States has always been
shaped by multiple competing forces. For example, beginning in 1938 with
Earnest Hollis’ book Philanthropy Foundations and Higher Education, many
researchers have documented the influence that private foundations have had
on reforming higher education. In a study published in 2011 by Cassie Hall—
using a review of academic literature, an analysis of public discourse from a
wide variety of media, ten years of secondary data on philanthropic giving
to higher education, and interviews with five senior-level professionals—Hall
shows that there has been a fundamental shift in the relationship between
foundations, higher education, and the control of public policy. Historically,
foundations shaped higher education primarily through direct incentives to
institutions with a focus on capital construction, academic research or pro-
grammatic efforts (Hall 16). But as Hall demonstrates in her analysis of the
changing relationship between foundations and higher education, “recent
foundation behavior suggests that a new approach to higher education philan-
thropy has emerged over the past decade, one that emphasizes broad-scale
reform initiatives and systemic change through focused, hands-on public pol-
icy work” (2). This new approach to foundation work is being referred to as
“advocacy philanthropy.” Hall argues that foundations’ “overt focus on public
policy advocacy within specific state and local contexts will have a significant
impact on higher education in the United States” (50).
As a conclusion to her study, Hall discusses the possible benefits, concerns,
and emerging outcomes of this shift. Potential benefits of advocacy philanthropy
45
Chapter 1
46
Accountability and Standardized Testing
NOTES
1. As evidence of this increase we can look to the Collegiate Learning Assessment
(CLA), which has grown the number of participating higher education institutions
to 700 from its inception in 2002. Further, CAE (Council for Aid to Education),
the organization that administers the CLA, is working with those developing Com-
mon Core State Standards Assessments to ensure alignment between their stan-
dardized tests and those used at the college level such as the CLA (Council for Aid
to Education).
2. Throughout this book we focus on PARCC, but there is another consortium that
has also developed CCSS aligned standardized tests—The Smarter Balanced As-
sessment Consortium. Because we don’t intend this book to focus primarily on an
analysis of these consortia, we chose to focus on PARCC as just one example of
the current state of standardized testing in relationship to high school and college
curricula both because it is the more controversial of the two consortia and because
we both happen to live in PARCC member states.
3. Many historians agree that the first standardized tests to include writing were ad-
ministered in China as early as 1115 A.D. These were known as “Imperial Exam-
inations” and covered the Six Arts: music, math, writing, knowledge of the ritu-
als of public and private life, archery, and horsemanship (Ward 44). The Imperial
Examination was essentially a civil service exam that was open to nearly all males
“and became the most important avenue to position, power, and prestige in China”
(Hanson, 186).
4. For more on the role of the National Defense of Education Act on the shape of
rhetoric and composition as a field, see Margaret Strain’s “In Defense of a Nation:
The National Defense Education Act, Project English, and the Origins of Empirical
Research in Composition.”
5. For more information, see http://www.learningoutcomesassessment.org/AboutUs
.html.
6. For more information, see http://digitalcommons.unl.edu/peerreviewteaching/.
7. For a fuller discussion of the CCSS, see http://www.corestandards.org/about-the
-standards.
47