Global Security and Intelligence Studies Volume 4, Number 1, Spring/Summer 2019 | Page 19
Global Security and Intelligence Studies
So, the rise of assessment in the United States is driven by the confluence of
these factors. Colleges and universities have continued to seek ways to address the
complex task of measuring student knowledge levels. Government institutions,
responding to political pressures, sought mechanisms to ensure measures of accountability,
so that public funds were not being wasted. The rise of accreditation
bodies can be seen as a way of mediating these two issues (Shavelson 2010).
Critics of Assessment
The power of these accreditation bodies is not viewed by everyone as a positive
development. Some suggest that these accrediting bodies actually inhibit the assessment
of student learning. In his 1996 book, Crisis in the Academy, Christopher
J. Lucas criticized the accreditation system as too expensive, onerously complicated,
incestuous in its organization, and not properly tied to quality (Lucas 1996).
Similarly, a 2002 report by George C. Leef and Roxana D. Burris of the American
Council of Trustees and Alumni (ACTA) argued that the system does not ensure
or protect educational quality, while still imposing significant costs (Leef and Burris
2002). Robert Shireman, a senior fellow for the Century Foundation, recently
echoed this sentiment. He argued that accrediting agencies often require institutions
to reduce SLOs to meaningless blurbs which “prevents rather than leads to
the type of quality assurance that has student work at the center” (Shireman 2017).
Some suggest that this pressure leads to poor methodology in the attempt to
demonstrate the causal relationship between education and student learning. This
is not to suggest that the education does not facilitate student learning, only that
the data gathering and analysis structures that are in place are poorly constructed.
David Eubanks, a board member for the Association for the Assessment of Learning
in Higher Education, criticized the approach to assessment structures, noting
“the whole assessment process would fall apart if we had to test it for reliability and
validity and carefully model interactions before making conclusions about cause
and effect” (Eubanks 2017, 6). He argues that applying common sense to dubious
data is akin to a Rorschach test.
Another critique of the current accreditation-driven assessment structure
is that it has little impact on educational improvement. As Banta and Blaich noted
in a study of assessment reports: “We scoured current literature, consulted experienced
colleagues, and reviewed our own experiences, but we could identify only
a handful of examples of the use of assessment findings in stimulating improvements”
(Banta and Blaich 2011, 22). Indeed, some suggest that the 6 percent of improvements
that were found in this study actually overestimate the impact because
the assessment reports were not selected randomly (Fulcher et al. 2014).
Indeed, among these scholars, the current model is self-sustaining due to
financial incentives. As Upton Sinclair once noted, “it is difficult to get a man to
understand something when his salary depends upon his not understanding.”
8