Dear all,
The death of Grant Wiggins early this week, coming on top of that of Paul Ginnis, makes one think how lucky we are to have enjoyed the influence of two great educational thinkers. There is no doubt that both these men have had a great impact on their sphere's of influence and upon us at ISHCMC.
There are a great number of tributes to Grant Wiggins on the internet that show his importance to the teaching profession. Although I regarded Paul as my educational mentor and more recently close friend, it was actually Grant Wiggins who ignited my desire to dig deep and to ask questions about education. I attended one of his weekend workshops on UbD at ISBangkok in the early 2000's and his insight, intelligence and practicality made me start to question how and why we do things in schools. Up to then, I'd just taught and managed by instinct on what felt right. The workshop on UbD changed all of that and I was inspired to learn more about children, how they learn and what we need to do to optimize that learning.That inspiration still persists today.
Here is an article I read years ago that made me think more about how and why we assess students. A topic that I am still grappling with today.
Have a good weekend,
Yours
Adrian
Grant Wiggins: Defining Assessment
Grant Wiggins is a nationally recognized assessment expert
who has been working in assessment reform for more than twenty-five years. He
is president of the educational consulting firm Authentic Education, and with Jay McTighe,
co-author of Understanding by Design, an award-winning framework
for curriculum design used around the world. In this interview, Wiggins shares
his thoughts on performance assessments, standardized tests, and more.
Wiggins has published several articles for Edutopia.org. In
2002, he wrote Toward Genuine Accountability: The Case for a New State
Assessment System. In 2006, he wrote Healthier Testing Made Easy: The Idea of Authentic
Assessment.
1. What
distinctions do you make between "testing" and
"assessment"?
Our line of argument is that testing is a small part of
assessment. It needs to be part of the picture. Many people who are
anti-testing end up sounding anti-evaluation and anti-measurement. A good test
has a role to play. The language that we like to use is, it's an audit. It's a
snapshot. You don't run your business for the audit. You want more than a
snapshot, you want a whole family album. But the audit and the snapshot have a
place in the larger picture.
What can the test do that more complex, performance-based,
project-based things can't do? Look for discrete knowledge and skill for the
individual student. Many projects, because they're so collaborative, end up
making you wonder, well, what about the individual student? What does the
individual student know?
For instance, in some state-based, performance-based
assessment, they always had a parallel paper-and-pencil test for the individual
student so that you had enough data on the individual. A different way to say
it -- and this is what scientists and researchers say -- is triangulate
the information. Match the quiz against the project, against the
PowerPoint® presentation. Now what's the whole picture say? So, what we would
say is "testing" is one piece of a portfolio.
2. What
is authentic assessment and why is it important?
Authentic assessment, to me, is not meant to be the charged
phrase, or jargony phrase that it has come to be for a lot of people. When we
first started using it fifteen years ago, we merely meant to signify authentic
work that big people actually do as opposed to fill-in-the-blanks,
paper-and-pencil, multiple-choice, short-answer quiz, school-based assessment.
So it's authentic in the sense [that] it's real. It's realistic. If you go into
the work place, they don't give you a multiple-choice test to see if you're
doing your job. They have some performance assessment, as they say in business.
Having said that, there is a misunderstanding. People say,
"Well, if it's not authentic, it can't possibly be a good
assessment." We never said that. We never implied it. There's a lot of
authentic work that doesn't make for good assessment because it's so messy and
squishy and it involves so many different people and so many variables that you
can't say with any certainty, "Well, what did that individual student know
about those particular objectives in this complex project that occurred over a
month?" So there's a place for unauthentic, non-real-world assessments.
We're just making the distinction that you shouldn't leave school not knowing
what big people actually do.
3. Why is
it important that teachers consider assessment before they begin planning
lessons or projects?
One of the challenges in teaching is designing, and to be a
good designer you have to think about what you're trying to accomplish and
craft a combination of the content and the instructional methods, but also the
assessment. And one of the things that we've done over the past years in
working with teachers is share with them how important it is to say, "What
are you going to assess? What's evidence of the goals that you have in
mind?" Otherwise your teaching can end up being hit-or-miss.
We call it backward design. Instead of jumping
to the activities -- '"Oh, I could have kids do this, oh, that'd be
cool" -- you say, "Well, wait a minute." Before you decide
exactly what you're going to do with them, if you achieve your objective, what
does it look like? What's the evidence that they got it? What's the evidence
that they can now do it, whatever the "it" is? So you have to think
about how it's going to end up, what it's going to look like. And then that
ripples back into your design, what activities will get you there. What
teaching moves will get you there?
4. How do
you assess project-based learning?
It all starts with, well, what are our goals? And how does
this project support those goals and how are we assessing in light of those goals?
So, you would expect to see for any project a scoring guideline, a rubric, in
which there are clear links to the project, to some criteria and standards that
we value that relate to some overarching objective -- quite explicitly, that
we're aiming for as teachers.
Sometimes we run into the problem that the project is so
much a creature of the student's interest that there's no question that lovely
learning occurs, but we sort of lose sight of the fact that now it's completely
out of our control. We don't even know what it's really accomplishing in terms
of our goals other than the kid is learning a lot and doing some critical and
creative work.
What we have to do is realize that even if we give this kid
free reign to do really cool projects, it's still got to fit within the context
of some objectives, standards, and criteria that we bring to it, and frame the
project in so that we can say by the end, "I have evidence. I can make the
case that you learned something substantial and significant that relates to
school objectives."
5. How
can technology support and enhance assessment?
Once we get beyond the idea that assessment is more than
just quizzes and tests -- and that it's the documentation of whereby you make
this case that the student has done something significant -- this body of
evidence, if we want to stick with that judicial metaphor, proves the student
actually learned something.
Technology is an obvious partner because whether it's on a
CD-ROM, floppies, or an old-fashioned technology like video cameras or even
overheads, the student is bringing together visual, three-dimensional, and
paper-and-pencil work. We want to be able to document and have a trace of what the
student has accomplished and how the student got there.
Having said that, I think sometimes technology is overused
and we don't think carefully enough about the evidence we need to give the
grade, put something on the transcript, and track that information over time.
Many well-intentioned people say, "Let's have student portfolios of the
student's work K-12." Well, that's fine for the student, but there's
hardly another human being other than the kid's family that wants to wade
through all that.
And that's actually another role of technology: It's a good
database system -- information management, storage, and retrieval whereby we
say, "I don't want to look through the whole portfolio. I want to just see
some samples, some rubrics to get a sense of the student's current level of
performance." Tracking information over time through technology is
actually an important part of it as well.
6. How do
you respond to the argument that teachers don't have enough time to design and
conduct authentic or performance-based assessments?
One of the criticisms often leveled at alternative forms of
assessment -- whether we call them performance, portfolio, authentic,
real-world, or project-based, -- is they're too time intensive, they're too
expensive. It's too big of a hassle. What's the payoff? What's the cost
benefit?
I can understand that argument at the state level. The state
is in the audit business. And one of the things I think we've learned over the
years is that given their need to save money, to not be too intrusive, to make
it reliable as an assessment, then they may have to not do some of this. But
many of those arguments that the critics make don't hold up at the district
level at all. On the contrary, it's not very expensive. You've got all your own
local people who are in the business of assessing. It's not inappropriate or a
waste of time because you can't meet the standards without doing performance-based
assessment.
7.
Standardized tests, such as the SAT, are used by schools as a predictor of a
student's future success. Is this a valid use of these tests?
Standardized testing has
a role to play as an audit, but one of the things that many policymakers and
parents forget, or don't know, is that these tests have a very narrow focus and
purpose as audits. They're just trying to find out if you really learned the
stuff you learned in school.
Whether these tests
predict future performance or success -- they do not. Even with the SAT, ETS
and the College Board are quite clear about what it does and does not predict.
It just predicts freshman grade point average in the first semester. That's
all. And there's plenty of studies to show that grades in college don't
correlate with later success.
So, one of the things
that people get in trouble with is assessment. It's like a bad game of
telephone. Remember the game you played as a kid? What starts out as a
perfectly intelligible sentence ends up being some wild distorted thing by the
end.
Ten or fifteen years
ago, the Secretary of Education was having wall charts about each state's SAT
performances -- as if that was a measure of school and school-system success.
But the SAT was invented as an aptitude test, not an achievement test linked to
curricula. It was just about general intelligence. Let's be very careful about
what we're making claims about, what these assessment results do and don't
mean. Most state and national tests are predicting very, very narrow results
about certain types of school performance. That's all.
No comments:
Post a Comment