Andrew Hill
April 1, 2017
Running Head: A CRITICAL REFLECTION ON ASSESSMENT 2
This course has proven enlightening and instructive. Weve studied sundry aspects of
assessment (and its superset-relationship to testing) and how to consider or employ each in a
classroom context. Appreciating the distinction between assessment and testing something that
untrained teachers fail to recognize is of critical importance. I have learned to value this
distinction (as well as our many classroom experiences1). Herein I describe those aspects I found
most compelling. I do so in the context of prior classroom experiences, namely my stint teaching
Assessment and testing in contrast to what many a hagwon (Korea) or eikeiwa (Japan)
teacher may believe are not synonymous. Testing is, in fact, a subcategory of assessment. Per
Brown and Abeywickrama (2010), tests are prepared administrative procedures (Brown and
Abeywickrama, 2010, p. 3) that occur at given times and locations whereat test-takers must tap
into all they have learned. A given test a method must quantify a test-takers performance
according to explicit procedures or rules (p. 3). Texas public school students of multiple
generations are familiar with standardized tests: TAAS, TAKS, and STAAR, among others.
My Korean friends and coworkers diligent students all spent countless hours studying for the
level or magnitude of some attribute of a person (p. 3). A formal test, then, is clearly an
assessment. But so are more informal activities. If I am teaching irregular past tense verbs to my
students and I see one struggling, I might informally assess her: JiWon, what is the past tense of
1
I truly appreciated the group work assignments in this course. I had never worked in
conjunction with classmates to write a single cohesive paper and work to establish a unifying
voice. I often wondered how academics and researchers collaborated on a single academic paper.
These group work projects helped me to understand how to effectively do so.
Running Head: A CRITICAL REFLECTION ON ASSESSMENT 3
fly? If she answers flew then I may determine that she understands this concept. If she
answers flied then I will know I must adjust my instruction to help her course-correct. This is
an example of formative assessment. It informs and shapes my teaching. This stands in contrast
to summative assessment (typical traditional testing), which customarily occurs at the end of a
unit, semester, or course and assesses a sizable chunk of what students have learned. I employed
summative assessment in the South Korean classroom (expounded in this papers fifth
paragraph) and formative assessment (the aforementioned daily targeted questioning of students
that helped me gauge how well they understood my instruction). I also used formative
assessment while leading our own class in my quickshare (Powerwalk Dictation) and reading
we studied in class. For example, at the end of my reading presentation, I asked students to 1)
recall a type of listening assessment, 2) paraphrase that type in tweet form, and 3) tweet out
appropriate and designated hashtags2. By monitoring each students work, I could assess how
well they had understood my presentation. I could, theoretically, use this information to prepare
I truly enjoyed these student-led activities. It strikes me that they served several purposes:
1) They gave students an opportunity to practice leading a class and engaging in clear and
coherent communication, a necessary skill for any instructor. 2) They gave us an opportunity to
discover new and relevant classroom exercise; activities we might employ as future teachers of
ESL, EFL, and Bilingual Education. Digressing a bit, but only slightly, I use a great file-storing
application called Evernote3. It permits users to take notes within the application, to store files in
2
I loved this opportunity to practice digital/multimodal literacy activities.
3
I have often contemplated how I might use this program, interactively or educationally, with
my students.
Running Head: A CRITICAL REFLECTION ON ASSESSMENT 4
an easily accessible way, and to clip web pages. Each saved file can then be tagged, indexed, and
easily found again. I have a wonderfully complex system of filling and tagging materials from
every class so that I may find again once I begin teaching with my MA-TESL. I have saved
every quickshare (and key assessment-information) as I know I can use it at a later date.
Moving on: Early in this course, we considered the five foremost principles of language
expound on these cardinal criteria (p. 25), but for requisite brevity and substance I will focus
on two I found myself frequently reflecting on, asterisked above. Per Brown and Abeywickrama,
giving, and scoring an assessment instrument (p. 25). I know from personal experience that
myriad educational institutions particularly private academies (or hagwons) in South Korea
demand EFL teachers administer tests under acutely impractical conditions. For example: In
Korea, I proctored summative tests monthly to elementary students. These tests had five
components (Listening, Reading, Vocabulary, Grammar, and Writing). The instructors test
guidelines specified that students would require 100 minutes to complete this comprehensive
test. Unfortunately, test-takers had only 80 minutes (two class periods) to do so. In other words,
this test was a far cry from practical. The critical time crunch triggered student anxiety, raised
the affective filter, and thus hindered student-related reliability (one of four reliability4 forms).
These tests also lacked test-administration reliability. The classroom walls were paper-thin. As
each teacher played the listening sections audio, students and teachers alike (in each classroom)
could hear the loud and echoing audio through the walls. Students could not listen well enough
to answer correctly. Test reliability was thus negatively affected. I appreciated Pophams (2007)
4
Per Brown and Abeywickrama (2010), a reliable test is consistent and dependable (p. 27).
Running Head: A CRITICAL REFLECTION ON ASSESSMENT 5
practical advice to teachers: We need not devote any time calculating the reliability of our own
classroom tests and risk hypertension (Popham, 2007, p. 42); we should simply understand its
I wish I could travel back in time and evaluate each tests validity, what Gronlund (1998)
generally defines as the extent to which inferences made from assessment results are
appropriate, meaningful, and useful in terms of the purpose of the assessment [emphasis added]
(Brown and Abeywickrama, 2010, p. 29) and authenticity, what Bachmann and Palmer (1996)
generally define as the degree of correspondence of the characteristics of a given language test
task to the features of a target language task (p. 36). Finally, monthly tests had scant washback,
what Hughes (2003) refers to as the effect of testing on teaching and learning (p. 37). I had no
real opportunity to review and discuss test results with students (or incorporate relevant and
organized quality review sessions. We were, unfortunately, afforded no time for this. Our class
Huang and Juns (2015) compelling look at the influence of rater differences. Native English
Speaking (NES) and Non-Native English Speaking (NNES) teachers may evaluate their
students work differently, which has direct implications for the reliability, validity, and fairness
of assessments (Huang and Jun, 2015, p. 3). Per Kim (2009), NES raters paid special heed to a
select group of Korean students language use, pronunciation, and vocabulary whereas NNES
raters emphasized pronunciation, vocabulary, and intelligibility (p. 3). While working in Korea
I met many NES teachers (from countries like the US, Canada, Ireland, New Zealand, and India)
and many NNES teachers (from countries like Korea, Japan, Singapore, and Thailand). At that
Running Head: A CRITICAL REFLECTION ON ASSESSMENT 6
time, I gave little thought to how differences among these groups might unduly influence the
Finally, I wish to make note the virtue of alternative assessments, particularly the e-
portfolio. This is what Genesee and Upshur (1996) refer to as a purposeful collection of
students work that demonstrates their efforts, progress, and achievement in given areas
(Brown and Abeywickrama, 2010, p. 130). Done electronically, students engage in effective
excited to engage in this assignment at the end of this course and in the future. I want my
students, as they prepare for their own future final portfolio assignments, to recognize their
growth over the course of the academic semester or year. I wish that I had the opportunity to
experienced some dread when my advisor recommended I enroll. I believe this was due to my
own prior experience with assessments. High-school Andrew would have declared
a student who believed standardized tests were the metric by which I was deemed a good or
poor student. I want my own students (past and future) to suffer from little anxiety. I want(ed)
to assess them in a traditional or alternative fashion and make them feel valued as students.
This course has given me so much information and will help me to effectively do this.
Running Head: A CRITICAL REFLECTION ON ASSESSMENT 7
Works Cited
Brown, H.D., & Abeywickrama, P. (2010). Language assessment principles and classroom
nd
practices (2 ed.). White Plains, New York: Pearson.
Huang, B., & Jun, S. (2015). Age matters, and so may raters. Studies in Second Language
Acquisition,1-28.
Popham, .J. (2007). Classroom assessment: what teachers need to know. Boston: Allyn & Bacon