Introduction
Understanding language as a dynamic system, designing engaging and multi-leveled lesson plans, and utilizing technology to assist in language learning are essential for creating language classrooms that are enjoyable and meet the learning objectives. However, how does one measure that the learning objectives have been met? Language assessment provides the framework for assessing a variety of areas within language learning; assessments can have a variety of purposes (achievement, placement, proficiency, diagnostic, aptitude), characteristics (summative, formative, formal, informal, norm-referenced, criterion-referenced), and benefits and drawbacks related to areas of practicality, reliability, validity, authenticity, and washback. With all these elements in mind, creating the right assessment for your learners and targeted learning objectives can be a daunting task. Through the LTS program, I’ve developed my skills in designing and applying assessments. The following three artifacts demonstrate my competence in designing assessments for language learning. All three focus mainly on assessing achievement—to measure learners’ achievement of meeting a specific learning objective (Brown & Abeywickrama, 2019)—yet each shows the diverse characteristics an assessment with this purpose can have, as well as the consideration of benefits and drawbacks that are essential for designing high-quality language assessments.
LT 549 - Measuring Language Ability
Assessment Creation Activity: Writing
The first artifact is a sample writing assessment created in LT 549. This assessment, designed for intermediate-mid EFL learners at a Japanese university, assesses their presentational writing skills through a narrative essay which utilizes an extended production task type that is prompt-based (Carr, 2011) and has two sample prompts to choose from. Students first write a draft that receives formative feedback, afterwards submitting a final draft. This assessment creation activity focuses on the formal and summative assessment of the final draft. This drafting process improves the face validity of this assessment as well. Writing is a process that requires revision; therefore, the formative feedback they receive and use will increase their confidence that the second draft is truly measuring their writing abilities. This essay task assesses five student learning outcomes related to the genre of narrative writing (e.g., using the past tense to describe a personal memory) to certify that students are at the level expected by the end of this course—in turn allowing for their advancement to the next course (Bachman & Damböck, 2018). An analytic rubric was adapted from a version found in Winke and Lim (2015) for this assessment; these types of rubrics can assist with both intra- and inter-rater reliability, as they provide more consistency throughout the grading process (Barkaoui, 2011 as cited by Brown & Abeywickrama, 2019). As a language learner, I’ve experienced assessments of this type in my classes and found them to be highly useful and practical to my language development. I plan to utilize this task type to assess writing skills while constantly checking in with students to make sure that both the prompts and the rubric are valid, reliable, and aligned with the learning outcomes of the course.
LT 549 - Measuring Language Ability
Assessment Creation Activity: Reading
During LT 549, I had the chance to develop sample assessment tasks for other skills, such as the interpretive skill of reading. Designing production and receptive tasks require different considerations to be made with regard to learners and areas of reliability and validity. This reading assessment targets the same learners and context as the writing assessment and is also a formal and summative assessment. For this assessment, learners are being assessed on their achievement of interpretive reading; specifically, it addresses macro skills in reading related to identifying the main idea and specific details in an informational passage (Brown & Abeywickrama, 2019). Learners read four short passages and answer either five-to-seven multiple choice questions (a selected response task) or three short-answer questions (a limited response task) for each reading passage (Carr, 2011); for short-answer questions, learners are asked to respond in their L1 (Japanese) to maintain construct validity. In this assessment creation activity, the example passage also shows the careful consideration of multiple factors. When it comes to reading, language teachers must determine the difficulty of the passage—tools such as the English Vocab Profiler and Flesch-Kincaid readability test allow language teachers to view quantitative data about the reading passage they’ve chosen. However, the difficulty of words within the passage is not the only aspect to consider; learners might have more or less familiarity with a specific topic or genre, and teachers must consider whether this could give an unfair advantage or disadvantage to certain students. In this assessment, the passage was about a generally familiar topic (holidays) in a now-familiar genre of text (general informational text), which increases construct validity as well. Students had been learning about this topic and reading passages of this genre in class before the assessment; therefore, the assessment aligned with course content (Coombe et al., 2007). The assessment itself is highly practical as well. It’s taken within the class period, and relative to grading extended response tasks, multiple-choice and short-answer questions—with the assistance of an easy-to-use rubric—can be quickly graded. This assessment exhibits the multitude of considerations present in creating assessments, and my ability to take those considerations into account when designing carefully crafted assessments that align with the learning context and learning outcomes.
LING 594 - English Grammar
Grammar Lesson Plan: Compound Sentences
My final artifact, a lesson plan from LING 594, demonstrates how formative and informal assessment can be integrated throughout a lesson plan to provide learners with feedback that can drastically benefit their learning. This lesson was designed for an EAP course and focuses on creating compound sentences to increase the complexity of learners’ writing. Students receive multiple opportunities for both peer feedback and feedback from the teacher. For example, in the first activity of the lesson, groups of students create compound sentences using conjunctive adverbs, but later on in the lesson—after learning about punctuation—they go back to their original compound sentences and try to correct any incorrect punctuation they’ve noticed they originally had. The teacher can provide more feedback if necessary, but this allows students to self-assess their own work in groups. Another example is present later in the lesson when learners turn in paragraphs they wrote after interviewing fellow students; although this paragraph is not graded based on accuracy, it is still given formative feedback by the teacher so students can learn from it. This activity acts as an example of formal, but formative assessment with positive washback as it can guide the teacher to either focus more on compound sentences after this lesson or move on to the next topic (Hughes, 2003 as cited by Brown & Abeywickrama, 2019). Lastly, learners are given the opportunity to peer review past written work; before this lesson, students were asked to write and describe a favorite memory from their past. Now, in class, students swap their homework, and provide feedback to each other on the specific areas covered in class—this maintains construct validity, face validity, and is highly practical. Developing peer feedback skills also helps them develop thorough academic skills for future coursework. This lesson plan showcases how the integration of multiple opportunities for different kinds of formative assessment within a class lesson can benefit language learning.
Conclusion
These artifacts demonstrate the diversity of types of and opportunities for assessment that I can bring to my language classrooms. Whether it be formal or informal, summative or formative, I can develop assessments that align with learning outcomes and maintain practicality, reliability, validity, authenticity, and washback. Each of these artifacts also includes specific information on the design of the assessments themselves—the assessment creation activities even have some test specifications (Davidson & Lynch, 2002), which increase the clarity of the assessment and make it operational for designing future iterations of similar assessments. Assessment creation is by no means an easy task, but throughout my time in LTS I developed the foundational understanding required to design high-quality assessments, and I am ready to continue developing these as a part of my future as a language teacher.
Bachman, L. & Damböck, B. (2018). Language assessment for classroom teachers. Oxford University Press.
Brown, H. D. & Abeywickrama, P. (2019). Language assessment: Principles and classroom practices. Pearson.
Carr, N. T. (2011). Designing and analyzing language tests. Oxford University Press.
Coombe, C., Folse, K., & Hubley, N. (2007). A practical guide to assessing English language learners. University of Michigan Press.
Davidson, F., & Lynch, B. K. (2002). Testcraft: A teacher’s guide to writing and using language test specifications. Yale University Press.
Winke, P., & Lim, H. (2015). ESL essay raters’ cognitive processes in applying the Jacobs et al. rubric: An eye-movement study. Assessing Writing, 25, 38–54.