Evolution of Language Learning Practices and Assessment
Before the emergence of AI in education, language learning assessment relied on long-established frameworks and one of the most influential was the Common European Framework of Reference for Languages (CEFR), which has been acknowledged in the Canadian education context. We thought it was a good entry point as it provides an an overarching goal of language assessment, with proficiency levels ranging from AI (beginner) to C2 (near-native). The historical language learning assessment practices can be examined through three different lenses: Classroom-based, tutor-based, and media based approaches.
For much of the 20th century, students were evaluated through grammar-translation practices: memorizing vocabulary lists, reciting grammar rules, and translating texts (Mao, 2022). Having been an English language learner myself in the early 2000s, I remember weekly vocabulary quizzes where long lists of words had to be memorized. Results were sometimes posted publicly, creating a tiered system that rewarded the “high scorers” while leaving others embarrassed. This approach emphasized accuracy and recall, but it often discouraged students who didn’t learn well through rote memorization.
By the mid-century, standardized testing became common, with multiple-choice tests measuring discrete points of grammar, vocabulary, and reading (Bachman, 1990). While these tests provided efficiency and comparability, they were not designed for diverse learners, and multiple-choice formats often reduced language to comprehension and context recognition rather than authentic use.
With the rise of Communicative Language Teaching, classrooms shifted toward communicative and performance-based assessments, such as skits, role plays, oral presentations, and situational tasks that aligned more closely with CEFR descriptors (Fulcher, 2000; Brown & Hudson, 1998). These assessments were more engaging and reflected real-world communication.
Private tutors played a major role in preparing learners for CEFR-aligned proficiency exams. Assessments in these settings were highly individualized, often focusing on targeted feedback for upcoming standardized tests such as IELTS or other proficiency exams. Tutors combined traditional drills with more communicative tasks, depending on the learner’s goals. The strength of this approach was the immediacy and personalization of feedback, which pre-AI computer-assisted tools could not replicate (Alderson, Clapham, & Wall, 1995). However, affordability and accessibility were significant barriers — not every learner had equal access to tutoring, which reinforced inequities in exam preparation. The culture also tended to be exam-driven, with instruction often shaped around test-taking strategies rather than holistic language growth. In this sense, tutoring offered valuable individualized support, but it also reflected the pressures and limitations of high-stakes assessment.
Before mobile apps and AI-driven personalization, media such as textbooks, CDs, audio programs, and language labs were key assessment tools. Learners completed exercises that mirrored CEFR descriptors — reading passages, listening comprehension, and writing prompts.
The advantage of these resources was their accessibility and growing interactivity: CDs and language labs, for example, provided exposure to authentic audio, and self-correcting exercises in textbooks encouraged independent practice. However, feedback was limited to answer keys or teacher interpretation, which meant learners often had little guidance on why they got something wrong.
The late 1990s introduced Computer-Assisted Language Learning (CALL), which enabled self-paced quizzes and adaptive testing, though these remained rule-based and restricted to item banks (Alderson et al., 1995). Media-based assessments made practice widely accessible but still lacked the nuance of human evaluation and the personalization that AI tools now promise.
References
Alderson, J. C., Clapham, C., & Wall, D. (1995). Language test construction and evaluation. Cambridge University Press.
Bachman, L. F. (1990). Fundamental considerations in language testing. Oxford University Press.
Brown, J. D., & Hudson, T. (1998). The alternatives in language assessment. TESOL Quarterly, 32(4), 653–675. https://doi.org/10.2307/3587999
Cambridge English. (n.d.). The Common European Framework of Reference (CEFR). Cambridge Assessment English. Retrieved September 29, 2025, from https://www.cambridgeenglish.org/exams-and-tests/cefr/
Council of Ministers of Education, Canada (CMEC). (n.d.). Common European Framework of Reference for Languages (CEFR). Retrieved September 29, 2025, from https://www.cmec.ca/136/Common_European_Framework_of_Reference_for_Languages_(CEFR).html
Fulcher, G. (2000). The ‘communicative’ legacy in language testing. Language Testing, 17(3), 251–268. https://doi.org/10.1177/026553220001700302
Mao, A. (2022). Literature review of language testing theories and approaches. Open Access Library Journal, 9, e8741. https://doi.org/10.4236/oalib.1108741