Hostname: page-component-586b7cd67f-g8jcs Total loading time: 0 Render date: 2024-11-24T15:48:27.206Z Has data issue: false hasContentIssue false

Issues in post-entry language assessment in English-medium universities

Published online by Cambridge University Press:  13 June 2013

John Read*
Affiliation:
University of [email protected]

Abstract

Students entering English-medium universities around the world come from diverse linguistic backgrounds and many find it difficult to cope with the language demands of their degree courses. This speech focuses first on the motivations for introducing what are known in Australia as post-entry language assessments (PELAs). I discuss in particular the Diagnostic English Language Needs Assessment (DELNA) at the University of Auckland, which has been successfully implemented throughout the university in the last ten years. This leads to a broader discussion of three significant issues in post-entry language assessment. One is the nature of the construct on which such assessments are based. A related issue is whether a PELA can appropriately be considered ‘diagnostic’ in nature. The third issue is how to undertake the validation of a PELA, particularly with regard to the consequential aspect: does the assessment contribute to enhancing the students’ academic language ability?

Type
Plenary Speeches
Copyright
Copyright © Cambridge University Press 2013 

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

References

AEI (Australian Education International) (2007). Final report: Outcomes from a national symposium: English language competence of international students. https://aei.gov.au/research/Publications/Documents/NS_Outcomes_Syposium.pdfGoogle Scholar
Alderson, J. C. (2005). Diagnosing foreign language proficiency. London: Continuum.Google Scholar
Arkoudis, S., Baik, C. & Richardson, S. (2012). English language standards in higher education: From entry to exit. Camberwell, Vic: ACER Press.Google Scholar
AUQA (Australian Universities Quality Agency) (2009). Good Practice Principles for English language proficiency for international students in Australian universities, Report to the Department of Education, Employment and Workplace Relations, Canberra. www.auqa.edu.au/files/otherpublications/good%20practice%20principles%20for%20english%20language%20proficiency%20report.pdfGoogle Scholar
Bachman, L. F. & Palmer, A. S. (2010). Language assessment in practice. Oxford: Oxford University Press.Google Scholar
Beglar, D. & Hunt, A. (1999). Revising and validating the 2000 Word Level and University Word Level Vocabulary Tests. Language Testing 16.2, 131162.CrossRefGoogle Scholar
Birrell, B. (2006). Implications of low English standards among overseas students at Australian universities. People and Place 14.4, 5364. Melbourne: Centre for Population and Urban Research, Monash University.Google Scholar
Bonanno, H. & Jones, J. (2007). The MASUS procedure: Measuring the academic skills of university students. A resource document. Sydney: Learning Centre, University of Sydney. http://sydney.edu.au/stuserv/documents/learning_centre/MASUS.pdfGoogle Scholar
Chapelle, C. A., Enright, M. K. & Jamieson, J. (eds.) (2008). Building a validity argument for the Test of English as a Foreign Language. New York: Routledge.Google Scholar
Clapham, C. (1996). The development of IELTS: A study of the effect of background knowledge on reading comprehension. Cambridge: Cambridge University Press.Google Scholar
Council of Europe (2001). Common European framework of reference for languages: Learning, teaching, assessment. Cambridge: Cambridge University Press.Google Scholar
Davies, A. (1975). Two tests of speeded reading. In Jones, R. L. & Spolsky, B. (eds.), Testing language proficiency. Washington, DC: Center for Applied Linguistics, 119130.Google Scholar
Davies, A. (1990). Principles of language testing. Oxford: Blackwell.Google Scholar
Davies, A. (2008). Assessing academic English: Testing English proficiency 1950–1989: The IELTS solution. Cambridge: Cambridge University Press.Google Scholar
Davies, A. & Elder, C. (2005). Validity and validation in language testing. In Hinkel, E. (ed.), Handbook of research on second language learning. Mahwah, NJ: Erlbaum, 795813.Google Scholar
Drury, H., Dunworth, K., Kralik, C., Moore, T. & Mulligan, D. (2012). Degrees of proficiency: Post-entry English language assessment in Australia's universities. Work-in-progress session presented at the Inaugural Conference of the Association for Language Testing and Assessment of Australia and New Zealand, University of Sydney, 8–10 November.Google Scholar
Dunworth, K. (2009). An investigation into post-entry English language assessment in Australian universities. Journal of Academic Language and Learning 3.1, 113.Google Scholar
Elder, C. & Erlam, R. (2001). Development and validation of the Diagnostic English Language Needs Assessment (DELNA). Unpublished final project report. Department of Applied Language Studies and Linguistics, University of Auckland.Google Scholar
Elder, C. & von Randow, J. (2008). Exploring the utility of a web-based English language screening tool. Language Assessment Quarterly 5.3, 173194.Google Scholar
Huhta, A. (2008). Diagnostic and formative assessment. In Spolsky, B. & Hult, F. M. (eds.), The handbook of educational linguistics. Malden, MA: Blackwell, 469482.Google Scholar
Hyland, K. (2004). Disciplinary discourses: Social interactions in academic writing. Ann Arbor, MI: University of Michigan Press.Google Scholar
Kunnan, A. J. & Jang, E. E.. (2009). Diagnostic feedback in language assessment. In Long, M. H. & Doughty, C. J. (eds.), The handbook of language teaching. Oxford: Wiley-Blackwell, 610625.CrossRefGoogle Scholar
Lane, B. (2012). National regulator sharpens focus on English language standards. The Australian, 22 August. www.theaustralian.com.au/higher-education/national-regulator-sharpens-focus-on-english-language-standards/story-e6frgcjx-1226455260799Google Scholar
Lea, M. R. & Street, B. (1998). Student writing in higher education: An academic literacies approach. Studies in Higher Education 23.2, 157172.Google Scholar
Lea, M. R. & Street, B. (1999). Writing as academic literacies: Understanding textual practices in higher education. In Candlin, C. N. & Hyland, K. (eds.), Writing: Texts, processes and practices. London: Longman, 6281.Google Scholar
Manning, W. H. (1987). Development of cloze-elide tests of English as a second language. TOEFL Research Reports, No. 23. Princeton, NJ: Educational Testing Service.Google Scholar
Messick, S. (1989). Validity. In Linn, R. L. (ed.), Educational measurement (3rd edn). New York: American Council on Education and Macmillan, 13103.Google Scholar
Nation, I. S. P. (1990). Teaching and learning vocabulary. New York: Newbury House.Google Scholar
Plakans, L. (2012). Writing integrated items. In Fulcher, G. & Davidson, F. (eds.), The Routledge handbook of language testing. London: Routledge, 249261.Google Scholar
Ransom, L. (2009). Implementing the post-entry English language assessment policy at the University of Melbourne: Rationale, processes and outcomes. Journal of Academic Language and Learning 3.2, 1325.Google Scholar
Read, J. (2008). Identifying academic language needs through diagnostic assessment. Journal of English for Academic Purposes 7.2, 180190.Google Scholar