John Field Universities of Bedfordshire and Cambridge ISLI, University of Reading 12 th November, 2013.

Slides:



Advertisements
Similar presentations
AS/A2 – Making Notes Supporting Students Learning.
Advertisements

Testing Relational Database
Sruti Akula (PhD ELE) EFL University Developing academic reading skills through strategy training.
Conversation Skills will be tested both as part of Formative & Summative Assessment.
Session 4: ASSESSING SPEAKING
Wynne Harlen. What do you mean by assessment? Is there assessment when: 1. A teacher asks pupils questions to find out what ideas they have about a topic.
How to teach heterogeneous groups
Language Testing Forum 2013, Nottingham
Good at listening or good at listening tests?
TESTING SPEAKING AND LISTENING
The Historical Investigation The York School
Helping Your Child Learn to Read
Speaking English Does Not Necessarily Mean Understanding English Hanadi Mirza
ENGLISH B HIGHER LEVEL The Mackay School – May 2014 Examinations.
Reading Comprehension Paper
Presented by Eroika Jeniffer.  We want to set tasks that form a representative of the population of oral tasks that we expect candidates to be able to.
Cornell Notes.
How to Adapt Assignments and Assessments for English Language Learners
John Clegg. Contents What is CLIL? CLIL objectives What to assess in CLIL Fairness issue Ways of addressing fairness reduce the language demands of the.
The New English Curriculum
© Cambridge International Examinations 2013 Component/Paper 1.
Chapter 3 Listening for intermediate level learners Helgesen, M. & Brown, S. (2007). Listening [w/CD]. McGraw-Hill: New York.
Adopting the Process Approach to Teaching Listening Dr. Jian Kang Loar Defense Language Institute October 15, 2011.
1 Testing Oral Ability Pertemuan 22 Matakuliah: >/ > Tahun: >
Effective Marking & Feedback in Writing
Teaching Listening.
Listening Task Purpose of the test:
Effective Questioning in the classroom
Test Taking Tips How to help yourself with multiple choice and short answer questions for reading selections A. Caldwell.
National Curriculum Key Stage 2
Jude Carroll, author of Tools for Teaching in an Educationally Mobile World (Routledge 2015) Supporting teaching across cultures: the role of good practice.
© Curriculum Foundation1 Section 2 The nature of the assessment task Section 2 The nature of the assessment task There are three key questions: What are.
Dr. MaLinda Hill Advanced English C1-A Designing Essays, Research Papers, Business Reports and Reflective Statements.
Language Assessment 4 Listening Comprehension Testing Language Assessment Lecture 4 Listening Comprehension Testing Instructor Tung-hsien He, Ph.D. 何東憲老師.
Supporting your child with reading.
Communicative Language Teaching (CLT)
Writing Literary Analysis Papers
Vocabulary Link Listening Pronunciation Speaking Language Link LESSON A Writting Reading Video Program.
VCE Learning. To unpack the challenge of enhancing the quality of VCE learning What does the student need to know about how to interpret the task ? Ho.
Principles in language testing What is a good test?
ELA Common Core Shifts. Shift 1 Balancing Informational & Literary Text.
The New English Curriculum September The new programme of study for English is knowledge-based; this means its focus is on knowing facts. It is.
Essay and Report Writing. Learning Outcomes After completing this course, students will be able to: Analyse essay questions effectively. Identify how.
Teaching Productive Skills Which ones are they? Writing… and… Speaking They have similarities and Differences.
LAS LINKS DATA ANALYSIS. Objectives 1.Analyze the 4 sub-tests in order to understand which academic skills are being tested. 2.Use sample tests to practice.
Academia Británica Pulling teeth UTTERANCE above ALL March ̍11 UTTERANCE above ALL Academia Británica Pulling teeth March ̍11 um, so...what are we talkin’about?
DEVELOPING LISTENING Alejandra Echague C DEVELOPING LISTENING IN A FOREIGN LANGUAGE 1. The foundation skill First to be acquired Mother skill 2.
Lectures ASSESSING LANGUAGE SKILLS Receptive Skills Productive Skills Criteria for selecting language sub skills Different Test Types & Test Requirements.
Teaching Reading Comprehension
Assessment. Workshop Outline Testing and assessment Why assess? Types of tests Types of assessment Some assessment task types Backwash Qualities of a.
 An article review is written for an audience who is knowledgeable in the subject matter instead of a general audience  When writing an article review,
Writing. Academic Writing Allow about 20 minutes In TASK 1 candidates are presented with a graph, table,chart or diagram and are asked to describe, summarise.
Listening comprehension is at the core of second language acquisition. Therefore demands a much greater prominence in language teaching.
Learning Through Failure. Reflect O Take a few moments to write down your answers to the following questions: O What was your reaction to the video? O.
Objectives of session By the end of today’s session you should be able to: Define and explain pragmatics and prosody Draw links between teaching strategies.
CERTIFICATE IN ASSESSING VOCATIONAL ACHIEVEMENT (CAVA) Unit 1: Understanding the principles and practices of assessment.
Key Stage 2 Portfolio. Llafaredd / Oracy Darllen / Reading Ysgrifennu / Writing Welsh Second Language.
UNIT 10 Teaching Reading. Aims of the unit In this unit,We are going to discuss how to teach reading. We will focus on the following: 1.How do people.
Overview of Standards for Literacy in History/Social Studies, Science, and Technical Subjects The Common Core State Standards.
To my presentation about:  IELTS, meaning and it’s band scores.  The tests of the IELTS  Listening test.  Listening common challenges.  Reading.
Academic Writing Fatima AlShaikh. A duty that you are assigned to perform or a task that is assigned or undertaken. For example: Research papers (most.
1. Chapter Preview Part 1 – Listening in the Classroom  Listening Skills: The Problem and the Goal  Listening Tasks in Class Part 2 – Listening outside.
Teaching Listening Why teach listening?
An –Najah National University Submitted to : Dr. Suzan Arafat
Dr Anie Attan 26 April 2017 Language Academy UTMJB
Testing testing: developing tests to support our teaching and learning
SPEAKING ASSESSMENT Joko Nurkamto UNS Solo 11/8/2018.
Teaching Listening Based on Active Learning.
SPEAKING ASSESSMENT Joko Nurkamto UNS Solo 12/3/2018.
National Curriculum Requirements of Language at Key Stage 2 only
Presentation transcript:

John Field Universities of Bedfordshire and Cambridge ISLI, University of Reading 12 th November, 2013

With thanks to Cambridge English language assessment and the ISLC, University of Reading, for their support for the research cited here

A theme of the talk Teachers and EFL materials writers tend to favour standard test formats: partly to prepare learners for international tests and partly through lack of alternatives But the needs of teachers and the designers of international high stakes tests are clearly very different Testers aim for: Reliability (avoiding ‘subjective judgement’) Ease of marking Teachers / local testers need: In depth information about the listening skills of individual learners so that testing can lead into instruction.

What do we think we test? Comprehension. So what do we mean by comprehension? ‘Understanding’ Understanding what? Giving correct answers to comprehension questions

What do we think we test? 2 ‘Listening for gist’ ‘Listening for information’ ‘Listening for main idea’ ‘Local listening’ ‘Global listening’

What do we think we test? 3 [CEFR B2 goals] ‘Can understand standard spoken language, live or broadcast, on both familiar and unfamiliar topics normally encountered in personal, social, academic or vocational life. Only extreme background noise, inadequate discourse structure and/or idiomatic usage influence the ability to understand Can understand the main ideas of propositionally and linguistically complex speech on both concrete and abstract topics delivered in a standard dialect, including technical discussions in his/her field of specialisation Can follow extended speech and complex lines of argument provide the topic is familiar and the direction of the talk is signposted by explicit markers’..

OK, so… That tells me what L2 listeners should aim for under my teaching/ testing … But how are they supposed to get there? These descriptors define the input or output in listening, but say nothing about the process we are trying to test. They cannot be said to support assessment for learning.

A cognitive account

The need for a cognitive approach There is a new interest among testers in what goes on in the mind of the test taker. We need to know whether high-stakes test actually test what they claim to test. Can a listening test, for example, accurately predict the ability of a test taker to study at an English medium university? At local level, we need to use tests to diagnose learner problems so that the tests can feed into learning. This is especially true of listening.

Cognitive validity asks… Does a test elicit from test takers the kind of process that they would use in a real-world context? In the case of listening, are we testing the kinds of process that listeners would actually use ? Or do the recordings and formats that we use lead test takers to behave differently from the way they would in real life?

Two possible approaches A. Ask learners to report on the processes they adopted when taking a test (e.g. by explaining how they got their answers) B. Use a model of listening that is supported by evidence from psychology. Match the processes produced by a test against the model.

Listening to learners

Learner report 1: Location Item: A demand for golf courses attracted the interest of both ………… and businessmen. Key: There was … enormous interest amongst landowners not to mention businessmen S: I think I um + I the key words. I think most + most useful for me is the ‘businessmen’ R: right S: because when I heard this before + I heard I heard ‘landowners’ and ‘businessmen’ R: so you you recognised the the word ‘landowners’ S: oh yeah R and [it was] close to the word ‘businessmen’ S: yeah this is ever close so I think maybe

Conclusion Test takers listen out for words from the (written) items in order to locate where the answer is likely to be.

Learner Report 2: order Professional Development for IATEFL R: is there anything that you heard that helped you? S: I have the problem about that because I am concentrate on the two of the questions so …I didn’t realise R: so S: his his + he’s already go to the 9 R: right ok so you were still listening out for number 8 S: yeah and number 7

Conclusions Professional Development for IATEFL Learners recognise and exploit the convention that questions are asked in the same order as the recording. This provides them with a rough sequential outline of the recording before they even begin to listen. If a listener fails to find the answer to one question, he/she may go on to miss the answer to the next one as well..

17 What a written set of items provides The items in (e.g.) a gap-filling task potentially provide a candidate with: An outline of what the recording covers A set of gapped sentences that follow the sequence of the recording Key words with which to locate information Sequences which may echo the wording of the recording or the order of words

Learner report 3: prominent words Professional Development for IATEFL Correct answer: Tom suggests that golf courses could also be used as NATURE RESERVES’ S: number 13 is I’m not sure but um + he said ‘crack’ R: you heard the word ‘crack’? S: crack …but I don’t know the meaning of ‘crack’ R: er you know it seemed to be an important word S: yes I think so R: ok + how did you spell ‘crack’ if if you don’t know the S: c-r-a-c-k R: right so you guessed the spelling did you? S: I guess yes Most importantly, courses should be designed to attract rather than drive away wildlife.

Conclusion Professional Development for IATEFL Learners sometimes simply listen out for prominent words – even if they do not understand them. This is partly a reflection of their level. At level B1 and below, listeners are very dependent upon picking up salient words rather than chunks or whole utterances. This tendency is increased by the use of gap filling tasks, which focus attention at word level.

General conclusions a. Conventional test formats provide much more information than is available in real-world contexts (and do so in a written form) BUT… b. Conventional test formats may also be more demanding than real-life listening because of divided attention effects, where the learner has to read and listen or read, write and listen.

Recordings Does the input impose similar listening demands to those of a real-world speaker?

Natural speech ( Recording Level B2) To what extent do these recordings resemble authentic everyday speech?

Recording origin Authentic Scripted Semi-scripted / re-recorded Improvised ‘All tests are based on authentic situations’ Cambridge ESOL PET Handbook

Why re-recorded material? Exam Boards prefer this type because it enables them to  ‘ Reduce noise’  Control speech rate  Simplify vocabulary and grammar if necessary  Introduce distractors  Eliminate redundancy (or add it with single-play tasks)

Some conclusions on studio recordings Actors adapt their delivery to fit punctuation. They pause regularly at the ends of clauses There are few hesitation pauses. No overlap between speakers

Speaker variables Accent Speech rate: speed and consistency Pausing Level and placing of focal stress Number of speakers Pitch of voice; familiarity of voice Precision of articulation

Normalisation and testing L2 listening Test takers need time to adjust (normalise) to the voice of an unfamiliar speaker. Best not to focus questions on the opening 10 seconds of a longer recording. Because of the need to normalise, it is best not to have too many speakers in a test recording. Listening difficulty increases as the number of voices increases beyond one M and one F (Brown & Yule, 1983). Adapting to voices is cognitively demanding. Testers must bear in mind the cognitive demands of normalising to speech rate and voice pitch. Is it fair to add to those demands by featuring a variety of accents?

Tasks Does the task elicit processes which resemble those that a listener would use in a real-world listening event?

Task types in international tests Multiple-choice Gap filling True/False/Don’t know Multiple matching: Identify which of the five speakers is a lorry driver / a politician / a musician Visual multiple choice Examination boards recognise that all of these have their drawbacks - which is why they argue for a mixture of tasks

Multiple choice questions You hear an explorer talking about a journey he’s making. How will he travel once he is across the river? A. by motor vehicle B. on horseback C. on foot (FCE Handbook, 2008: 60)

Recording 1 (FCE Sample Test 1:1 ) trucks carry foot then use horses rather than trucks pick up the vehicles The engine’s full of water at the moment, it’s very doubtful if any of the trucks can get across the river in this weather. The alternative is to carry all the stuff across using the old footbridge, which is perfectly possible …and then use horses rather than trucks for the rest of the trip all the way instead of just the last 10 or 15 kilometres as was our original intention. We can always pick up the vehicles again on the way down…

Recording 1 (FCE Sample Test 1:1 ) trucks carry foot then use horses rather than trucks pick up the vehicles The engine’s full of water at the moment, it’s very doubtful if any of the trucks can get across the river in this weather. The alternative is to carry all the stuff across using the old footbridge, which is perfectly possible …and then use horses rather than trucks for the rest of the trip all the way instead of just the last 10 or 15 kilometres as was our original intention. We can always pick up the vehicles again on the way down…

Conclusion Conventional formats require the listener to: Map from written information to spoken Eliminate negative possibilities as well as identify positive ones (esp with MCQ and T/F) Read and write as well as listen (esp gap filling) Engage in complex tasks which take us well beyond listening (esp. multiple matching)

The task: solutions for the teacher / local tester

Suggestions for using conventional tasks Provide items after a first playing of the recording and before a second. This ensures more natural listening, without preconceptions or advance information other than general context. Keep items short. Loading difficulty on to items (especially MCQ ones) just biases the test in favour of reading rather than listening. Items should avoid echoing words in the recording Favour tasks (e.g. multiple matching) that allow items to ignore the order of the recording and to focus on global meaning rather than local detail.

More natural tasks Ignore the questions in the coursebook or present them orally. Ask questions and get answers in the first language Use whole class oral summary (What have you understood so far?), then replay the recording At lower levels of English, ask learners to transcribe small parts of a recording At higher levels, use note-taking and reporting back Get learners to work in pairs and compare notes

Items: What to target in a listening test?

Five phases of listening ( Field 2008) Speech signal Words Meaning 38 Decoding Word search Parsing Meaning construction Discourse construction

Targets An item in a test can target any of these levels: Decoding: She caught the (a) 9.15 (b) 9.50 (c) 5.15 (d) 5.50 train. Lexical search: She went to London by ……. Factual information: Where did she go and how? Meaning construction: Was she keen on going by train? Discourse construction. What was the main point made by the speaker?

Targeting levels of listening In theory, a good test should target all levels of listening in order to provide a complete picture of the test taker’s command of all the relevant processes. In practice, higher levels may be too demanding in the early stages of L2 listening. Novice listeners focus quite heavily on word-level decoding, which does not leave them enough spare attention to give to wider meaning. In addition, certain test formats may tap almost exclusively into one level. Gap-filling is a good example

Higher-level listening

Higher processes (Field 2008) PROPOSITION MEANING REPRESENTATION ENRICH MEANING HANDLE INFO DISCOURSE REPRESENTATION

Implications for testing Questions may and should be asked at three levels: Factual: Factual: local information Meaning in context: Meaning in context: requiring the listener to relate what the speaker says to the context or to draw conclusions which are not expressed by the speaker Discourse: Discourse: showing a global understanding of what was said (including speaker intentions etc.)

Meaning representation The listener has to: context Relate what was said to its context Enrich Enrich the meaning (drawing upon world knowledge) inferences Make inferences reference Resolve reference (she, it, this, did so) Interpret Interpret the speaker’s intentions All of these indicate possible question types

Discourse building / handling information ChooseConnectCompareConstruct Is it important? Is it relevant? How is it linked to the last utterance? Is what I think I heard consistent with what was said so far? What is the overall line of argument?

Spread of targets

Why is information handling omitted in present test design? Choose: the tester chooses which information points to focus on – sometimes choosing points that are not central to the recording Connect: Much testing focuses on single points, with no connection to those before and after Compare: Tests rarely ask learners to check information (for example, comparing two accounts of an accident) Construct. Tests rarely seek for evidence that learners can construct an outline based upon macro-and micro points / headings and subheadings

Solutions for local testers Ask questions at discourse level: What is the main point of the recording? / Give three main points. What is the connection between Point A and Point B? Complete a skeleton summary of the text with main points and sub-points Ask learners to compare two recordings for similarities and differences Ask learners to summarise a recording orally or in the form of notes (in L1 or L2)

Some thoughts on teacher testing of listening and its impact on teaching

The inflexibility of high stakes tests Large scale high-stakes tests Large scale high-stakes tests have major constraints which prevent them from testing listening in a way that fully represents the skill. Reliability and ease of marking Highly controlled test methods, using traditional formats that the candidate knows Little attention possible to individual variation or alternative answers

Advantages of more local tests and tasks Local smaller-scale tests Local smaller-scale tests afford the possibility of testing a wider rage of listening processes with: More open ended questions More scope for testing information handling Marking on an individual basis Possible acceptance of alternative answers

Progress testing / diagnostic teaching Properly designed, progress tests might enable the tester to diagnose specific listening problems. In a follow up (ideally soon after), the teacher/tester can ask: Why did you give that answer? What do you think you heard? In this way, a test can help to determine which aspects of listening should be focused on in later small-scale practice exercises. In other words, this kind of test can be formative rather than just judgemental.

References Field, J. (2008) Listening in the Language Classroom. Cambridge: CUP Field, J. (2009). The cognitive validity of the lecture listening section of the IELTS listening paper. IELTS Research Reports 9, Cambridge Field, J. (2013) Cognitive validity. In Geranpayeh, A. & Taylor, L. (eds.) Examining Listening. Cambridge: Cambridge University Press

Thanks for listening!

What makes listening a special case? Transitory: no long-term record to refer to Happens in real time Need to store while analysing Need to carry forward information in the mind Speech rate is not under listener's control Few word boundary markers The speech signal is highly variable as compared to spelling / fonts.

Two consequences of high variability Knowledge of a word  recognition of the word in connected speech Many high level errors of comprehension have their origins in low-level errors of word recognition