1 The Swiss ‘IEF’ Project - Assessment Instruments Supporting the ELP by Peter Lenz University of Fribourg/CH Voss/N, 3/06/05.

Slides:



Advertisements
Similar presentations
Assessment types and activities
Advertisements

Training teachers to use the European Language Portfolio Former les enseignants à lutilisation du Porfolio européen des langues.
Training teachers to use the European Language Portfolio Former les enseignants à lutilisation du Porfolio européen des langues.
Training teachers to use the European Language Portfolio Former les enseignants à lutilisation du Porfolio européen des langues.
Victorian Curriculum and Assessment Authority
Peter Lenz IBE SeminarWarsaw, 20/10/2011 A Language Assessment Kit – Relating to the CEFR – for French and English.
Japanese University Students’ Attitudes toward the Teacher’s English Use Koji Uenishi Hiroshima University.
How does DIALANG use the CEF?
Spiros Papageorgiou University of Michigan
KEMENTERIAN PENDIDIKAN DAN KEBUDAYAAN BADAN PENGEMBANGAN SUMBER DAYA MANUSIA PENDIDIKAN DAN KEBUDAYAAN DAN PENJAMINAN MUTU PENDIDIKAN AUTHENTIC ASSESSMENT.
AMEP Assessment Task Bank Professional Development Kit Listening Developed by Marian Hargreaves for NEAS 2013 © NEAS Ltd
Using the CEFR in Catalonia Neus Figueras
The reform of A level qualifications in the sciences Dennis Opposs SCORE seminar on grading of practical work in A level sciences, 17 October 2014, London.
DMe - Dick Meijer Talen Consultancy 1 CEFR and ELP seminar Introduction SKOPJE 16th and 17th February 2007 Dick Meijer.
Relating language examinations to the Common European Framework of Reference for Languages (CEFR) Waldemar Martyniuk Waldemar Martyniuk Language Policy.
VeldwERK: What happens when you step into the CEFR Seminar on Curriculum Convergences Council of Europe, Strasbourg 29th November, 2011 Daniela Fasoglio,
Common European Framework of Reference for Languages (CEFR): Learning, Teaching, Assessment Nuppu Tuononen Palmenia Centre for Continuing Education
6 th semester Course Instructor: Kia Karavas.  What is educational evaluation? Why, what and how can we evaluate? How do we evaluate student learning?
Raili Hildén University of Helsinki Relating the Finnish School Scale to the CEFR.
* Discussion: DO YOU AGREE OR DISAGREE WITH THESE STATEMENTS? WHY OR WHY NOT? 1.The difficulty of a text depends mostly on the vocabulary it contains.
Classroom Assessments Checklists, Rating Scales, and Rubrics
EDU 385 Education Assessment in the Classroom
TASK-BASED ASSESSMENT: INTEGRATING RESEARCH AND PROFESSIONAL DEVELOPMENT Geoff Brindley, Stephen Moore, Helen Slatyer & Steven Ross Macquarie University,
Communicative Language Teaching
Supplementary materials
Workshop: assessing writing Prepared by Olga Simonova, Maria Verbitskaya, Elena Solovova, Inna Chmykh Based on material by Anthony Green.
ELT 415 Material Assessment PART IV 1. THREE PIECES OF ADVICE Try to get as much information as possible by asking for it specifically or by trying to.
Score the Washington Educational Technology Assessments Educational Technology.
RUBRICS AND CHECKLISTS KEITHA LUCAS HAMANN ASSESSMENT IN ARTS EDUCATION.
Military Language Testing at the National Defence University and the Common European Framework BILC CONFERENCE BUDAPEST.
Teacher Candidates Practice Oral Language Functions Using Digital Video Recording Martha E. Castañeda Miami University Ohio Foreign Language Association.
Issues for Introducing Early Foreign Language Learning No theoretical optimum age to start teaching Early learning of non-mother tongue should be integrated.
European language portfolio: promoting a lifetime of vocationally-oriented language learning Project BG/0/B/F/LA
Assessment and Testing
1 Using the Learning Progression Framework (LPF) to Enhance the Learning, Teaching and Assessment of English Language at Primary Level 8 & 10 December.
How Much Do We know about Our Textbook? Zhang Lu.
Assessment. Workshop Outline Testing and assessment Why assess? Types of tests Types of assessment Some assessment task types Backwash Qualities of a.
National Standards in Reading & Writing Sources : NZ Ministry of Education websites. G Thomas, J Turner.
Relating examinations to the CEFR – the Council of Europe Manual and supplementary materials Waldek Martyniuk ECML, Graz, Austria.
COURSE AND SYLLABUS DESIGN
The CEFR in the Netherlands Erna van Hest, Cito Strasbourg, 6-8 February 2007.
1 Collecting and Interpreting Quantitative Data Deborah K. van Alphen and Robert W. Lingard California State University, Northridge.
Monitoring and Assessment Presented by: Wedad Al –Blwi Supervised by: Prof. Antar Abdellah.
Midterm Presentation- HOSPITALITY LANGUAGE IN DIFFERENT PLACE Min-Han Tsai (Tony) AFL 1A.
Session 2 English Language Proficiency Benchmarks Assessment Primary Professional Development Service.
Selection and Use of Supplementary Materials and Activities
RelEx Introduction to the Standardization Phase Relating language examinations to the Common European Framework of Reference for Languages Gilles Breton.
The Future for Assessment? Assessing Pupil Progress (APP) as a tool for effective Teacher Assessment in Primary Science.
To my presentation about:  IELTS, meaning and it’s band scores.  The tests of the IELTS  Listening test.  Listening common challenges.  Reading.
AAPPL Assessment Follow Up June What is AAPPL Measure? The ACTFL Assessment of Performance toward Proficiency in Languages (AAPPL) is a performance-
Relating Foreign Language Curricula to the CEFR in the Maltese context
Classroom Assessments Checklists, Rating Scales, and Rubrics
Information for Parents Key Stage 3 Statutory Assessment Arrangements
Introduction to the Specification Phase
ECML Colloquium2016 The experience of the ECML RELANG team
Assessment in Language Teaching: part 1 Lecture # 23
Chapter 6: Checklists, Rating Scales & Rubrics
Unit 3 The National English Curriculum
SMART objectives for smart teachers
Key findings on comparability of language testing in Europe ECML Colloquium 7th December 2016 Dr Nick Saville.
Introduction to the Validation Phase
Introduction to the Validation Phase
Classroom Assessments Checklists, Rating Scales, and Rubrics
Parent Reading Workshop
RELATING NATIONAL EXTERNAL EXAMINATIONS IN SLOVENIA TO THE CEFR LEVELS
English Language Proficiency Benchmarks Assessment
Why do we assess?.
FInal Exam SpecIfıcatıons
Training teachers to use the European Language Portfolio
OCS: Putting the Assessment Pieces Together
Presentation transcript:

1 The Swiss ‘IEF’ Project - Assessment Instruments Supporting the ELP by Peter Lenz University of Fribourg/CH Voss/N, 3/06/05

EYL: Launch of ELP 15+ in Switzerland In 2001 the Swiss Conference of Cantonal Ministers of Education recommends to the cantons  to consider the CEFR  in curricula (objectives and levels)  in the recognition of diplomas  to facilitate wide use of the ELP 15+  make ELP accessible to learners  help teachers to integrate ELP in their teaching  to develop ELPs for younger learners

3 Integrating the ELP – School Teachers’ Wishlist : More descriptors taylored to young learners ‘ needs : Less abstract formulations : Self-assessment grid and checklists with finer levels : Tools facilitating “hard” assessment :  Test tasks relating to descriptors  Marked and assessed learner texts  Assessed spoken learner performances on video  Assessment criteria relating to finer levels for Speaking and Writing

4 Meeting the Needs for ELP IEF Project ( ) “Instruments for the Assessment of Foreign-Language Competences - English, French” FL German-speaking cantons of Switzerland Principality of Liechtenstein Bildungsplanung Zentralschweiz  Peter Lenz & Thomas Studer (UniFR)

5 IEF: Overview of Expected Products Bank of validated test tasks (  5 “skills”; C-tests) Benchmark performances (Speaking, Writing) Bank of target-group-specific descriptors (levels A1.1-B2.1) Tests available from publisher Tests for evaluations Assessment criteria (Speaking, Writing) (Self-)assessment checklists Training packages for teacher training ELP

6 Phase I A New Bank of Can-do Statements How did the new descriptors take shape? 1) Collecting from written sources (ELPs, textbooks, other sources)  Teachers decide on relevance for target learners and on suitability for assessment  Teachers complement collection 2) Validating, amending the collection in workshops 3) Fine-tuning and selecting descriptors  Make formulations non-ambiguous and accessible; add examples  Select descriptors to cover whole range of levels A1.1 - B2.1  Represent wide range of skills and tasks  ~330 descriptors for empirical phase

7 Assessment questionnaires – Teachers assess their pupils Following Schneider & North‘s methodology for the CEFR Phase I Calibrating Additional Descriptors

8 Linked and anchored assessment questionnaires of 50 descriptors each for different levels 2 parallel sets of descrip- tors of similar difficulty per assumed level Identical descriptors as links (& sometimes CEFR anchors) Too few learners at B2

9 Phase I Calibrating Additional Descriptors Statistical analysis and scale-building (A1.1 - B1.2)

10 Phase II Adapting Descriptors for Self-assessment (Self-)assessment checklists ELP Bank of target-group-specific descriptors (levels A1.1-B2.1)

11 Phase II Reformulations – Can …  I can... 1.Some Can do s are transformed into I can s 2.Learners are asked for feedback: learners assess themselves and give feedback on that 3.Whole bank of Can do s is transformed into I-can statements

12 Phase II Checklists for the New Swiss ELP (Self-)assessment checklists Bank of target group-specific descriptors (levels A1.1-B2.1) Drawing on 3 sources

13 ELP II: Self-assessment in Relation to Finer Levels

14 Phase III Developing Test Tasks and Instruments Bank of validated test tasks (Self-)assessment checklists Bank of target-group-specific descriptors (levels A1.1-B2.1) ELP

15 Phase III Test Tasks and Instruments  Speaking tasks (production and interaction)  Writing tasks  Listening tasks  Reading tasks 1) Test tasks relating to communicative language ability 2) C-Tests (integrative tests)  C-Tests (type of CLOZE) are said to provide reliable information on a learner‘s linguistic resources esp. for (written) Production.  Most test tasks are related to one descriptor, sometimes two – but descriptor difficulty vs. task diff.?  The test tasks are field-tested and attributed to a level at least tentatively  Validation: tests + teacher questionnaires

16 Phase IV Assessment Criteria for Performances Assessment criteria for Speaking and Writing Bank of target-group-specific descriptors (levels A1.1-B2.1) (Self-)assessment checklists ELP Bank of validated test tasks (mainly performance-oriented)

17 Phase IV Developing Criteria for Speaking How did the criteria take shape? – Steps taken:  Collect criteria from various sources: CEFR, examination schemes... 1) Collecting criteria  Teachers bring video recordings  Teachers describe differences between learner performances they can watch on video  more criteria  Teachers adopt and apply descriptors from existing collection  Teachers agree on essential categories (e.g. Vocab range, Pronunciation/Int. ) and build a scale for each analytical category 2) Assessing spoken performances in workshops 3) Preparing empirical validation  Decide on categories to be retained  Revise and complete proposed scales of analytical criteria

18 Phase IV Producing Video Tapes With Spoken Performances One learner - different tasks in various settings

19 Phase IV Empirical Validation of Speaking Criteria Methodology A total of 35 teachers (14 Fr, 21 En) apply  58 analytical criteria (some from CEFR ) belonging to 5 categories  28 task-based descriptors (matching performed tasks)  to 10 or 11 video-taped learners per language, each performing 3-4 spoken tasks Criteria categories  Interaction  Vocabulary range  Grammar  Fluency  Pronunciation & Intonation

20 Phase IV Calibrating Criteria for Speaking Criteria and questionnaires - a linked and anchored design CEFR Anchors 3 assessment questionnaires for three different learner levels “Statement applies to this pupil but s/he can do clearly better” “Statement generally applies to this pupil ” “Statement doesn‘t apply to this pupil” For reasons of practicality: only 3-step rating scale for Can descriptors/criteria !! Links between questionnaires

21 Phase IV Criteria for Speaking – Analysis (1) The 5 analytical categories retained – Correlations and Fit InteractionVocabGrammarFluencyPronuncia- tion Overall Interaction 1.00 Vocab Grammar Fluency Pronuncia- tion Overall Disattenuated correlations between pupil measures suggest proximity of categories/competences – except Pronunciation/Intonation FACETS indicates slight misfit for Fluency ; overfit for Interaction (.88)

22 Phase IV Criteria for Speaking – Analysis (2) Criteria applied to French and English – Diagnosing DIF

23 Phase IV Criteria for Speaking – Analysis (3) Teacher severity and consistency Consistency: 5 out of 35 raters were removed from the analysis due to misfit of up to 2.39 logits (infit mean square) Severity: Some extreme raters (severe or lenient) show a strong need for rater training although every criterium makes a meaningful (though somewhat abstract) statement on mostly observable aspects of competence. Map for English

24 Phase IV Criteria for Speaking – Anchoring (1)  11 analytical criteria from the CEFR linked design (mostly Interaction and Vocabulary)  A total of 28 task-oriented descriptors from the IEF bank in a linked design  Known scores of 3 learners of English rated in teacher workshop on CEFR basis Used here Potential anchors towards the CEFR in the data:

25 Phase IV Criteria for Speaking – Anchoring (2)  CEFR difficulties (x-axis) and IEF difficulties (y-axis) of criteria are plotted (blue diamonds) using a scaling factor for equating the separate calibrations.  Lines visualize the usual 95% confidence interval that helps detect items that are not suitable as anchors (outliers). Outlier – perceived more difficult in IEF (over 3 logits): Can link groups of words with simple connectors like ‘and’, ‘but’ and ‘because’. Outlier removed

26 Phase IV Criteria for Speaking – Taking Stock  The calibrations of the video-taped learners are very plausible.  IEF now has video-taped examples of learners from below A1 (French only) to a very high B2.  The additional, newly-developed criteria are well spread across the targeted level range A1-B2 (A1.1?). But: What will the assessment instruments to be used in schools look like?

27 Phase IV Assessment instruments for Speaking Problem: middle category “Statement generally applies to this pupil “ – desirable (because of its meaningfulness) but possibly too broad Range here: to logits Solutions?  Other formulations for narrower categories?  Use e.g. B1.2 descrip- tor to establish A2.1 of a learner?  ???

28 Phase IV Assessment instruments for Speaking Narrower categories - Can the middle category be divided up into three? Range of middle category: -1.2 to +0.8 logits Main problem: Raters have the impression to apply modifiers upon modi- fiers, new restrictions upon restrictions already present in the criteria. 0…0… 1 Pupil has this ability only partially 2 Pupil generally has this ability. 3 Pupil fully has this ability 4…4…