Fightin’ Words Points for discussion Points for review Looking for a report structure?

Slides:



Advertisements
Similar presentations
Test Construction A workshop. Activity 1 Using the information you read Brown (2004) (particularly on pages 51-64), develop criteria for evaluating an.
Advertisements

Mywish K. Maredia Michigan State University
Putting things together: Textbooks (framework/ context/ language focus) Small readers Poems / Plays Big Books …. Other resources: Supplementary Grammar.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
ISHIKAWA DIAGRAM – Tool for quality management Marit Laos IS Project Management
Assessment. Schedule graph may be of help for selecting the best solution Best solution corresponds to a plateau before a high jump Solutions with very.
Teacher Performance Assessment (edTPA) Partner School Visits Winter, 2014.
Supporting Teachers to make Overall Teacher Judgments The Consortium for Professional Learning.
Literacy Test Preparation Grade 10 English Booklet 1, Section II: Writing Page 7 Booklet 1, Section V: Writing Page 15 Booklet 2, Section VIII: Writing.
Psychology 202b Advanced Psychological Statistics, II February 1, 2011.
Summary 1 l The Analytical Problem l Data Handling.
1 Canonical Analysis Introduction Assumptions Model representation An output example Conditions Procedural steps Mechanical steps - with the use of attached.
Universal Access: More People. More Situations Content or Graphics Content or Graphics? An Empirical Analysis of Criteria for Award-Winning Websites Rashmi.
Midterm Review Evaluation & Research Concepts Proposals & Research Design Measurement Sampling Survey methods.
As with averages, researchers need to transform data into a form conducive to interpretation, comparisons, and statistical analysis measures of dispersion.
NAACL Workshop on MTE 3 rd in a Series of MTE adventures 3 June, 2001.
IACT303 – INTI 2005 World Wide Networking Welcome and Introduction to Subject. Penney McFarlane The University of Wollongong.
Robert delMas (Univ. of Minnesota, USA) Ann Ooms (Kingston College, UK) Joan Garfield (Univ. of Minnesota, USA) Beth Chance (Cal Poly State Univ., USA)
Web Site Creation: Good Practice Guidelines Self Assessment Of Project Web Sites Marieke Napier NOF-digitise Advisor UKOLN University of Bath UKOLN is.
Introduction to NYC Teacher Data Initiative Training for Schools Fall 2008.
Meteorology 415 Fall  Syllabus  Forecasting is like….  Appreciating the methods…10,000 hours  Basic course plan  Welcome to the web page 
Teaching Reading and Writing to English Language Learners CEDEI Dr. Kathleen McInerney.
OCTOBER ED DIRECTOR PROFESSIONAL DEVELOPMENT 10/1/14 POWERFUL & PURPOSEFUL FEEDBACK.
Your PowerQuest Title A ? PowerQuest for ? Grade By: Insert Your Name and School.
1 Name: Wael Eid Grade: 11 English project The SAT ® What Does It Mean for Students?
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
Esri International User Conference | San Diego, CA Technical Workshops | Spatial Statistics: Best Practices Lauren Rosenshein, MS Lauren M. Scott, PhD.
Issues in structural equation modeling Hans Baumgartner Penn State University.
Multivariate Data Analysis Chapter 8 - Canonical Correlation Analysis.
Audience, Purpose, & Context Who? Why? What?. Analyzing the Audience  Who will be reading, listening to, or using this material?  What particular characteristics.
Building the Performance Culture. Introduction: Reflections from the Dean  University Context »Transparent and defensible process. »Evidence based »Academic.
Evaluation of software engineering. Software engineering research : Research in SE aims to achieve two main goals: 1) To increase the knowledge about.
Basic Impact Evaluation Course Poverty Analysis Initiative (C102) Poverty Analysis Initiative (C102) World Bank Institute Attacking Poverty Program.
BREAKOUT 1: Identifying the Gap (or Journey) (13.45 – 15.00)
Advantages of Using Children’s Literature provides a motivating introduction to complex curriculum topics mathematical vocabulary can be reinforced and.
Technology in the classroom. UM weather Great way to begin a day or class –Links to 300 weather sites –Links to 700 web cams to view weather –Radar and.
Reliability Chapter 3.  Every observed score is a combination of true score and error Obs. = T + E  Reliability = Classical Test Theory.
Evaluating and measuring impact in career development: extension workshop Presented by – Date – Just to identify strengths and areas to improve are no.
Linguistic Markers to Improve the Assessment of students in Mathematics: An Exploratory Study Sylvie NORMAND-ASSADI, IUFM de Créteil Lalina COULANGE, IUFM.
Grade 1 Books. GRADE 2 Justice for All, tackles big ideas: justice, equality, liberty, and freedom. Students examine our nation’s promise of equality.
January 14, 2004 The three r's of Software Testing © James Bullock, Brian Branagan contact: 1 Seattle QA-SIG The three r's of.
Chapter 16 Data Analysis: Testing for Associations.
Remote Sensing Classification Accuracy
Chapter 2 Doing Sociological Research Key Terms. scientific method Involves several steps in research process, including observation, hypothesis testing,
Multivariate Data Analysis Chapter 1 - Introduction.
JOE CHIHADE - CHERRY DANIELSON ELLEN IVERSON HHMI ASSESSMENT WORKSHOP.
Michigan State University Gathering and Timely use of Feedback from Individualized On-Line Work. Matthew HallJoyce ParkerBehrouz Minaei-BigdoliGuy AlbertelliGerd.
Mehmet Dali Öztürk, Ph.D. 3CSN and ELI Collaborate: Data Collection Workshop September 6, 2013.
IDEA STUDENT EVALUATION REPORTS Insight Improvement Impact ® Using IDEA as a Tool for Reflection about Your Teaching Raritan Valley Community College January.
Concluding your Answer. New Ideas The purpose of the conclusion is not simply to summarise the points already made The conclusion should cover new ground.
Data for Student Success May, 2010 Ann Arbor, MI “It is about focusing on building a culture of quality data through professional development and web based.
A questions B questions C questions D questions E questions Skills RPE mock – Paper 1.
Evaluation of statistical methods for meta-analysis Julian Higgins School of Social and Community Medicine University of Bristol, UK 1 Cochrane Methods.
January 22, 2014 Dr. Warren Wessel – Instructor ECS 410 – Assessment and Evaluation.
Dan Thompson Oklahoma State University Center for Health Science Evaluating Assessments: Utilizing ExamSoft’s item-analysis to better understand student.
29 th May 2012 Waingels College. Aims To explain the role of literacy in PE Demonstrate and share ideas to support the impact of literacy in PE Recognise.
Funders typically looking for applications to demonstrate: Evidence of need for what you are proposing, and why your project is the right approach to meeting.
AQA Unit 1: Exploring modern texts English Literature: 1 hour and 30 minutes.
BY MADELINE GELMETTI INCLUDING STUDENTS WITH DISABILITIES AND ENGLISH LEARNERS IN MEASURES OF EDUCATOR EFFECTIVENESS.
Logistic Regression: Regression with a Binary Dependent Variable.
Academic Conversations
Introduction to the Validation Phase
Monitoring and Assessment Assignment
A Level English Language
CVE.
Multidimensional Scaling
CPEA RESULTS.
Product moment correlation
Legacy transcripts Planning Meeting March 7, 2019.
Presentation transcript:

Fightin’ Words Points for discussion Points for review Looking for a report structure?

System Assessment What metrics are suitable for assessing what system characteristics? What system characteristics reflect what user needs? Is there a radical difference between evaluation focusing on research or development needs and evaluation focusing on end-user needs? When should real-world data be used? What is the impact of using it?

Metric Choice What constitutes a valid metric How can you demonstrate that a metric is valid? What metrics can be automated What are the advantages/disadvantages of specific metrics? For the metric(s) selected, what are the difficulties in applying them?

Metric Choice - II For a given metric, what variations in scores are typically produced? What are the statistical error variances? For a given metric, what are the score ranges for ‘good’ and ‘bad’ systems Are there metrics which correlate with one another? Are there metrics which indicate an overall quality score? Are there metrics which work better for specific language pairs?

Where to from here? MTE Primer? Book with papers, and commentary fitting them into the right context (including the taxonomy?) Exercises the reader can do Showing how to fit new ideas into taxonomy Contribution to the language community in general MT Summit Workshop Other workshops?

BIG HUG

Web Sites Classification (taxonomy) Open to everyone Working area Possibly password protected Rite of initiation? Data for people - at least until next workshop URL’s