Presentation is loading. Please wait.

Presentation is loading. Please wait.

Evaluation. formative 4 There are many times throughout the lifecycle of a software development that a designer needs answers to questions that check.

Similar presentations


Presentation on theme: "Evaluation. formative 4 There are many times throughout the lifecycle of a software development that a designer needs answers to questions that check."— Presentation transcript:

1 Evaluation

2 formative 4 There are many times throughout the lifecycle of a software development that a designer needs answers to questions that check whether his or her ideas match with those of the user(s).Such evaluation is known as formative evaluation because it (hopefully) helps shape the product. User-centred design places a premium on formative evaluation methods. 4 Summative 4 Summative evaluation, in contrast, takes place after the product has been developed.

3 Context of Formative Evaluation usability specific particulardefinite 4 Evaluation is concerned with gathering data about the usability of a design or product by a specific group of users for a particular activity within a definite environment or work context. 4 Regardless of the type of evaluation it is important to consider users –characteristics of the users activities –types of activities they will carry out –environment –environment of the study (controlled laboratory? field study?) –nature –nature of the artefact or system being evaluated? (sketches? prototype? full system?)

4 Reasons for Evaluation 4 Understanding 4 Understanding the real world –particularly important during requirements gathering 4 Comparing 4 Comparing designs –rarely are there options without alternatives –valuable throughout the development process 4 Engineering 4 Engineering towards a target –often expressed in the form of a metric 4 Checking conformance 4 Checking conformance to a standard

5 Classification of Evaluation Methods 4 Observation and Monitoring –data collection by note-taking, keyboard logging, video capture 4 Experimentation and Benchmarking –statement of hypothesis, control of variables 4 Collecting users’ opinions –surveys, questionnaires, interviews 4 Interpreting situated events 4 Predicting usability

6 Observation and Monitoring - Direct Observation Protocol 4 Usually informal in field study, more formal in controlled laboratories 4 data collection by direct observation and note- taking –users in “natural” surroundings –“objectivity” may be compromised by point of view of observer –users may behave differently while being watched (Hawthorne Effect) –ethnographic, participatory approach is an alternative

7 Observation and Monitoring - Indirect Observation Protocol 4 data collection by remote note taking, keyboard logging, video capture briefedpolicy justifiedprioritised –Users need to be briefed fully; a policy must be decided upon and agreed about what to do if they get “stuck”; tasks must be justified and prioritised (easiest first) –Video capture permits post-event “debriefing” and avoids Hawthorne effect (However, users may behave differently in unnatural environment) –with data-logging vast amounts of low-level data collected; difficult and expensive to analyse –interaction of variables may be more relevant than a single one (lack of context)

8 Experimentation and Benchmarking 4 “Scientific” and “engineering” approach 4 utilises standard scientific investigation techniques 4 Selection of benchmarking criteria is critical…and sometimes difficult (e.g., for OODBMS) 4 Control of variables, esp. user groups, may lead to “artificial” experimental bases

9 Collecting User’s Opinions 4 Surveys –critical mass and breadth of survey are critical for statistical reliability –Sampling techniques need to be well-grounded in theory and practice –Questions must be consistently formulated, clear and not “lead” to specific answers

10 Collecting User’s Opinions - Verbal Protocol 4 (Individual) Interviews –can be during or after user interaction during: immediate impressions are recorded during: may be distracting during complex tasks after: no distraction from task at hand after: may lead to misleading results (short-term memory loss, “history rewritten” etc.) –can be “structured” or not a structured interview is like a personal questionnaire - prepared questions

11 Collecting Users Opinions 4 Questionnaires –“open” (free form reply) or “closed” (answers “yes/no” or from a wider range of possible answers) latter is better for quantitative analysis –important to use clear, comprehensive and unambiguous terminology, quantified where possible e.g., daily?, weekly?, monthly? Rather than “seldom”, “often ” and there should always be a “never” –Needs to allow for “negative” feedback –All Form Fillin guidelines apply!

12 Relationship between Types of Evaluation and Reasons for Evaluation Observing and Monitoring Users’ Opinions Experiments etc. InterpretivePredictive Understanding Real World Comparing Designs Engineering to target Standards conformance Y Y Y Y Y Y Y Y Y Y Y Y Y


Download ppt "Evaluation. formative 4 There are many times throughout the lifecycle of a software development that a designer needs answers to questions that check."

Similar presentations


Ads by Google