Presentation on theme: "Online Peer Evaluation System Team Green Apple Team Members Ada Tse Amber Bahl Tom Nichols Matt Anderson Faculty Mentor Prof. M Lutz Project Sponsor Richard."— Presentation transcript:
Online Peer Evaluation System Team Green Apple Team Members Ada Tse Amber Bahl Tom Nichols Matt Anderson Faculty Mentor Prof. M Lutz Project Sponsor Richard Fasse RIT Online Learning
Agenda Project Overview Current System Our Product Features Requirements Process Product Architecture and Design Software Process Risk Analysis and Mitigation Metrics Status & Future Goals Demo
Problem Statement The RIT Online Learning department is in need of an online peer evaluation system that can enhance the collaborative learning experience. Existing Tool: –Paper Alternative –Clipboard Survey System
Importance Group work is an important aspect in today's education system The average SE graduate does about 16 group projects
Current System: Clipboard Create, Deploy and Analyze –Does provide different views for analysis but more effective for analyzing surveys then Peer Evaluations. –Very Hard to identify problem groups Not integrated with myCourses Survey System Can’t deploy evaluations per group Hard to setup Reporting does not show group weaknesses No control over who takes the survey
Current System: Reporting View View: Percentage/ Graph
Peer Evaluation System Integrated with existing system –Login pass-through –Course and group data imported directly from myCourses Setup Workflow –Tailored for peer evaluations Question Templates –Reusable –Shared between instructors
Application Workflow Instructor Main -Create Eval Student main -Take Eval Instructor Main -Reporting 1. Create Evaluation 2. Take Evaluation 3. Analyze Results WOW!!
Instructor Main List of global and personal questions templates Evaluation status Evaluations listed per course
Solution: Create Evaluation Select Template Eval Setup Info
Process: Scrum What is Scrum? –Scrum is an iterative, incremental process for developing any product or managing any work. It produces a potentially shippable set of functionality at the end of every iteration (Sprint).
Scrum: Sprint Typical team size 2 to 4 members Delivers working software –Typically between 1-4 week iterations Cross-functional tasks per team member New work may be uncovered by the team during development
Our Methodology Flavor of Scrum Differences: –Upfront requirements –Postponed the Sprint one delivery date by 2 weeks Similarities: –The whole project was implemented in chunks (Sprints) depending on the requirements prioritization (Sprint Backlogs). –Team meetings
Risk New Technologies –.NET Integration with myCourses –XML Feeds –Testing LDAP Authentication Complexity of business requirements
Risk Mitigation: Task Planning New Technologies –Allocated tasks according to skill set –Team members started off with small/simple programs –Experienced team members educated the team
Risk Mitigation: Development LDAP & myCourses integration –Great help from the Online Learning Complex business requirements –Incremental development & comprehensive requirements gathering
Risk Mitigation Plan: Software Process Use of Scrum User Feedback (Allows for midcourse corrections) Increased Product Visibility Increased Progress Visibility –Sprint Planning Through many sprints the requirements were revised many times to ensure that clarity is achieved. Throughout every sprint, each decision will be evaluated to make sure that it aligns with the overall goals of the project.
Risk Mitigation: Tooling Subversion for revision control Google groups Trac provides web based management –View files and changesets Automated synchronization of project documents to web site Trac provided an integrated bug tracking system
Metrics Backlogs –Product –Sprint Number of tasks completed for a particular sprint (work effort distributed for each sprint) Number of bugs –By Feature –By Severity –Per Sprint Total effort (man hours) for all phases
Bugs Per Feature Total # of bugs: 53 Major: 22 Minor: 11 Trivial: 20
Current Status Progress Key FeaturesProgress Requirements ElicitationDONE Requirements Analysis (SRS)DONE High Level ArchitectureDONE Initial Setup (DB, Environment)DONE Requirements PrioritizationDONE Sprint 1 (5 th Week)DONE Sprint 2 (7 th Week)DONE Sprint 3 (9 th Week)DONE Integration TestingDONE Final Release05/19/2006
Future Enhancements More views for reporting –Currently our application supports 2 views High-level groups + students Team member + responders + questions Better support for answer type –Currently our application supports Text Type Radio Button
Reflections Great Team!!! –All team members were new to the group Appropriate Software Process Model Delays in Sprint 1 –Unknown Technologies –.NET 2.0