November 8, 2005NSF Expedition Workshop Supporting E-Discovery with Search Technology Douglas W. Oard College of Information Studies and Institute for.

Slides:



Advertisements
Similar presentations
Agnes Adjabeng Librarian/Principal Programme Officer, EPA/GHANA Mainstreaming eLearning for Environment UNEP Pre-Conference Seminar at 4 th eLearning Africa,
Advertisements

GMD German National Research Center for Information Technology Darmstadt University of Technology Perspectives and Priorities for Digital Libraries Research.
Strategic decision making with exploratory search Toby Mostyn CTO Polecat.
Ranked Retrieval INST 734 Module 3 Doug Oard. Agenda  Ranked retrieval Similarity-based ranking Probability-based ranking.
November 17, 2005TREC 2005 TREC-2006 Legal Track Planning Session Jason Baron Dave Lewis Doug Oard.
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
Technology-Assisted Review Can be More Effective and More Efficient Than Exhaustive Manual Review Gordon V. Cormack University of Waterloo
Search Engines and Information Retrieval
Information Retrieval Review
Modern Information Retrieval
MALACH Multilingual Access to Large spoken ArCHives Survivors of the Shoah Visual History Foundation Human Language Technologies IBM T. J. Watson Research.
Information Retrieval in Practice
Information Retrieval Interaction CMSC 838S Douglas W. Oard April 27, 2006.
Modern Information Retrieval Chapter 5 Query Operations.
Evaluating the Performance of IR Sytems
Information Access Douglas W. Oard College of Information Studies and Institute for Advanced Computer Studies Design Understanding.
Lessons Learned from Information Retrieval Chris Buckley Sabir Research
December 2, 2004TDT-2004 Adaptive Topic Tracking at Maryland Tamer Elsayed, Douglas W. Oard, David Doermann University of Maryland, College Park Gary Kuhn.
Data-PASS/NDIIPP: A new effort to harvest our history IASSIST/IFDO 2005 May, 25, 2005.
ISP 433/633 Week 6 IR Evaluation. Why Evaluate? Determine if the system is desirable Make comparative assessments.
ICAIL 2007 DESI Workshop Panel presentation Marie-Francine Moens Centre for Law and ICT/ Department of Computer Science Katholieke Universiteit Leuven,
LIS618 lecture 11 i/r performance evaluation Thomas Krichel
Information Technology & Computer Science E-Discovery Lab Identification and Collection Seminar on E-Discovery, February 9th, 2012, College of Information.
European Public Sector Information Systems Conference -- September 30, 1998 Case Study: Building the Skills that Produce Success - A Case Study from the.
The Evolution of Shared-Task Evaluation Douglas W. Oard College of Information Studies and UMIACS University of Maryland, College Park, USA December 4,
A National Resource Working in the Public Interest © 2006 The MITRE Corporation. All rights reserved. KM at MITRE Jean Tatalias KM TEM, December 2007.
August 21, 2002Szechenyi National Library Support for Multilingual Information Access Douglas W. Oard College of Information Studies and Institute for.
Search Engines and Information Retrieval Chapter 1.
Text REtrieval Conference (TREC) The TREC Conferences Ellen Voorhees.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
CS523 INFORMATION RETRIEVAL COURSE INTRODUCTION YÜCEL SAYGIN SABANCI UNIVERSITY.
Human Resource Management Lecture 27 MGT 350. Last Lecture What is change. why do we require change. You have to be comfortable with the change before.
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
U.S. Department of the Interior U.S. Geological Survey CDI Webinar Sept. 5, 2012 Kevin T. Gallagher and Linda C. Gundersen September 5, 2012 CDI Science.
Evaluation INST 734 Module 5 Doug Oard. Agenda Evaluation fundamentals  Test collections: evaluating sets Test collections: evaluating rankings Interleaving.
1 Information Retrieval Acknowledgements: Dr Mounia Lalmas (QMW) Dr Joemon Jose (Glasgow)
Evaluation INST 734 Module 5 Doug Oard. Agenda  Evaluation fundamentals Test collections: evaluating sets Test collections: evaluating rankings Interleaving.
P RINCIPLES 1-7 FOR E LECTRONIC D OCUMENT P RODUCTION Maryanne Post.
Policy and IT Security Awareness Amy Ginther Policy Develoment Coordinator University of Maryland Information Technology Security Workshop April 2, 2004.
The Challenge of Rule 26(f) Magistrate Judge Craig B. Shaffer July 15, 2011.
Unidata Support Tools and Plans Unidata Policy and Users Committee May 2005.
IR System Evaluation Farhad Oroumchian. IR System Evaluation System-centered strategy –Given documents, queries, and relevance judgments –Try several.
Search Engine Architecture
IR Theory: Relevance Feedback. Relevance Feedback: Example  Initial Results Search Engine2.
Evaluation INST 734 Module 5 Doug Oard. Agenda Evaluation fundamentals Test collections: evaluating sets  Test collections: evaluating rankings Interleaving.
WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries.
The Structure of Information Retrieval Systems LBSC 708A/CMSC 838L Douglas W. Oard and Philip Resnik Session 1: September 4, 2001.
Information in the Digital Environment Information Seeking Models Dr. Dania Bilal IS 530 Spring 2005.
Basics of Information Retrieval and Query Formulation Bekele Negeri Duresa Nuclear Information Specialist.
Planning for the TREC 2008 Legal Track Douglas Oard Stephen Tomlinson Jason Baron.
Performance Measurement. 2 Testing Environment.
Information Retrieval
Search and Access Technologies for Large Scale Web Archives Joseph JaJa, Sangchul Song, and Mike Smorul Institute for Advanced Computer Studies Department.
Enterprise Archiving, Retention and Discovery System Jim Albert Deputy Director Department of Information Services April 19 th 2007.
Knowledge building in the 21 st century at The Geelong College: Information-to-Knowledge Continuum “As we increasingly move toward an environment of instant.
Evaluation INST 734 Module 5 Doug Oard. Agenda Evaluation fundamentals Test collections: evaluating sets Test collections: evaluating rankings Interleaving.
Augmenting (personal) IR Readings Review Evaluation Papers returned & discussed Papers and Projects checkin time.
Evaluation of Information Retrieval Systems Xiangming Mu.
Complex legal requirements in electronic discovery search and retrieval: A NARA case study NSF Collaborative Expedition Workshop # 45 Advancing Information.
A Framework for Assessing Needs Across Multiple States, Stakeholders, and Topic Areas Stephanie Wilkerson & Mary Styers REL Appalachia American Evaluation.
Cross-Language Information Retrieval Applied Natural Language Processing October 29, 2009 Douglas W. Oard.
Enterprise Track: Thread-based Retrieval Enterprise Track: Thread-based Retrieval Yejun Wu and Douglas W. Oard Goal Explore -- document expansion.
Loyola University Maryland ImageNow Users Group February 23, 2015.
Search Engine Architecture
IR Theory: Evaluation Methods
Proposed Formative Evaluation Adaptive Topic Tracking Systems
Evaluation of IR Performance
Search Engine Architecture
Retrieval Evaluation - Reference Collections
Structure of IR Systems
Presentation transcript:

November 8, 2005NSF Expedition Workshop Supporting E-Discovery with Search Technology Douglas W. Oard College of Information Studies and Institute for Advanced Computer Studies University of Maryland, College Park

Outline Search process Evaluation Research agenda

The E-Discovery Context Classic “Information Retrieval” –Goal: satisfy a visceral information need –Understanding of need evolves during search –Personal view of relevance E-Discovery –Goal: identify a set of responsive documents –Negotiated information need (with iteration) –Agreed / defensible / explainable process

Supporting Personal Searching Source Selection Search Query Selection Ranked List Examination Document Delivery Document Query Formulation IR System Query Reformulation and Relevance Feedback Source Reselection NominateChoosePredict

Current E-Discovery Support Search Query Result Set Review Delivery Responsive Documents Query Formulation Indexing Index Acquisition Collection

Future E-Discovery Support Source Selection Search Query Selection Ranked List Incremental Review Result Set Delivery Responsive Documents Query Formulation IR System Indexing Index Acquisition Collection

Incremental Ranked Review Limit to:Marlboro NOT “Upper Marlboro” Rank by:tobacco, {policy staff} Limit to:Marlboro NOT “Upper Marlboro” Rank by:regulation, manipulation

Evaluation Design Issues Real information needs are easily captured –Byproduct of negotiation process Repeatable process models lack some fidelity –Tune system to query (not query to system) –Pooling reveals relative (not absolute) recall –Solution: augment lab studies with real cases Confidentiality / privilege limit access to data –Corporate collections are incomplete –Public collections are less representative

TREC-2006 Legal Track Collection options –Tobacco settlement (scanned/OCR) –State department cables –Enron ( ) Schedule –Nov 05Planning conference –Mar 06Guidelines (collection, topics, …) –Jul 06Experiments –Sep 06Ground truth judgments –Nov 06Report Results

Research at Maryland Center for Information Policy –Best practices: records management NDIIPP –Law firm records –Policy: retention, access Joint Institute for Knowledge Discovery –Enron , phone calls, and databases –Linking: text-metadata, cross-source MALACH –Oral history recordings –Search: conversational speech

Looking Forward Continue this dialog –Technology drives process innovation –Process requirements drive technical innovation Craft a balanced investment strategy –Use TREC to explore system design Create reusable test collections Foster development of a research community –User studies –Process standards