The CLEF 2003 cross language image retrieval task Paul Clough and Mark Sanderson University of Sheffield

Slides:



Advertisements
Similar presentations
SINAI-GIR A Multilingual Geographical IR System University of Jaén (Spain) José Manuel Perea Ortega CLEF 2008, 18 September, Aarhus (Denmark) Computer.
Advertisements

Bogdan Vrusias © 2003 Scene of Crime Information System: Playing at St. Andrews 22nd August 2003 Bogdan Vrusias, Mariam Tariq, Lee Gillam Department of.
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
Sheffield at ImageCLEF 2003 Paul Clough and Mark Sanderson.
Search Results Need to be Diverse Mark Sanderson University of Sheffield.
A Maximum Coherence Model for Dictionary-based Cross-language Information Retrieval Yi Liu, Rong Jin, Joyce Y. Chai Dept. of Computer Science and Engineering.
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
The Challenges of Multilingual Search Paul Clough The Information School University of Sheffield ISKO UK conference 8-9 July 2013.
Multilingual experiments of CLEF 2003 Eija Airio, Heikki Keskustalo, Turid Hedlund, Ari Pirkola University of Tampere, Finland Department of Information.
The XLDB Group at GeoCLEF 2005 Nuno Cardoso, Bruno Martins, Marcírio Chaves, Leonardo Andrade, Mário J. Silva
Modern Information Retrieval
Information Access I Measurement and Evaluation GSLT, Göteborg, October 2003 Barbara Gawronska, Högskolan i Skövde.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
 Ad-hoc - This track tests mono- and cross- language text retrieval. Tasks in 2009 will test both CL and IR aspects.
 Official Site: facility.org/research/evaluation/clef-ip-10http:// facility.org/research/evaluation/clef-ip-10.
LREC Combining Multiple Models for Speech Information Retrieval Muath Alzghool and Diana Inkpen University of Ottawa Canada.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Evaluating Cross-language Information Retrieval Systems Carol Peters IEI-CNR.
August 21, 2002Szechenyi National Library Support for Multilingual Information Access Douglas W. Oard College of Information Studies and Institute for.
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( Bridging Languages for Question Answering: DIOGENE at CLEF-2003.
CLEF Ǻrhus Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Oier Lopez de Lacalle, Arantxa Otegi, German Rigau UVA & Irion: Piek Vossen.
Impressions of 10 years of CLEF Donna Harman Scientist Emeritus National Institute of Standards and Technology.
1 Intra- and interdisciplinary cross- concordances for information retrieval Philipp Mayr GESIS – Leibniz Institute for the Social Sciences, Bonn, Germany.
1 The Domain-Specific Track at CLEF 2008 Vivien Petras & Stefan Baerisch GESIS Social Science Information Centre, Bonn, Germany Aarhus, Denmark, September.
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( The Multiple Language Question Answering Track at CLEF 2003.
Philosophy of IR Evaluation Ellen Voorhees. NIST Evaluation: How well does system meet information need? System evaluation: how good are document rankings?
JULIO GONZALO, VÍCTOR PEINADO, PAUL CLOUGH & JUSSI KARLGREN CLEF 2009, CORFU iCLEF 2009 overview tags : image_search, multilinguality, interactivity, log_analysis,
PATENTSCOPE Patent Search Strategies and Techniques Andrew Czajkowski Head, Innovation and Technology Support Section Centurion September 11, 2014.
1 Cross-Lingual Query Suggestion Using Query Logs of Different Languages SIGIR 07.
CLEF 2004 – Interactive Xling Bookmarking, thesaurus, and cooperation in bilingual Q & A Jussi Karlgren – Preben Hansen –
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
“ SINAI at CLEF 2005 : The evolution of the CLEF2003 system.” Fernando Martínez-Santiago Miguel Ángel García-Cumbreras University of Jaén.
Cross-Language Evaluation Forum CLEF Workshop 2004 Carol Peters ISTI-CNR, Pisa, Italy.
The PATENTSCOPE search system: CLIR February 2013 Sandrine Ammann Marketing & Communications Officer.
Image Retrieval in Radiology: The ImageCLEF Challenge Charles E. Kahn, Jr. Medical College of Wisconsin Jayashree Kalpathy-Cramer Oregon Health & Science.
Cross-Language Evaluation Forum (CLEF) IST Expected Kick-off Date: August 2001 Carol Peters IEI-CNR, Pisa, Italy Carol Peters: blabla Carol.
MIRACLE Multilingual Information RetrievAl for the CLEF campaign DAEDALUS – Data, Decisions and Language, S.A. Universidad Carlos III de.
Péter Schönhofen – Ad Hoc Hungarian → English – CLEF Workshop 20 Sep 2007 Performing Cross-Language Retrieval with Wikipedia Participation report for Ad.
Multilingual Relevant Sentence Detection Using Reference Corpus Ming-Hung Hsu, Ming-Feng Tsai, Hsin-Hsi Chen Department of CSIE National Taiwan University.
IIIT Hyderabad’s CLIR experiments for FIRE-2008 Sethuramalingam S & Vasudeva Varma IIIT Hyderabad, India 1.
UA in ImageCLEF 2005 Maximiliano Saiz Noeda. Index System  Indexing  Retrieval Image category classification  Building  Use Experiments and results.
Ben Carterette Paul Clough Evangelos Kanoulas Mark Sanderson.
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
How robust is CLIR? Proposal for a new robust task at CLEF Thomas Mandl Information Science Universität Hildesheim 6 th Workshop.
Iterative Translation Disambiguation for Cross Language Information Retrieval Christof Monz and Bonnie J. Dorr Institute for Advanced Computer Studies.
CLEF 2007 Workshop Budapest, Hungary, 19–21 September 2007 Nicola Ferro Information Management Systems (IMS) Research Group Department of Information Engineering.
The CLEF 2005 Cross-Language Image Retrieval Track Organised by Paul Clough, Henning Müller, Thomas Deselaers, Michael Grubinger, Thomas Lehmann, Jeffery.
CLEF Kerkyra Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Arantxa Otegi UNIPD: Giorgio Di Nunzio UH: Thomas Mandl.
Cross-Language Evaluation Forum CLEF 2003 Carol Peters ISTI-CNR, Pisa, Italy Martin Braschler Eurospider Information Technology AG.
From Text to Image: Generating Visual Query for Image Retrieval Wen-Cheng Lin, Yih-Chen Chang and Hsin-Hsi Chen Department of Computer Science and Information.
Thomas Mandl: GeoCLEF Track Overview Cross-Language Evaluation Forum (CLEF) Thomas Mandl, (U. Hildesheim) 8 th Workshop.
Medical retrieval 2008 New data set with almost 66,000 images Thirty topics were made available, ten in each of three categories: visual, mixed, and semantic.
Stiftung Wissenschaft und Politik German Institute for International and Security Affairs CLEF 2005: Domain-Specific Track Overview Michael Kluck SWP,
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
What’s happening in iCLEF? (the iCLEF Flickr Challenge) Julio Gonzalo (UNED), Paul Clough (U. Sheffield), Jussi Karlgren (SICS), Javier Artiles (UNED),
The Cross Language Image Retrieval Track: ImageCLEF Breakout session discussion.
The Loquacious ( 愛說話 ) User: A Document-Independent Source of Terms for Query Expansion Diane Kelly et al. University of North Carolina at Chapel Hill.
Combining Text and Image Queries at ImageCLEF2005: A Corpus-Based Relevance-Feedback Approach Yih-Cheng Chang Department of Computer Science and Information.
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
The CLEF 2005 interactive track (iCLEF) Julio Gonzalo 1, Paul Clough 2 and Alessandro Vallin Departamento de Lenguajes y Sistemas Informáticos, Universidad.
1 The Domain-Specific Track at CLEF 2007 Vivien Petras, Stefan Baerisch & Max Stempfhuber GESIS Social Science Information Centre, Bonn, Germany Budapest,
Analysis of Experiments on Hybridization of different approaches in mono and cross-language information retrieval DAEDALUS – Data, Decisions and Language,
Manual Query Modification and Automated Translation to Improve Cross-language Medical Image Retrieval Jeffery R. Jensen William R. Hersh Department of.
CLEF Workshop ECDL 2003 Trondheim Michael Kluck slide 1 Introduction to the Monolingual and Domain-Specific Tasks of the Cross-language.
CLEF Budapest1 Measuring the contribution of Word Sense Disambiguation for QA Proposers: UBC: Agirre, Lopez de Lacalle, Otegi, Rigau, FBK: Magnini.
F. López-Ostenero, V. Peinado, V. Sama & F. Verdejo
Experiments for the CL-SR task at CLEF 2006
CLIR PATENTSCOPE search system
Dictionaries First Grade Unit 3 INTRO
Cheshire at GeoCLEF 2008: Text and Fusion Approaches for GIR
Presentation transcript:

The CLEF 2003 cross language image retrieval task Paul Clough and Mark Sanderson University of Sheffield

Introduction A pilot experiment in CLEF 2003 Called ImageCLEF Combination of image retrieval and CLIR An ad hoc retrieval task 4 entries –NTU (Taiwan) –Daedalus (Spain) –Surrey (UK) –Sheffield (UK)

Why a new CLEF task? No existing TREC-style test collection Broadens the CLEF range of CLIR tasks Facilitates CL image retrieval research International forum for discussion

The ad hoc task Given a user need expressed in a language different from the document collection, find as many relevant images as possible Fifty user needs (topics): –Expressed with a short (title) and longer (narrative) textual description –Also expressed with an example relevant image (QBE) –Titles translated into 5 European languages (by Sheffield) and Chinese (by NTU) Two retrieval challenges –Matching textual queries to visual documents (use captions) –Matching non-English queries to English captions (use translation) Essentially a bilingual CLIR task No retrieval constraints specified

Creating the test collection Pooled all submitted runs from entrants Work undertaken at Sheffield Publicly- available ImageCLEF resources Evaluated all runs using trec_eval CLEF results Selected & translated 50 topics Topics Title (translated) Description Example image Judged relevance of pools plus ISJ Relevance assessments 4 relevance sets created by 2 assessors per topic using a ternary relevance scheme based on image and caption Created an image collection Image collection + caption

Evaluation Evaluation based on most stringent relevance set (strict intersection) Compared systems using –MAP across all topics –Number of topics with no relevant image in the top participants evaluated (used captions only): –NTU – Chinese->English, manual and automatic, Okapi and dictionary- based translation, focus on proper name translation –Daedalus – all->English (except Dutch and Chinese), Xapian and dictionary-based + on-line translation, Wordnet query expansion, focus on indexing query and ways of combining query terms –Surrey – all->English (except Chinese), SoCIS system and on-line translation, Wordnet expansion, focus on query expansion and analysis of topics –Sheffield – all->English, GLASS (BM25) and Systran translation, no language-specific processing, focus on translation quality

Results Surrey had problems NTU obtained highest Chinese results –approx. 51% mono and 12 failed topics (NTUiaCoP) Sheffield obtained highest –Italian: 72% mono and 7 failed topics –German: 75% mono and 8 failed topics –Dutch: 69% mono and 7 failed topics –French: 78% mono and 3 failed topics Daedalus obtained highest –Spanish: 76% mono and 5 failed topics (QTdoc) –Monolingual: and 1 failed topic (Qor) For more information … see the ImageCLEF working notes

Summary – what next? ImageCLEF 2004 … a full CLEF task Image collection –Use a larger and more general image collection –Translate the image captions – multilingual Ad hoc task –Refine the topic set –Translate the entire topic statement –Deal with difficulties expressed by assessors in relevance assessments for CL image retrieval –Define more clearly the topic narrative –Include more examples of relevant images Introduce an interactive CL image retrieval task

Finally… I need you …. … to express your interest in ImageCLEF 2004 … NOW! me: