Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #12 Intelligent Digital Forensics September 30, 2009.

Slides:



Advertisements
Similar presentations
SDMX in the Vietnam Ministry of Planning and Investment - A Data Model to Manage Metadata and Data ETV2 Component 5 – Facilitating better decision-making.
Advertisements

DIGIDOC A web based tool to Manage Documents. System Overview DigiDoc is a web-based customizable, integrated solution for Business Process Management.
Database System Concepts and Architecture
CAPTURE SOFTWARE Please take a few moments to review the following slides. Please take a few moments to review the following slides. The filing of documents.
Domain Engineering Arie van Deursen May Software Product Line  Group of products  sharing a common, managed set of features  to satisfy the.
Guide to Computer Forensics and Investigations, Second Edition
Chapter 12: Expert Systems Design Examples
Xyleme A Dynamic Warehouse for XML Data of the Web.
Guide to Computer Forensics and Investigations Fourth Edition
Database Management: Getting Data Together Chapter 14.
Steganography Part 2 – Detection and Research. Introduction to Steganalysis What is steganalysis?  The art of detecting messages hidden by steganography.
1 Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang, Assistant Professor Dept. of Computer Science & Information Engineering National Central.
Chapter 14 The Second Component: The Database.
COEN 252 Computer Forensics Forensic Duplication of Hard Drives.
Databases & Data Warehouses Chapter 3 Database Processing.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Review for Final Exam November 19, 2010.
MDC Open Information Model West Virginia University CS486 Presentation Feb 18, 2000 Lijian Liu (OIM:
By N.Gopinath AP/CSE. Why a Data Warehouse Application – Business Perspectives  There are several reasons why organizations consider Data Warehousing.
CLARIN tools for workflows Overview. Objective of this document  Determine which are the responsibilities of the different components of CLARIN workflows.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence From Data Mining To Knowledge.
A summary of the report written by W. Alink, R.A.F. Bhoedjang, P.A. Boncz, and A.P. de Vries.
Aurora: A Conceptual Model for Web-content Adaptation to Support the Universal Accessibility of Web-based Services Anita W. Huang, Neel Sundaresan Presented.
Katanosh Morovat.   This concept is a formal approach for identifying the rules that encapsulate the structure, constraint, and control of the operation.
Data Management Turban, Aronson, and Liang Decision Support Systems and Intelligent Systems, Seventh Edition.
Ihr Logo Chapter 5 Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization Turban, Aronson, and Liang.
1 The BT Digital Library A case study in intelligent content management Paul Warren
Chapter 9 Designing Databases Modern Systems Analysis and Design Sixth Edition Jeffrey A. Hoffer Joey F. George Joseph S. Valacich.
Of 33 lecture 10: ontology – evolution. of 33 ece 720, winter ‘122 ontology evolution introduction - ontologies enable knowledge to be made explicit and.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #8 Computer Forensics Data Recovery and Evidence Collection September.
PLoS ONE Application Journal Publishing System (JPS) First application built on Topaz application framework Web 2.0 –Uses a template engine to display.
Summary Marie Yarbrough. Introduction History of Image Forgery Method Segmentation Classification Common-Sense Reasoning Conclusion.
 Three-Schema Architecture Three-Schema Architecture  Internal Level Internal Level  Conceptual Level Conceptual Level  External Level External Level.
Ihr Logo Chapter 5 Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization Turban, Aronson, and Liang.
Knowledge Representation of Statistic Domain For CBR Application Supervisor : Dr. Aslina Saad Dr. Mashitoh Hashim PM Dr. Nor Hasbiah Ubaidullah.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #4 Data Acquisition September 8, 2008.
Guest Lecture Introduction to Data Mining Dr. Bhavani Thuraisingham September 17, 2010.
Computer Systems & Architecture Lesson 4 8. Reconstructing Software Architectures.
1 Technology in Action Chapter 11 Behind the Scenes: Databases and Information Systems Copyright © 2010 Pearson Education, Inc. Publishing as Prentice.
A radiologist analyzes an X-ray image, and writes his observations on papers  Image Tagging improves the quality, consistency.  Usefulness of the data.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture 23 Intelligent Digital Forensics October 22, 2007.
XML and Database.
Using and modifying plan constraints in Constable Jim Blythe and Yolanda Gil Temple project USC Information Sciences Institute
Data Structures and Algorithms Dr. Tehseen Zia Assistant Professor Dept. Computer Science and IT University of Sargodha Lecture 1.
Data and Applications Security Developments and Directions Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #15 Secure Multimedia Data.
Selective and Intelligent Imaging Using Digital Evidence Bags.
INTRODUCTION TO GIS  Used to describe computer facilities which are used to handle data referenced to the spatial domain.  Has the ability to inter-
Using automation to enhance the process of Digital Forensic analysis Daniel Walton School of Computer and Information Science
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
44220: Database Design & Implementation Introduction to Module Ian Perry Room: C49 Ext.: 7287
Detecting False Captioning Using Common Sense Reasoning James Byrd.
Introduction to Biometrics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #6 Guest Lecture + Some Topics in Biometrics September 12,
COEN 252 Computer Forensics Forensic Duplication of Hard Drives.
Forensic Investigation Techniques Michael Jones. Overview Purpose People Processes Michael Jones2Digital Forensic Investigations.
Copyright © 2016 Pearson Education, Inc. Modern Database Management 12 th Edition Jeff Hoffer, Ramesh Venkataraman, Heikki Topi CHAPTER 11: BIG DATA AND.
Copyright © 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 28 Data Mining Concepts.
Introduction: Databases and Database Systems Lecture # 1 June 19,2012 National University of Computer and Emerging Sciences.
Databases (CS507) CHAPTER 2.
Visual Information Retrieval
Fundamentals of Information Systems, Sixth Edition
INTRODUCTION TO GEOGRAPHICAL INFORMATION SYSTEM
Introduction Multimedia initial focus
Data and Applications Security Developments and Directions
Fundamentals of Information Systems, Sixth Edition
Digital Forensics Dr. Bhavani Thuraisingham
Digital Forensics 2 Lecture 2: Understanding steganography in graphic files Presented by : J.Silaa Lecture: FCI Based on Guide to Computer Forensics and.
Copyright © 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 2 Database System Concepts and Architecture.
Data Warehousing and Data Mining
ece 627 intelligent web: ontology and beyond
Dr. Bhavani Thuraisingham The University of Texas at Dallas
Presentation transcript:

Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #12 Intelligent Digital Forensics September 30, 2009

Reading for Lecture #12 l l XIRAF – XML-based indexing and querying for digital forensics l Selective and intelligent imaging using digital evidence bags l l Detecting false captioning using common-sense reasoning

Abstract of Paper 1 l This paper describes a novel, XML-based approach towards managing and querying forensic traces extracted from digital evidence. This approach has been implemented in XIRAF, a prototype system for forensic analysis. XIRAF systematically applies forensic analysis tools to evidence files (e.g., hard disk images). Each tool produces structured XML annotations that can refer to regions (byte ranges) in an evidence file. XIRAF stores such annotations in an XML database, which allows us to query the annotations using a single, powerful query language (XQuery). XIRAF provides the forensic investigator with a rich query environment in which browsing, searching, and predefined query templates are all expressed in terms of XML database queries

Introduction l Framework for forensic analysis called XIRAF l A clean separation between feature extraction and analysis - Features extracted are stored in XML format l A single, XML-based output format for forensic analysis tools l The use of XML database technology for storing and querying the XML output of analysis tools.

XIRAF Framework l Consists of three components l Feature extraction manager - Features are extracted from BLOBs (Binary large objects) using feature extraction tools - Output of the tools are coded in XML for the forensics analyzer l Tool repository - Tools are wrapped (e.g., object wrappers) l Storage subsysystem - Stores BLOBs and XML annotations - XQuery used to query XML data

Forensic Applications l Authors have implemented following applications - Timeline browser: Through web browser examiner can look at data/time of interest - Photo search l Search for images satisfying certain conditions - Child pornography detection l Using hashing carried out matching

Summary and Directions l The separation of feature extraction and analysis brings benefits to both phases. XIRAF extracts features automatically, which is essential when processing large input sets. l The use of XML as a common, intermediate output format for tools allows the integration of the output of diverse, independent tools that produce similar information. This handles both the heterogeneity present in the input data (e.g., different browser types) and with the diversity of forensic analysis tools. l These benefits are demonstrated both by the timeline browser and by child pornography detection program. l By storing extracted features in an XML database system one can analyze those features using a single, general-purpose, powerful query language. In addition, we benefit automatically from advances that are made in the area of XML database systems l Directions: Use semantic web technologies?

Abstract of Paper 2 l This paper defines what selective imaging is, and the types of selective imaging that can be performed. This is contrasted with intelligent imaging and the additional capabilities that have to be built into an imager for it to be ‘intelligent’. A selective information capture scenario is demonstrated using the digital evidence bag (DEB) storage format. A DEB is a universal container for digital evidence from any source that allows the provenance to be recorded and continuity to be maintained throughout the life of the investigation. The paper concludes by defining the ‘ultimate test’ for an intelligent and selective imager

Selective Imaging l Selective imaging is a term that is generally associated with the decision not to acquire all the possible information during the capture process. l It is now recognized that ‘partial or selective file copying may be considered as an alternative’ when it may not be practical to acquire everything. l Techniques include manual selection, semi-automatic selection, automatic selection

Intelligent Imaging l Include the domain experts in the imaging process l How do you go about capturing the knowledge of the technical experts that are familiar with digital technical complexities and legal domain experts and combine them? l How do you know that you have captured everything relevant to the case under investigation or have not missed evidence of other offences?

Digital Evidence Bags l Both selective and intelligent imaging techniques offer many more options and capabilities than current bit stream imaging. l There are currently no commercial tools that perform selective imaging and adequately record the provenance of the selected information. l Furthermore, no method has existed that captured the criteria or method used by the examiner in deciding what to acquire. For example, was an arbitrary manual selection used or was information captured based on category of information, file extensions, file signature or hash set. l Authors solution to these problems is by the use of the digital evidence bag (DEB) format. A DEB is a universal container for digital information from any source. It allows the provenance of digital information to be recorded and continuity to be maintained throughout the life of the exhibit. l Additionally, DEBs may be encapsulated within other DEBs. This feature differentiates the DEB structure from that used by current monolithic formats commonly in use.

“The Ultimate Test” l The method and storage container used must be able to store sufficient information about the provenance of the information captured such that when the information is restored it is identical to that which would have been acquired should a bit stream image have been taken.

Summary and Directions l The methodology described and demonstrated by the authors is claimed to be a big improvement over bit stream imaging methods currently used. l Directions - Better selection methods, more accurate?

Abstract of Paper 3 l Detecting manipulated images has become an important problem in many domains (including medical imaging, forensics, journalism and scientific publication) largely due to the recent success of image synthesis techniques and the accessibility of image editing software. Many previous signal-processing techniques are concerned about finding forgery through simple transformation (e.g. resizing, rotating, or scaling), yet little attention is given to examining the semantic content of an image, which is the main issue in recent image forgeries. Here, the authors present a complete workflow for finding the anomalies within images by combining the methods known in computer graphics and artificial intelligence. They first find perceptually meaningful regions using an image segmentation technique and classify these regions based on image statistics. We then use AI common-sense reasoning techniques to find ambiguities and anomalies within an image as well as perform reasoning across a corpus of images to identify a semantically based candidate list of potential fraudulent images. Their method introduces a novel framework for forensic reasoning, which allows detection of image tampering, even with nearly flawless mathematical techniques.

Introduction l Detecting manipulated images has become an important problem in many domains l Many previous signal-processing techniques are concerned about finding forgery through simple transformation (e.g. resizing, rotating, or scaling), l Need to examine the semantic content of an image l Authors present a complete workflow for finding the anomalies within images by combining the methods known in computer graphics and artificial intelligence

Introduction l In Photo fakery, photo manipulation techniques may fall into four categories: l Deletion of details: removing scene elements l Insertion of details: adding scene elements l Photomontage: combining multiple images l False captioning: misrepresenting image content

Technical Approach l Authors find perceptually meaningful regions using an image segmentation technique and classify these regions based on image statistics. l They then use AI common-sense reasoning techniques to find ambiguities and anomalies within an image as well as perform reasoning across a corpus of images to identify a semantically based candidate list of potential fraudulent images. l They claim their method introduces a novel framework for forensic reasoning, which allows detection of image tampering, even with nearly flawless mathematical techniques.

Technical Approach l Image Segmentation - Segment the source into regions of importance - Compare across images in a corpus l Classification - Segment based classification l Common sense reasoning - Handles classification ambiguities

Summary and Directions l Introduces a hybrid method for image forensics. - Given a subset of a corpus as a suspicious candidate set, analyze the candidates through specific metrics that are optimized to find fakery given the image’s qualitative classification. This use of common-sense reasoning goes l Directions - To integrate the facts discovered in a photo corpus to help identify what evidence may be missing as well as what fact might be unique to this scenario.

Lecture 13 - Locating and recovering graphic files l Reference: Chapter 9 am 10 of Textbook - Steganography and Steganalysis l h/2004_03_research01.htm h/2004_03_research01.htm