UT DALLAS Erik Jonsson School of Engineering & Computer Science FEARLESS engineering Integrity Policies Murat Kantarcioglu.

Slides:



Advertisements
Similar presentations
TOPIC CLARK-WILSON MODEL Ravi Sandhu.
Advertisements

Title ON FOUR DEFINITIONS OF DATA INTEGRITY Ravi Sandhu George Mason University FIVE.
CS691 – Chapter 6 of Matt Bishop
Dan Fleck CS 469: Security Engineering
CMSC 414 Computer (and Network) Security Lecture 12 Jonathan Katz.
Security Models and Architecture
Access Control Intro, DAC and MAC System Security.
Chapter 2.  CIA Model  Host Security VS Network Security  Least Privileges  Layered Security  Access Controls Prepared by Mohammed Saher2.
Hybrid Policies Overview Chinese Wall Model Clinical Information Systems Security Policy ORCON RBAC Introduction to Computer Security ©2004 Matt Bishop.
Chapter 6: Integrity Policies Overview Requirements Biba’s models Clark-Wilson model Introduction to Computer Security ©2004 Matt Bishop.
ITIS 3200: Introduction to Information Security and Privacy Dr. Weichao Wang.
June 1, 2004Computer Security: Art and Science © Matt Bishop Slide #6-1 Chapter 6: Integrity Policies Overview Requirements Biba’s models Lipner’s.
May 4, 2004ECS 235Slide #1 Biba Integrity Model Basis for all 3 models: Set of subjects S, objects O, integrity levels I, relation ≤  I  I holding when.
Verifiable Security Goals
Courtesy of Professors Chris Clifton & Matt Bishop INFSCI 2935: Introduction of Computer Security1 September 18, 2003 Introduction to Computer Security.
Courtesy of Professors Chris Clifton & Matt Bishop INFSCI 2935: Introduction of Computer Security1 October 7, 2004 Introduction to Computer Security Lecture.
1 Clark Wilson Implementation Shilpa Venkataramana.
1 Integrity Policies CSSE 490 Computer Security Mark Ardis, Rose-Hulman Institute March 22, 2004.
6/26/2015 6:12 PM Lecture 5: Integrity Models James Hook (Some materials from Bishop, copyright 2004) CS 591: Introduction to Computer Security.
Chapter 6: Integrity Policies Overview Requirements Biba’s models Clark-Wilson model Introduction to Computer Security ©2004 Matt Bishop.
CS526Topic 21: Integrity Models1 Information Security CS 526 Topic 21: Integrity Protection Models.
User Domain Policies.
November 1, 2004Introduction to Computer Security ©2004 Matt Bishop Slide #6-1 Chapter 6: Integrity Policies Overview Requirements Biba’s models Clark-Wilson.
Sicurezza Informatica Prof. Stefano Bistarelli
CMSC 414 Computer and Network Security Lecture 19 Jonathan Katz.
1 ISA 562 Internet Security Theory and Practice Integrity Policies Chapter 6 of Bishop ’ s book.
Mandatory Security Policies CS461/ECE422 Spring 2012.
Slide #6-1 Integrity Policies CS461/ECE422 – Computer Security I Fall 2009 Based on slides provided by Matt Bishop for use with Computer Security: Art.
3/16/2004Biba Model1 Biba Integrity Model Presented by: Nathan Balon Ishraq Thabet.
Computer Security 3e Dieter Gollmann
Week 8 - Wednesday.  What did we talk about last time?  Authentication  Challenge response  Biometrics  Started Bell-La Padula model.
Session 2 - Security Models and Architecture. 2 Overview Basic concepts The Models –Bell-LaPadula (BLP) –Biba –Clark-Wilson –Chinese Wall Systems Evaluation.
Week 8 - Monday.  What did we talk about last time?  Access control  Authentication.
Security Architecture and Design Chapter 4 Part 3 Pages 357 to 377.
Dr. Bhavani Thuraisingham Cyber Security Lecture for July 2, 2010 Security Architecture and Design.
ITIS 3200: Introduction to Information Security and Privacy Dr. Weichao Wang.
Next-generation databases Active databases: when a particular event occurs and given conditions are satisfied then some actions are executed. An active.
Chapter 5 Network Security
Chapter 6: Integrity Policies  Overview  Requirements  Biba’s models  Clark-Wilson model Introduction to Computer Security ©2004 Matt Bishop.
CS426Fall 2010/Lecture 251 Computer Security CS 426 Lecture 25 Integrity Protection: Biba, Clark Wilson, and Chinese Wall.
14.1/21 Part 5: protection and security Protection mechanisms control access to a system by limiting the types of file access permitted to users. In addition,
Trusted OS Design and Evaluation CS432 - Security in Computing Copyright © 2005, 2010 by Scott Orr and the Trustees of Indiana University.
12/4/20151 Computer Security Security models – an overview.
Chapter 5 – Designing Trusted Operating Systems
Design Principles and Common Security Related Programming Problems
A security policy defines what needs to be done. A security mechanism defines how to do it. All passwords must be updated on a regular basis and every.
A Comparison of Commercial and Military Computer Security Presenter: Ivy Jiang1 A Comparison of Commercial and Military Computer Security Policies Authors:
IS 2150 / TEL 2810 Introduction to Security
Slide #6-1 Chapter 6: Integrity Policies Overview Requirements Biba’s models Clark-Wilson model.
6/22/20161 Computer Security Integrity Policies. 6/22/20162 Integrity Policies Commercial requirement differ from military requirements: the emphasis.
CS526Topic 19: Integrity Models1 Information Security CS 526 Topic 19: Integrity Protection Models.
Lecture 2 Page 1 CS 236 Online Security Policies Security policies describe how a secure system should behave Policy says what should happen, not how you.
Chapter 7. Hybrid Policies
TOPIC: Web Security Models
Banking and Bookkeeping
Verifiable Security Goals
Chapter 6 Integrity Policies
Chapter 6: Integrity Policies
Chapter 5: Confidentiality Policies
Basic Security Theorem
Integrity Models and Hybrid Models
Dr. Bhavani Thuraisingham Cyber Security Lecture for July 2, 2010 Security Architecture and Design.
Chapter 6: Integrity Policies
Lecture 18: Mandatory Access Control
Integrity Policies Dr. Wayne Summers Department of Computer Science
Biba Integrity Model Basis for all 3 models:
Chapter 6: Integrity Policies
Computer Security Integrity Policies
Presentation transcript:

UT DALLAS Erik Jonsson School of Engineering & Computer Science FEARLESS engineering Integrity Policies Murat Kantarcioglu

FEARLESS engineering Requirements of Policies for Commercial Applications [Lipner 1982] 1.Users will not write their own programs, but will use existing production programs and databases. 2.Programmers will develop and test programs on a non-production system; if they need access to actual data, they will be given production data via a special process, but will use it on their development system. 3.A special process must be followed to install a program from the development system onto the production system. 4.The special process in requirement 3 must be controlled and audited. 5.The managers and auditors must have access to both the system state and the system logs that are generated. The emphasis of these requirements is on integrity.

FEARLESS engineering Requirements and Principles of Operation The requirements suggest several principles –Separation of duty. If two or more steps are required to perform a critical function, at least two different subjects should perform them. Moving an application from the development system to the production system is an example of critical function. –Separation of function. Different functions are executed on different sets of data. For example, developers do not develop new programs in the production environment. Also they do not process production data in the development environment. If they need data, depending on the sensitivity of data, sanitized versions of these data may be given to them. –Auditing. It is the process of analyzing systems to determine what actions took place and who performed them

FEARLESS engineering Biba Integrity Model The Biba model associates an integrity level with both objects and subjects These levels form the basis for expressing integrity policies that refer to the corruption of ‘clean’ high level entities by ‘dirty’ low level entities In the integrity lattice, information may only flow downwards

FEARLESS engineering Biba Integrity Model Set of subjects S, objects O, integrity levels I, relation ≤  I  I holding when the second dominates the first i: S  O  I gives integrity level of an object or of a subject r  S  O means s  S can read o  O w  S  O means s  S can write o  O x  S  S defines the ability of a subject to invoke (execute) another subject.

FEARLESS engineering Intuition for Integrity Levels The higher the level, the more confidence –That a program will execute correctly –That data is accurate and/or reliable Important point: integrity levels are not security levels Integrity labels are assigned and maintained separately, because the reasons behind the labels are different

FEARLESS engineering Biba’s Model Access Control Rules 1.s  S can read o  O iff i(s) ≤ i(o) 2. s  S can write to o  O iff i(o) ≤ i(s) 3. s 1  S can execute s 2  S iff i(s 2 ) ≤ i(s 1 ) No actual implementations in real products have been reported for the Biba model The problem of integrity of information requires articulated solutions

FEARLESS engineering Clark-Wilson Integrity Model This model is based on two important principles: –Separation of duties –Well-formed transactions – these transactions constrain the ways in which users can modify the data. The main idea is that a data item can be modified only by a given set of transactions that are certified to work with that data item Unlike the Bell-LaPadula security model, which relies on access mediation in the operating system kernel (or DBMS), Clark and Wilson’s approach relies on application-level controls

FEARLESS engineering Clark-Wilson Integrity Model Main points of the Clark-Wilson model 1.Subjects have to be identified and authenticated 2.Objects can be manipulated only by a restricted set of programs 3.Subjects can execute only a restricted set of programs 4.A proper audit log has to be maintained 5.The system has to be certified to work properly

FEARLESS engineering Clark-Wilson Integrity Model Integrity defined by a set of constraints –Data in a consistent or valid state when it satisfies these constraints Example: Bank –D today’s deposits, W withdrawals, YB yesterday’s balance, TB today’s balance –Integrity constraint: D + YB –W Well-formed transactions move system from one consistent state to another Issue: who examines, certifies transactions done correctly?

FEARLESS engineering Entities CDIs: constrained data items –Data subject to integrity controls UDIs: unconstrained data items –Data not subject to integrity controls IVPs: integrity verification procedures –Procedures that test that the CDIs conform to the integrity constraints TPs: transaction procedures –Procedures that take the system from one valid state to another

FEARLESS engineering Certification Rules 1 and 2 CR1When any IVP is run, it must ensure all CDIs are in a valid state CR2For some associated set of CDIs, a TP must transform those CDIs in a valid state into a (possibly different) valid state –CR2 defines as certified a relation that associates a set of CDIs with a particular TP –CR2 implies that a TP may corrupt a CDI if it is not certified to work with that CDI –Bank example: TP balance, CDIs accounts. Let C be a certified relation, then (balance, account 1 ), (balance, account 2 ),…., (balance, account n )  C

FEARLESS engineering Enforcement Rules 1 and 2 –CR2 implies that a TP may corrupt a CDI if it is not certified to work with that CDI –Example: the TP that invests money in the bank’s stock portfolio would corrupt account balances even if the TP were certified to work on the portfolio, because the actions of the TP make no sense on the bank account –This leads to the first enforcement rule –The second enforcement rule is motivated by the fact that not all users are allowed to use all the TPs; the model must thus also account for the person performing the TP

FEARLESS engineering Enforcement Rules 1 and 2 ER1The system must maintain the certified relations and must ensure that only TPs certified to run on a CDI manipulate that CDI. ER2The system must associate a user with each TP and set of CDIs. The TP may access those CDIs on behalf of the associated user. The TP cannot access that CDI on behalf of a user not associated with that TP and CDI. –System must maintain, enforce certified relation –System must also restrict access based on user ID (allowed relation)

FEARLESS engineering Enforcement Rule 2 – allowed relation The allowed relation A specifies which user execute which TP on which CDI Let U be the set of users in the system Let T be the set of TP’s in the system Let C be the set of CDI’s in the system A is defined as: { | u  U, tp  T, cdi_s  2 C }

FEARLESS engineering Users and Rules CR3The allowed relations must meet the requirements imposed by the principle of separation of duty. ER3The system must authenticate each user attempting to execute a TP –Type of authentication undefined, and depends on the instantiation –Authentication not required before use of the system, but is required before manipulation of CDIs (requires using TPs)

FEARLESS engineering Logging CR4All TPs must append enough information to reconstruct the operation to an append- only CDI. –This CDI is the log –Auditor needs to be able to determine what happened during reviews of transactions

FEARLESS engineering Handling Untrusted Input CR5Any TP that takes as input a UDI may perform only valid transformations, or no transformations, for all possible values of the UDI. The transformation either rejects the UDI or transforms it into a CDI. –In bank, numbers entered at keyboard are UDIs. TPs must validate numbers (to make them a CDI) before using them; if validation fails, TP rejects UDI –Therefore CR5 says that any TP that takes a UDI as input must either convert the UDI into a CDI or reject the UDI and perform no transformation at all

FEARLESS engineering Separation of Duty In Model ER4Only the certifier of a TP may change the list of entities associated with that TP. No certifier of a TP, or of an entity associated with that TP, may ever have execute permission with respect to that entity. –Enforces separation of duty with respect to certified and allowed relations

FEARLESS engineering Comparison With Requirements 1.Users cannot certify TPs, so CR5 and ER4 enforce this 2.Procedural, so model doesn’t directly cover it; but special process corresponds to using TP No technical controls can prevent programmer from developing program on production system; usual control is to delete software tools 3.TP does the installation, trusted personnel do certification

FEARLESS engineering Comparison With Requirements 4.CR4 provides logging; ER3 authenticates trusted personnel doing installation; CR5, ER4 control installation procedure New program UDI before certification, CDI (and TP) after 5.Log is CDI, so appropriate TP can provide managers, auditors access Access to state handled similarly

FEARLESS engineering Comparison to Biba Biba –No notion of certification rules; trusted subjects ensure actions obey rules –Untrusted data examined before being made trusted Clark-Wilson –Explicit requirements that actions must meet –Trusted entity must certify method to upgrade untrusted data (and not certify the data itself)

FEARLESS engineering Key Points Integrity policies deal with trust –As trust is hard to quantify, these policies are hard to evaluate completely –Look for assumptions and trusted users to find possible weak points in their implementation Biba based on multilevel integrity Clark-Wilson focuses on separation of duty and transactions