Continuous improvement of macromolecular crystal structures Tom Terwilliger (Los Alamos National Laboratory) DDD WG member ECM 2012: Diffraction Data Deposition.

Slides:



Advertisements
Similar presentations
Group 5: Historical control data Follow-up of 2005 IWGT where the use of historical control data in interpretation of in vitro results was identified as.
Advertisements

Data Curation in Crystallography: Publisher Perspectives JISC Data Cluster Consultation Workshop CCLRC, Didcot, Oxon 10 October 2006.
Chapter 9. Performance Management Enterprise wide endeavor Research and ascertain all performance problems – not just DBMS Five factors influence DB performance.
A vision involving raw data archiving via local archives as a supplement to the existing processed data archives (PDB, CSD, ICDD etc) John R. Helliwell,
From Objectives to Methods (d) Research methods A/Prof Rob Cavanagh April 7, 2010.
Primary ICT Assessment What does good assessment look like? The ICT Assessment Toolkit © NEWLICT North East and West London ICT Consultants' Group.
PSAE Practice Session Science Mr. Johns Room 2012.
The Experience Factory May 2004 Leonardo Vaccaro.
UI Standards & Tools Khushroo Shaikh.
Effective systems development requires a team effort from stakeholders, users, managers, systems development specialists, and various support personnel,
Fundamentals of Information Systems, Second Edition
Software Process and Product Metrics
1 CMPT 275 Software Engineering Requirements Analysis Process Janice Regan,
Science and Engineering Practices
Architecture and Software Product Lines A software architecture represents a significant investment of time and effort, usually by senior talent. So it.
Sharif University of Technology Session # 4.  Contents  Systems Analysis and Design Sharif University of Technology MIS (Management Information System),
The Re-engineering and Reuse of Software
Chapter : Software Process
Chapter 7 Requirement Modeling : Flow, Behaviour, Patterns And WebApps.
Torrington, Hall & Taylor, Human Resource Management 6e, © Pearson Education Limited 2005 Slide 7.1 Importance of Selection The search for the perfect.
Overview of the Database Development Process
DATA FOUNDATION TERMINOLOGY WG 4 th Plenary Update THE PLUM GOALS This model together with the derived terminology can be used Across communities and stakeholders.
Chapter 2 The process Process, Methods, and Tools
Module 4: Systems Development Chapter 13: Investigation and Analysis.
NSI/ISI Statistical software Issues and a way forward to maximise re-use and minimise integration efforts by Andrea Toniolo Staggemeier.
Using the SAS® Information Delivery Portal
Semester 2: Lecture 9 Analyzing Qualitative Data: Evaluation Research Prepared by: Dr. Lloyd Waller ©
Computing Fundamentals Module Lesson 19 — Using Technology to Solve Problems Computer Literacy BASICS.
This chapter is extracted from Sommerville’s slides. Text book chapter
Update on Selective Editing and Implications for Staff Skills International Trade Conference September 2008 Ken Smart.
INTERACTIVE ANALYSIS OF COMPUTER CRIMES PRESENTED FOR CS-689 ON 10/12/2000 BY NAGAKALYANA ESKALA.
 Fundamentally, data mining is about processing data and identifying patterns and trends in that information so that you can decide or judge.  Data.
EBI is an Outstation of the European Molecular Biology Laboratory. Annotation Procedures for Structural Data Deposited in the PDBe at EBI.
- Ahmad Al-Ghoul Data design. 2 learning Objectives Explain data design concepts and data structures Explain data design concepts and data structures.
Towards an Experience Management System at Fraunhofer Center for Experimental Software Engineering Maryland (FC-MD)
Regional Data Management Structure. Questions We Hear  Why are we reporting differently than other agencies? What are we being asked to do that is different.
Fundamentals of Information Systems, Second Edition 1 Systems Development.
New Advanced Higher Subject Implementation Events Biology: Unit Assessment at Advanced Higher.
Data Integration and Management A PDB Perspective.
Configuration Management and Change Control Change is inevitable! So it has to be planned for and managed.
Basic Nursing: Foundations of Skills & Concepts Chapter 9
MODELING MATTER AT NANOSCALES 3. Empirical classical PES and typical procedures of optimization Classical potentials.
Theme 2: Data & Models One of the central processes of science is the interplay between models and data Data informs model generation and selection Models.
A centre of expertise in digital information managementwww.ukoln.ac.uk Quality Assurance For Museum Web Sites: Review Brian Kelly UKOLN University of Bath.
TSS Database Inventory. CIRA has… Received and imported the 2002 and 2018 modeling data Decided to initially store only IMPROVE site-specific data Decided.
The Marketing Research Process and Proposals Copyright © 2010 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
CIWQS Review Phase II: Evaluation and Final Recommendations March 14, 2008.
Atomic structure model
Process Asad Ur Rehman Chief Technology Officer Feditec Enterprise.
Modelling the Process and Life Cycle. The Meaning of Process A process: a series of steps involving activities, constrains, and resources that produce.
Requirement engineering & Requirement tasks/Management. 1Prepared By:Jay A.Dave.
Fundamentals of Information Systems, Third Edition2 An Overview of Systems Development: Participants in Systems Development Development team –Responsible.
Developing our Metadata: Technical Considerations & Approach Ray Plante NIST 4/14/16 NMI Registry Workshop BIPM, Paris 1 …don’t worry ;-) or How we concentrate.
1 The XMSF Profile Overlay to the FEDEP Dr. Katherine L. Morse, SAIC Mr. Robert Lutz, JHU APL
Introduction to Machine Learning, its potential usage in network area,
Presented by Edith Ngai MPhil Term 3 Presentation
PDBe Protein Interfaces, Surfaces and Assemblies
Software Project Configuration Management
Marketing Research and Information Systems
(Additional materials)
Brain Initiative Informational Conference Call
Martin Rajman, Martin Vesely
Reduce the need for human intervention in protein model building
Goal, Question, and Metrics
1.b What are current best practices for selecting an initial target ligand atomic model(s) for structure refinement from X-ray diffraction data?
Chapter # 8 Quality Management Standards
Computer Literacy BASICS
Applying principles of computer science in a biological context
M. Kezunovic (P.I.) S. S. Luo D. Ristanovic Texas A&M University
Ton Spek Utrecht University The Netherlands Vienna –ECM
Presentation transcript:

Continuous improvement of macromolecular crystal structures Tom Terwilliger (Los Alamos National Laboratory) DDD WG member ECM 2012: Diffraction Data Deposition Working Group Meeting August 6, 2012 Bergen, Norway

Background Deposition of data allows reanalysis:...in the long term the community will probably have to face the issue of whether the structural database should be static or dynamic. As methodology improves, it seems likely that re-refinement of older models (either on a case-by-case basis, or as one large-scale project) might provide better models and, hopefully, increase our understanding of the chemistry and biology of the molecules under study. --Kleywegt, Harris, Zou, Taylor, Wählby, Jones (2004). The Uppsala Electron Density Server. Acta Cryst D60, 2240.

Deposition of data and metadata allows model improvement:...improvements in crystallographic software and validation tools, combined with the deposition of X-ray data into the PDB, have enabled the development of automated rerefinement protocols... which can improve most structure models compared with their initially deposited form......greater sophistication can be achieved [from deposition of ] metadata......users of the PDB and software developers will greatly benefit...as it will turn what was previously a static archive of frozen models into a repository of self-improving results.... Depositors will also benefit... because it will make their structural results more futureproof leading to more citations and to higher visibility --Joosten, Womack, Vriend, Bricogne (2009). Re-refinement from deposited X-ray data can deliver improved models for most PDB entries. Acta Cryst. D65, 176–185.

The current paradigm The PDB is the definitive repository of macromolecular structures The crystallographic community uses the PDB as an archive of models and data The biological community uses the PDB as its window into structures Other repositories would need to add major value to structures to be widely viewed

The current paradigm, cont... One-time structure determination and validation is standard Determine structure Deposit data and model (interpretation of data) Possibility of updating structures exists but infrequently used Original depositor can update a structure (removing original) Other researchers can (in principle) re-analyze and deposit another interpretation

The current paradigm, cont... Historical focus has been on the model, not the data Data are used to obtain and validate model Model is used by the community There is only rare reference to the data A PDB entry is a model, with data and metadata added

Why should there be continuous improvement of structures? Reinterpretation is feasible PDB-REDO has demonstrated the feasibility of systematic, continuous reinterpretation of the data in the PDB Software and algorithmic improvements will extend this capability Standardized validation procedures exist to evaluate model quality Availability of raw images will extend reinterpretation even further

Why should there be continuous improvement of structures? (cont...) Reinterpretation is desirable Minor errors (or major ones) can be fixed Consistency among related structures can be improved by using consistent methods New structural information (loops, ligands) can be obtained New formalism describing a crystallographic structure (multi-model interpretations) can be employed Models can be optimized to provide maximal information about specific features (a particular inter-atomic distance) Joint refinement of groups of structures can improve all of them

Why isnt continuous improvement already fully in place? Cultural issues The interpretation is customarily owned by the person who solved the structure The depositor invested major effort to obtain the structure Improvements in the model may be looked at as criticism of the depositor Practical issues Which model should a person look at? When is a new model improved? The original depositor may have had access to critical information or employed specialized approaches Deposition is a big job

What might continuous improvement of macromolecular structures include? Large-scale efforts Systematic optimization of all structures in the PDB Redetermination of groups of related structures Redetermination of groups of structures focusing on specific questions Small-scale efforts Optimization of structures that are of interest to specific investigators

What is needed for successful continuous improvement of macromolecular structures? Validated processed data and metadata Data and considerable metadata are already collected by the PDB Ideal would be for deposited data and metadata to be validated by full re-determination of the structure Unmerged data Raw diffraction data Raw diffraction images Multiple X-ray wavelengths Data from heavy-atom derivatives

What is needed for successful continuous improvement of macromolecular structures? (cont...) Multiple views of the PDB The views of the archive of structures presented by the PDB can have a major influence on biological interpretation of structure The model that is appropriate may vary because our models are not full descriptions of what is in the crystals The appropriate model will depend on the question being asked (What is the distance between these atoms? What is the buried surface area? What is the electrostatic potential? What is the difference between these two structures?) A way to choose the right model for the question being asked Continuous improvement will need be accompanied by a mechanism for retrieving models appropriate for a variety of uses Choice of model might include validation metrics, resolution, method used for determining model, model parameterization