Chair: Professor Dame Ann Dowling Sub-panel Chairs: Panel Advisers: SP07: Professor David Price Dr Karen Ness SP08: Professor Richard Catlow Ms Lesley.

Slides:



Advertisements
Similar presentations
GSOE Impact Workshop Impact and the REF 19 th May 2010 Lesley Dinsdale.
Advertisements

Research Excellence Framework Jane Boggan Planning Division Research Staff Forum - January 2010.
Main Panel A: Subpanels and Chairs A1: Clinical Medicine - Christopher Day, Newcastle University A2: Public Health, Health services and Primary Care -
National Professional Qualification for Headship
REF2014 HODOMS Birmingham 8 th April Ann Dowling: Chairman of REF Main Panel B John Toland: Chairman of REF Sub-Panel B10: Mathematical Sciences.
Service to the University, Discipline and Community Academic Promotions Briefing Session Chair, Academic Board Peter McCallum.
RAE 2008 Engineering Ann Dowling, Chairman Main Panel G - Engineering EPC Congress March 2005.
Supporting & promoting Equality & Diversity through REF Dianne Berry, Chair REF E&D Advisory Panel Ellen Pugh, Senior Policy Officer ECU.
The Research Excellence Framework Assessment framework, guidance on submissions and panel criteria.
Guidance on submissions Chris Taylor, Deputy REF Manager Graeme Rosenberg, REF Manager.
These slides have been produced by the REF team, and were last updated on 3 September 2011 They provide a summary of the assessment framework and guidance.
Communicating the outcomes of the 2008 Research Assessment Exercise A presentation to press officers in universities and colleges. Philip Walker, HEFCE.
The Research Assessment Exercise in the United Kingdom Paul Hubbard International colloquium “Ranking and Research Assessment in Higher Education” 13 December.
2005/6 ATN Research Quality Framework (RQF) Trial Curtin University of Technology Queensland University of Technology RMIT University University of South.
The Research Excellence Framework Panel criteria [Main Panel Chair] Graeme Rosenberg.
Achieving and Demonstrating Research Impact John Scott.
Demonstrating research impact in the REF Graeme Rosenberg REF Manager
The Research Excellence Framework. Purpose of REF The REF replaces the RAE as the UK-wide framework for assessing research in all disciplines. Its purpose.
The Research Excellence Framework. Presentation outline The REF assessment framework and guidance on submissions: - Overview - Staff - Outputs - Impact.
Consultation on panel criteria and working methods.
REF Information Session August Research Excellence Framework (REF)
2015 Commendations and Citations Information Session.
The Research Excellence Framework Data and Audit May 2012.
The UK Experience of Quality Assurance in Research and Doctoral Education Dr Robin Humphrey Director of Research Postgraduate Training Faculty of Humanities.
Research Quality Assessment following the RAE David Sweeney Director, Research, Innovation, Skills.
The REF assessment framework and guidance on submissions Linda Tiller, HEFCW 16 September 2011.
SSHRC Partnership and Partnership Development Grants Rosemary Ommer 1.
Introduction to the Research Excellence Framework.
The Research Excellence Framework Briefing events for HEI contacts 21 May: Glasgow 23 May: Manchester 24 May: London 28 May: Cardiff 31 May: London.
Research Assessment Exercise RAE Dr Gary Beauchamp Director of Research School of Education.
Academic Research Enhancement Award (AREA) Program Erica Brown, PhD Director, NIH AREA Program National Institutes of Health 1.
Page 1 RESEARCH EXCELLENCE FRAMEWORK : RESEARCH IMPACT ASESSMENT LESSONS FROM THE PILOT EXERCISE Professor John Marshall Director Academic Research Development.
The Research Excellence Framework Expert Advisory Groups round 1 meetings February 2009 Paul Hubbard Head of Research Policy.
Professor Andrew Wathey Vice-Chancellor and Chief Executive Northumbria University.
Research Quality Framework Presentation to APSR - ARROW - Repository Market Day 4 May 2007 Sandra Fox Department of Education Science and Training.
Main Panel D Criteria and Working Methods Main Panel D covers: Area Studies Modern Languages and Linguistics English Language and Literature History Classics.
The Research Excellence Framework Impact: the need for evidence Professor Caroline Strange 22 June 2011.
THE IMPACT OF RAE ON SERIAL PUBLICATION Professor Judith Elkin UK Serials Group March 2004.
12/9/10 Pilot assessment impact- paperwork Findings of the expert panels- report + appendix Lessons learned- feedback from pilot institutions Examples.
The REF assessment framework (updated 23 May 2011)
Delivering Strength Across the Piece David Sweeney Director, Research, Education and Knowledge Exchange HEPI, Royal Society 31 March 2015.
Main Panel A Criteria and Working Methods Cardiff School of Biosciences Ole H Petersen Chair.
Assessment Validation. MORE THAN YOU IMAGINE ASQA (Australian Skills Quality Authority) New National Regulator ASQA as of 1 July, 2011.
What is impact? What is the difference between impact and public engagement? Impact Officers, R&IS.
ACADEMIC PROMOTIONS Promotions Criteria Please note, these slides only contain a summary of the promotions information – full details can be found.
Research Excellence Framework 2014 Michelle Double Hyacinth Gale Sita Popat Edward Spiers Research and Innovation Support Conference.
Impact and the REF Consortium of Institutes of Advanced Study 19 October 2009 David Sweeney Director (Research, Innovation and Skills)
The Research Excellence Framework Assessment framework and guidance on submissions Graeme Rosenberg, REF Manager.
Towards REF 2020 What we know and think we know about the next Research Excellence Framework Dr. Tim Brooks, Research Policy & REF Manager, RDCS Anglia.
Impact and the REF Tweet #rfringe17
WP2. Excellent university for the researchers
Consultation on draft panel criteria and working methods
REF 2021 Briefing 25 January 2018.
REF 2021 What we know and thought we knew, in preparation for the next Research Excellence Framework Dr. Tim Brooks, Research Policy & REF Manager, RDCS.
THE OFFICE FOR SCHOLARLY COMMUNICATION/ Responsible Metrics at Kent
Law Sub-panel Generic Feedback - Impact
These slides have been produced by the REF team, and were last updated on 30 January 2012 They provide a summary of the assessment framework and guidance.
REF 2021 Briefing Consultation on the draft guidance
These slides have been produced by the REF team, and were last updated on 30 January 2012 They provide a summary of the assessment framework and guidance.
Anglia Ruskin REF Awayday 2017
Consultation on draft panel criteria and working methods
Research Update GERI May 2010.
Consultation on the REF 2021 guidance and criteria
Research Excellence Framework: Past and Future
Towards Excellence in Research: Achievements and Visions of
Consultation on draft panel criteria and working methods
Rating in 2002 for funding from 2003
REF and research funding update
UCML, London 18 January 2019 REF 2021 Susan Hodgett (D25)
Understanding Impact Stephanie Seavers, Impact Manager.
Presentation transcript:

Chair: Professor Dame Ann Dowling Sub-panel Chairs: Panel Advisers: SP07: Professor David Price Dr Karen Ness SP08: Professor Richard Catlow Ms Lesley Dinsdale SP09: Professor Brian FosterPanel Secretary: SP10: Professor John TolandSimon Kerridge SP11: Professor Steve Furber SP12: Professor John Clarkson SP13: Professor Steve Williamson SP14: Professor Roger Falconer SP15: Professor Philip Nelson Main Panel B

Format of document Section 1: Submissions and units of assessment Section 2: Assessment criteria: outputs Section 3: Assessment criteria: impact Section 4: Assessment criteria: environment Section 5: Working methods Main Panel B Criteria and Working Methods

UOAs 7 – 15, covering physical & mathematical sciences and engineering Encourage use of research groups to structure submissions Eligibility for multiple submission requests depends on nature of disciplines covered: o Expected for UOAs 12 (Aeronautical, Mechanical, Chemical & Manufacturing Engineering) and 13 (Electrical & Electronic Engineering, Metallurgy & Materials) o Requests must meet criteria set out in Guidance on Submissions (GOS), para 50 Submissions and UOAs Main Panel B CWM: Submissions and Units of Assessment

Originality, significance and rigour If appropriate to output type, may consider editorial and refereeing standards as part of the indication of rigour For some UOAs, citation information will be used as part of the indication of academic significance Types of research output All types of research output eligible for submission but HEIs should be mindful that purpose is to assess original research; for reviews will assess only original research or new insights reported Assessment Criteria: Outputs Main Panel B CWM: Outputs Assessment Criteria

Co-authored outputs A co-authored output may not be listed against more than one member of staff returned within the same submission (may be listed in different submissions, either same or different HEIs) Sub-panel9 only o For outputs with 5 or more co-authors, requires information on co-authorship (maximum 100 words) o Used to assess whether significant contribution o Then assess output as a whole All other sub-panels o Assume co-authors have made substantial contribution (could request audit information); will ignore any information provided o Assess output as a whole Assessment Criteria: Outputs Main Panel B CWM: Outputs Assessment Criteria

Double weighted outputs Allowed, but expected to occur only rarely for Main Panel B disciplines Must be requested by submitting institution (maximum 100 words) Assess request for double weighting, then assess quality No reserves Assessment Criteria: Outputs Main Panel B CWM: Outputs Assessment Criteria

Additional information on outputs Clarification of research process and/or content (maximum 300 words) o For non-text or practice-based outputs: description of research process and content if not evident within output o For reviews: identification of original research or new insights reported Additional factual information (maximum 100 words) o Sub-panels 11, 12, 13, 14, 15, only: factual information about significance of outputs not evident in outputs eg how gained recognition, led to further developments or been applied; opinion will be ignored o Not required for sub-panels 7, 8, 9, 10: will ignore any information provided in this category Assessment Criteria: Outputs Main Panel B CWM: Outputs Assessment Criteria

Additional information on outputs (continued) Clarification of content of outputs with significant material in common with output published before 1 January 2008 (GOS para 113) o General principle that panels will assess the original content or new insights reported in the output. o Recognition that some would not be considered by the panels as representing the published version of the work, for example, preliminary findings disseminated to a limited audience, technical reports, or some forms of conference contribution. These will be assessed in full. o HEIs provide details of outputs, and, where appropriate, information on how far the original work has been revised to incorporate new findings, including identification of the new research or new insights reported, additional to the earlier published output (maximum 300 words) Assessment Criteria: Outputs Main Panel B CWM: Outputs Assessment Criteria

Citation data Sub-panels 7, 8, 9, and 11 will make use of citation data, where it is available, as part of the indication of academic significance, as follows: o Data will be provided by the REF team; in addition sub- panel 11 will use Google Scholar; no other sources of bibiometric analysis will be used o Will not be a primary tool in assessment o Absence of citation data will not automatically be taken to mean absence of significance o Mindful that citations can be an unreliable indicator – some forms of output, equalities issues Sub-panels 10, 12, 13, 14 and 15 will not receive or make use of citation data, or any other sources of bibliometric analysis Assessment Criteria: Outputs Main Panel B CWM: Outputs Assessment Criteria

Range of impacts The panel expects to see impact across a wide range of types and will consider all equitably in the assessment of submitted case studies Case studies may describe more than one type of impact Impacts may be manifest in a wide variety of ways and the list of types provided is only indicative; the panels will consider any impact that meets the general definition in the GOS Assessment criteria: Impact

Evidence of Impacts An indicative list of types of evidence is provided but this is not exhaustive; sub-panels will consider any appropriate qualitative or quantitative evidence that is independently verifiable Highly confidential or sensitive impacts - consultation proposal on how to deal with these Impacts arising from public engagement activity – importance of link to underpinning research and evidence of reach beyond just dissemination Assessment criteria: Impact

Underpinning research Submitting unit must demonstrate that it made a significant contribution to the underpinning research Importance of demonstrating how the research led to, underpinned or contributed to the impact Threshold quality of research defined as 2* Up to six key references to the underpinning research may be listed in the case study; up to three of these that best indicate the quality should be highlighted Assessment criteria: Impact

Reach and Significance Reach is the extent and breadth of the beneficiaries of the impact Significance is the degree to which the impact has enabled, enriched, influenced, informed or changed the products, services, performance, practices, policies or understanding of commerce, industry or other organisations, governments, communities or individuals Impacts may be submitted at any stage of maturity but early stage or interim impacts may score less highly than more mature impacts Assessment Criteria: Impact

Impact template Main Panel B’s CWM give further information on the range of information it expects to see in each section Sections b. (Approach to impact) and d. (Strategies and plans for supporting impact) will be assessed holistically. Sections a. (Context) and d. (Relationship between the approach and the submitted case studies) are for information and will not be assessed Assessment criteria: Impact

Criteria are o Vitality: extent to which a unit provides an encouraging environment for research, has an effective strategy, is engaged with the national and international research and user communities, and is able to attract excellent postgraduate and postdoctoral researchers. o Sustainability: consideration of leadership, vision for the future and investment in people and infrastructure and, where appropriate for the subject area, the extent to which activity is supported by a portfolio of research funding. Will interpret ‘environment’ as relating to both the research environment within the submitting unit and its participation in, and contribution to, its subject discipline and academic community Assessment Criteria: Environment Main Panel B CWM: Environment Assessment Criteria

Environment template (REF5) Headings – and weightings in assessment o Overviewinformation only o Strategy20% o People30% (staff and students) o Income, infrastructure & facilities 30% o Collaboration & contribution to 20% the discipline Environment quantitative data (REF 4a/b/c) assessed under o People (doctoral awards) o Income (research income including in-kind) Panel criteria specifies the kinds of information sub- panels would wish to see under these headings Assessment Criteria: Environment Main Panel B CWM: Environment Assessment Criteria

Environment template (REF5) - continued Some sub-panels are requesting additional quantitative information to be included in REF5 to be considered under specific headings: In relation to People: o Sub-panels 7, 8, 9 and 11: FTE number of research assistants on staff census date (31 October 2013) o Sub-panels 8 and 11: FTE number of registered postgraduate research students on 31 July for each year of the assessment period In relation to Income, infrastructure & facilities: o Sub-panel 9 only: additional information on usage of major national and international facilities not supported by the Research Councils Assessment Criteria: Environment Main Panel B CWM: Environment Assessment Criteria

Interdisciplinary and multidisciplinary research welcomed and treated equally. Sub-panels members have been selected to embrace broad- ranging experience to enable assessment of such work and work that crosses UOA boundaries. Additional assessors (both academic and user) will be appointed to each sub-panel to assist with the assessment phase where required. Sub-panel requirements for assessors in will be informed by HEI submission intentions. Assessment Criteria: Working Methods

Main Panel will work with sub-panels to ensure adherence to assessment criteria and consistent application of standards. Details given in Panel Criteria Sub-panels will ensure that submissions are assessed using appropriate expertise: approaches defined in Panel Criteria User members and user assessors will contribute significantly to the assessment of impact and may also assess outputs where appropriate All outputs will be examined in a sufficient level of detail sufficient to develop a robust quality profile Assessment Criteria: Working Methods

Further information From our website: From your nominated institutional contact or data contact Or