Open Reforms Pilots: Spring 2013 Fellowships Competition Report

Table of Contents

  1. Introduction
  2. Competition Overview
  3. Methods
  4. Summary of Results
  5. Conclusion and summary of changes as a result of the Pilot

Introduction

The success of the transition to the new Open Suite of Programs and peer review processes rests largely on our ability to efficiently and effectively pilot and test the functionality of peer review design elements in advance of their full implementation. As we complete each of the pilot studies, it is our intention to make findings available to contribute to the body of literature on peer review and program design.

Piloting peer review design elements will allow CIHR to adjust and refine processes and systems in order to best support applicants and reviewers. In order to identify areas for improvement, CIHR is collecting feedback from applicants, CIHR staff institutions, and peer reviewers through surveys.

It is important to note that when conducting pilots within existing competitions, CIHR will ensure that the quality of the peer review process is not compromised and that every application receives a fair, transparent and high-quality review.

Note: CIHR is also using data collected from this pilot to validate the optimal number of reviewers to assign to an application; and how often reviewers modify their reviews after viewing other reviewer’s reviews/participating in the asynchronous online discussion. These analyses are being conducted separately and the results will be made available at a later date.

Competition Overview

The Fellowship program was chosen for this pilot as it has been using a remote review process for many years as well as for the high volume of applications it receives. Therefore, the functionality implemented within this competition not only tested peer review design elements but also represented an enhancement to the existing peer review process. For this competition, 608 applications were reviewed by 209 reviewers.

Within this first pilot, a number of program design elements remained unchanged:

  1. Funds available
  2. Evaluation criteria
  3. Application-reviewer assignment process
  4. Funding decision process
  5. Rating scale
  6. Application structure

The objectives of the pilot were to test proposed functionalities and processes related to remote review and asynchronous on-line discussion. The following design elements of the proposed Project Scheme were tested:

  1. Increase the number of reviewers per application (from 2 to 3)
  2. Remote review functionality (submitting preliminary and final reviews remotely)
  3. Asynchronous online discussion

Note: The quality of individual reviews was not assessed during this pilot.

Table 1: Design Elements from the Project Scheme tested in the Fellowship Phase 1 pilot

Design Element Objective Pilot Description
Increased number of reviewers per application
  • Improve reliability of results
  • Eliminate the current re-review process
  • Determine the optimal number of reviewers per application
  • Increase the number of reviewers per application from 2 to 3
Remote review - submitting preliminary & final reviews
  • Validate the proposed process
  • Improve consistency between reviewers
  • Monitor and collect data on all of the review components of the proposed design
  • Analyse trends if viewing the other reviewers reviews and how the discussion has influenced the reviewer’s rating
  • Reviewers submitted preliminary reviews and then had access to other reviewers’ ratings and written reviews.
  • Reviewers were able to discuss applications/reviews, as appropriate, through an asynchronous online discussion tool.
Asynchronous online discussion
  • Improve consistency between reviewers
  • Eliminate the current re-review process
  • Validate the proposed process
  • Monitor and collect data on all of the review components of the proposed design
  • The asynchronous online discussion allowed reviewers to consult and view each other’s ratings and reviews and to discuss any areas of disagreement prior to submitting their final reviews.

It is important to note the limitations to this pilot. The Fellowships program is a training awards program and therefore there are some notable program design differences when compared to a grant program. Notably, applications are assigned to one of five pools based on the candidate’s background and their proposed research project. Reviewers appointed to each pool are then assigned randomly to applications within their respective pool. This is different than the more comprehensive expertise based assignment done in most CIHR grant programs.

Methods

Peer review was conducted from April 2nd, 2013 to May 10th, 2013 by 209 reviewers. Reviewers had three weeks to participate in the asynchronous online discussion prior to submitting their final reviews. On May 15th, 2013, following the closure of the review process, an invitation to complete a survey was sent directly to the 209 reviewers using the Fluid Survey Invitation tool. It provided a URL specific to each participant, allowing them to stop, save and return to the survey at any time. CIHR monitored the response rates and two reminder emails were sent to ensure that the target response rate was achieved.

Applicants were not surveyed in this pilot as the enhancements only impacted reviewers. Applicants were surveyed for the second Fellowship pilot in fall 2013, as enhancements to the structured application were implemented.

Summary of Results

The objective of Survey 1 (Appendix 2: Fellowships Reviewer Survey) was to assess the enhanced remote review process in ResearchNet by the reviewers with respect to:

  1. Their ResearchNet experience;
  2. The usefulness of accessing other reviewer’s reviews, and;
  3. The usefulness of the asynchronous online discussion Tool.

The detailed survey results and summary of key findings and observations can be found in Appendix 1.

The survey response rate of 70% exceeded targets. However, the small overall number of reviewers will limit the statistical analysis that is possible. We will continue to build the dataset with subsequent pilots and surveys. Nevertheless, the overall results of this survey are positive within the context of this competition. A number of areas for improvements were identified by survey respondents and many of them have already been implemented within the subsequent pilots.

i. ResearchNet Experience:

The overwhelming majority (91%) of reviewers found the ResearchNet technology user friendly although recommendations were presented to enhance the system’s interface. Overall, 84% of respondents agreed that the peer review process was efficient. There were a few technical problems and a number of suggestions for improvement were provided. Many of these improvements have been implemented within subsequent pilots (Fellowship Pilot Phase 2 & Knowledge Synthesis Pilot) as noted within the conclusion and summary of changes sections of this report.

ii. Accessing other peer reviewers reviews:

A majority of reviewers (89%) viewed the reviews submitted by other reviewers, and based on the comments provided to CIHR, they felt this functionality was very useful. On average, reviewers spent 2 hours or less reading the other reviews. However, there is currently no strong evidence to suggest that reading the other’s reviews helped reviewers finalize their submitted score. Only 15% of the respondents said that the other reviews helped them finalize their submitted score often, very often or always. CIHR will explore the inconsistency between these two statements further through analysis of competition data, future pilots and will also refine the relevant questions within subsequent surveys.

iii. Asynchronous online discussion:

The asynchronous online discussion was used by the majority of reviewers (88%). Scoring discrepancy was the most common factor (87%) used to determine if an application required discussion. A minority of reviewers (29%) commented that direction from CIHR on whether or not to discuss an application would reduce their workload; however, 61% of reviewers felt that it should be at the reviewers’ discretion whether or not to initiate a discussion. CIHR is currently exploring options to moderate the discussions to facilitate this process.

Conclusion and summary of changes as a results of the Pilot

The enhancements to the remote review process within this pilot represented many positive changes to the review process and will serve as the foundation for subsequent pilots. In response to the data analysis, observation and reviewer feedback in this first pilot, enhancements were made to some of the processes and functionalities for subsequent pilots. To improve the effectiveness of the new adjudication process, several new features, listed below, were added to ResearchNet for the Knowledge Synthesis Grant and Fellowship Phase 2 pilots:

  1. Deadline dates to complete the various task within the adjudication process are now displayed on ResearchNet.
  2. The asynchronous online discussion tool and process will be improved by:
    1. Allowing reviewers to access application and review information without closing the discussion tool;
      1. A table will display the ratings from each reviewer at the top of the discussion screen;
    2. Limiting the indentation of the text box to prevent the text from collapsing, regardless of the number of replies;
    3. Opening the discussion period once all reviewers assigned to a particular application have submitted their preliminary reviews or by a pre-assigned date in the system, whichever comes earliest;
    4. Shortening the timeframe for discussion;
    5. Increasing accountability within the process by revealing the identities of all reviewers at the preliminary review due date, regardless if they have submitted or not;
      1. Reviewers who have not submitted their review will see the identity of the other reviewers of the application but will not see their reviews;
  3. The reviewer interface on ResearchNet will be improved by:
    1. Displaying all of a reviewer’s assigned applications on one page;
    2. Providing reviewers with the functionality to expand the application details and view the other reviewers assessment;
    3. Displaying ratings based on their actual value and not the calculated weighted value;
    4. Creating a dynamic application rank order that will move applications up or down the rank list as ratings are entered for each adjudication criteria;
  4. Written comments will be made mandatory as part of the structured review template.

Appendix 1: Detailed Survey Results

Response Rate

Survey was sent to 209 reviewers and 145 responses were received resulting in a response rate of 70%.

Reviewer Characteristics

The reviewers were comprised of Mid-Career (45%) and Senior (42%) Scientists primarily from the Biomedical research domain (79%). Half of the respondents are Associate professors (50%) while 34% are Professors.

Note: the distribution (by sector) of reviewers within this competition is representative of the distribution (by sector) of the applications.

Figure 1. Would you classify yourself as a:

Figure 1 long description

Figure 2. Which research position(s) do you currently hold at this time?

Figure 2 long description

Figure 3. Which of the following is your primary research domain:

Figure 3 long description

Accessing the other reviewers reviews

89% of the reviewers took the extra time to read the other reviewer’s reviews. Most (93%) spent a total of 2 hours or less reading the other reviews. However, there is no strong evidence to suggest that reading the other’s reviews helped finalize the reviewer’s submitted score. See figure 6.

Figure 4. Did you look at the other reviews of the applications assigned to you?

Figure 4 long description

Figure 5. How much additional time in total (all applications) did you spend reading reviews of other reviewers?

Figure 5 long description

Figure 6. How often did the reading of other reviews help finalize your submitted score?

Figure 6 long description

Online Discussion

The asynchronous online discussion was used by 88% of the reviewers. Scoring discrepancy was the common factor (87%) to determine if a discussion was needed. However, comments suggest that CIHR should provide more guidelines on when applications should be discussed. There was no clear indication if the discussion helped to modify their final scores. See figure 12.

Representative Comments:

  • Discuss application only when close to fundable.
  • Some applications may not need discussion (scores are congruent amongst reviewers).
  • CIHR should prompt/moderate discussions when necessary.
  • Reviewers should NOT be able to submit final score without discussion.

Figure 7. Did you initiate an online discussion thread?

Figure 7 long description

Figure 8. What factor(s) did you use to determine whether an online discussion was needed for an application?

Figure 8 long description

Figure 9. Who should decide when an online discussion must take place?

Figure 9 long description

Figure 10. Did you participate in an online discussion?

Figure 10 long description

Figure 11. Did online discussion increase the amount of time required to complete the review of an application (on average) in this competition?

Figure 11 long description

Figure 12. How often did the online discussion help finalize your submitted score?

Figure 12 long description

Reviewing applications

While reviewing applications, reviewers spent most of their time reading the application (78 minutes) and less time writing their reviews (38 minutes). Contrary to grant programs, the Fellowship program has a structured review format and a much shorter grant application where reviewers are asked to provide brief comments on each criterion.

Figure 13. How much time, on average, did you spend on:

Figure 13 long description

Peer Review Process & Technology

Most reviewers believed that the peer review process is efficient (84%) and that the technology is user-friendly (91%). However, numerous recommendations to improve the process and technology were raised and some will be implemented in future pilots. See the Summary of changes section above.

Figure 14. Was the peer review process efficient?

Figure 14 long description

Representative Comments: How can the peer review process be improved?

  • Cannot see scores, comments and application data when participating in the online discussion; must go back and forth between the screens;
  • Provide options for some form of verbal communication;
  • Encourage reviewers to submit reviews on time – or have consequences for not respecting the process;
  • Standardized form of review – divide comments into categories: proposals, publications, etc.
  • Common CV has excess irrelevant information. Awful format; some columns are too narrow, which makes reading entries cumbersome;
  • Remote review process is fine for scholarships/fellowships at trainee stage but CANNOT possibly substitute for a face to face interaction which is necessary for grant applications;
  • Reviewers should be evaluated;
  • Shorter application.

Figure 15. Was the technology employed to conduct the reviews user-friendly?

Figure 15 long description

Representative Comments: How can the technology employed to conduct the reviews be improved?

  • Ability to see comments/scores while completing the online discussion;
  • CCV has duplicated information which appears in different sections with different formats. Presentation of the information complicates things;
  • Scores should be updated as they are revised for other reviewers to see;
  • Add video conferencing or give option to set up another form of communication.
  • Asynchronous discussion is inefficient and unreliable;
  • Designate a leader/moderator for the discussion.

Appendix 2: Fellowship Reviewer Survey

  1. Would you classify yourself as a:
    • Early career scientist (<5 years as an independent researcher)
    • Mid-career scientist (5-10 years as an independent researcher)
    • Senior scientist (>10 years as an independent researcher)
  2. Which research position(s) do you currently hold at this time?
    Please select all that apply:
    • Professor
    • Associate Professor
    • Assistant Professor
    • Researcher
    • Clinician
    • Other (specify):
  3. Which of the following is your primary research domain:
    • Biomedical
    • Clinical
    • Health systems/services
    • Social, cultural, environmental and population health
  4. a) Did you look at the other reviews of the applications assigned to you?
    • Yes
    • No
  5. b) Did you look at the other reviews of the applications assigned to you?
    • Under an hour
    • 1 to 2 hours
    • 3 hours or more
  6. c) How often did the reading of other reviews help finalize your submitted score?
    • Never (0%)
    • Rarely (up to 25%)
    • Occasionally (>25% to 50%)
    • Often (> 50% to 75%)
    • Very Often (>75%)
    • Always (100%)
  7. a) Did you initiate an online discussion thread?
    • Yes
    • No
  8. b) What factor(s) did you use to determine whether an online discussion was needed for an application?
    Choose all that apply
    • Scoring discrepancy
    • Content clarification required
    • Quality check
    • Other:
  9. a) Who should decide when an online discussion must take place?
    • CIHR (e.g., using a decision rule within ResearchNet)
    • Reviewers (at their discretion)
    • Don’t know
  10. b) What explicit criteria should be used to determine whether an online discussion must take place:
  11. a) Did you participate in an online discussion?
    • Yes
    • No
  12. b) Did online discussion increase the amount of time required to complete the review of an application (on average) in this competition?
    • Under an hour
    • 1 to 2 hours
    • 3 hours or more
  13. c) How often did the online discussion help finalize your submitted score?
    • Never (0%)
    • Rarely (up to 25%)
    • Occasionally (>25% to 50%)
    • Often (> 50% to 75%)
    • Very Often (>75%)
    • Always (100%)
  14. How much time, on average, did you spend on (rounded to periods of 15 minutes):
    Please choose a value of "0" if the column is not applicable.
    Hours | Minutes
    1. Reading a single application
    2. Writing the review of a single application
    3. Reviewing a single application
    4. Participating in an online discussion for an application
  15. a) Was the peer review process efficient?
    • Yes
    • No
  16. b) How can it be improved?
  17. a) Was the technology employed to conduct the reviews user-friendly?
    • Yes
    • No
  18. b) How can it be improved?
  19. a) Did you take the online peer review training for this Fellowships Pilot?
    • Yes
    • No
  20. b) How can it be improved?
  21. Do you have any additional feedback?
Date modified: