Skip to content
  • Categories
  • CSPaper Review
  • Recent
  • Tags
  • Popular
  • Paper Copilot
  • OpenReview.net
  • Deadlines
  • CSRanking
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
CSPaper

CSPaper: peer review sidekick

  1. Home
  2. Peer Review in Computer Science: good, bad & broken
  3. Transparent Peer Review in Higher Education: Lessons for Computer Science

Transparent Peer Review in Higher Education: Lessons for Computer Science

Scheduled Pinned Locked Moved Peer Review in Computer Science: good, bad & broken
peer reviewtransparencyanonymityacademic publishingopen scienceaccountabilitytrustcomputer science
1 Posts 1 Posters 65 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • rootR Offline
    rootR Offline
    root
    wrote last edited by
    #1

    Recently, the European Journal of Higher Education (EJHE) became the first journal in its field to adopt a Transparent Peer Review (TPR) policy, beginning in April 2023. The policy publishes anonymous peer review reports alongside accepted articles, allowing readers to see the feedback and revisions that shaped the final paper. After 18 months, the journal conducted surveys of authors and reviewers, and the results provide valuable insights that resonate strongly with the ongoing debates in computer science peer review.

    This article summarizes the key findings of the EJHE case study and discusses their implications for computer science researchers.

    Screenshot 2025-09-17 at 10.18.40.png

    Key Findings from EJHE’s Transparent Peer Review

    Awareness and Experience

    • 62% of authors and 41% of reviewers already had prior experience with transparent peer review.
    • Before engaging with EJHE’s policy, 46% of authors and 32% of reviewers had a positive perception of TPR.

    Screenshot 2025-09-17 at 10.20.52.png
    Positive perception before TPR experience


    Author Satisfaction

    After engaging with TPR at EJHE:

    • 82% of authors were satisfied with review quality.
    • 76% were satisfied with review speed.
    • 80% were satisfied with decision making.

    Screenshot 2025-09-17 at 10.22.48.png

    When comparing TPR to previous review processes:

    • 40% said TPR was better,
    • 58% noticed no difference,
    • Only 2% thought it was worse.

    Screenshot 2025-09-17 at 10.23.35.png


    Reviewer Perspectives

    • 76% of reviewers preferred not to have their names included.
    • 86% said transparency had no impact on how they wrote their reviews.
    • Only 24% indicated that TPR would influence their future decision to review.

    Please see charts here: https://insights.taylorandfrancis.com/research-impact/a-case-study-on-implementation-and-impact-of-transparent-peer-review-tpr/


    Future Intentions

    • 73% of authors were very likely to submit again.
    • 94% said the TPR policy had no negative influence on their decision to submit.
    • 40% of authors indicated TPR had a positive influence on their willingness to review.

    Please see charts here: https://insights.taylorandfrancis.com/research-impact/a-case-study-on-implementation-and-impact-of-transparent-peer-review-tpr/


    Qualitative Feedback

    Authors praised TPR for:

    • High-quality, actionable reviewer comments.
    • Professional, constructive review tone.
    • Improved quality of final papers.

    Usage data reinforces this: peer review reports averaged 141 views and 63 downloads per article, showing strong community interest in reviewer feedback.


    Implications for Computer Science Domain

    1. Increasing Trust in the Review Process

    Computer science has faced repeated concerns about review quality, transparency, and fairness, particularly in large conferences (NeurIPS, ICML, CVPR). TPR offers a concrete mechanism to:

    • Increase accountability by showing how papers were evaluated.
    • Provide a learning resource for early-career researchers, who can study real reviews.

    2. Preserving Anonymity While Sharing Insight

    The EJHE experience highlights a key tension: reviewers overwhelmingly want to remain anonymous (76% opposed naming). This finding aligns with CS researchers’ fears of retaliation or reputational risk in competitive fields.
    👉 For CS venues, adopting TPR without revealing identities seems both feasible and widely acceptable.

    3. Reviewer Workload and Incentives

    A common worry is that transparency could increase reviewer burden. However, 86% of reviewers said their approach did not change. This suggests that TPR can be introduced without harming reviewer efficiency.
    Further, public review reports could be integrated into reviewer recognition systems (e.g., ORCID, Publons), providing credit without breaching anonymity.

    4. Impact on Submissions

    In computer science, authors often worry about bias or inconsistency in reviews. The EJHE data show 40% of authors found TPR better than traditional processes, and 94% were not negatively influenced in submission decisions. This suggests TPR could help retain authors in top CS venues by signaling rigor and fairness.


    Potential Takeaways for CS Peer Review Reform

    The EJHE case study offers strong evidence that:

    • TPR improves transparency and trust without discouraging submissions or reviews.
    • Reviewer anonymity remains crucial.
    • Usage metrics confirm that published reviews are valued by the wider research community.

    For computer science, where conference review crises (e.g., overloaded reviewer pools, inconsistent feedback) are frequent topics of debate, Transparent Peer Review could be a pragmatic reform.
    It balances openness with practicality and may address long-standing calls for greater accountability in our field.


    Final Thoughts

    The EJHE study demonstrates that Transparent Peer Review is not just an idealistic vision but a workable model that benefits authors, reviewers, and readers alike.

    As computer science grapples with scaling peer review for massive conferences and journals, TPR offers a way to:

    • Build trust in the process,
    • Encourage constructive, professional reviewing, and
    • Provide the community with valuable insights into the scholarly dialogue behind published work.

    Perhaps it’s time for CS conferences and journals to experiment with TPR pilots — starting small, but with the potential for transformative impact.


    1 Reply Last reply
    0
    Reply
    • Reply as topic
    Log in to reply
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes


    • Login

    • Don't have an account? Register

    • Login or register to search.
    © 2025 CSPaper.org Sidekick of Peer Reviews
    Debating the highs and lows of peer review in computer science.
    • First post
      Last post
    0
    • Categories
    • CSPaper Review
    • Recent
    • Tags
    • Popular
    • Paper Copilot
    • OpenReview.net
    • Deadlines
    • CSRanking