Skip to content
  • Categories
  • CSPaper Review
  • Recent
  • Tags
  • Popular
  • Paper Copilot
  • OpenReview.net
  • Deadlines
  • CSRanking
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
CSPaper Forum

CSPaper: peer review sidekick

  1. Home
  2. Peer Review in Computer Science: good, bad & broken
  3. Artificial intelligence & Machine Learning
  4. 📈 ICLR 2026 Submission Storm: 19,797 Papers, New Review Rules, and the Age of AI Convergence

📈 ICLR 2026 Submission Storm: 19,797 Papers, New Review Rules, and the Age of AI Convergence

Scheduled Pinned Locked Moved Artificial intelligence & Machine Learning
iclr2026submissionpeer reviewtopicstrendbrazilthemellmcspaper
1 Posts 1 Posters 537 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • rootR Offline
    rootR Offline
    root
    wrote last edited by root
    #1

    Screenshot 2025-10-16 at 02.17.21.png

    📈 Record-Breaking Submissions: ICLR 2026 Hits Nearly 20,000 Papers

    ICLR 2026 has officially broken all previous records with 19,797 valid submissions on OpenReview. That’s a 70% jump from ICLR 2025 (11,672 submissions), confirming ICLR’s position as the largest machine learning venue on Earth.

    Year Submissions Acceptances Acceptance Rate
    ICLR 2023 4,955 1,575 31.78%
    ICLR 2024 7,304 2,260 30.94%
    ICLR 2025 11,672 3,704 31.73%
    ICLR 2026 19,797 TBD (expected ~30%)

    The growth isn’t just quantitative; it mirrors an ongoing paradigm shift across the ML landscape: from building models to governing and applying them.


    🧭 Macro Trends: From Model Construction to Model Civilization

    Based on a large-scale analysis of 19,658 paper titles, keywords, and primary research areas, ICLR 2026 showcases a decisive turn in machine learning research priorities:

    “The era of model construction has ended; we are now entering the era of model application and governance.”

    🔥 Dominant Themes

    1. LLMs Everywhere — “Foundation or frontier models, including LLMs” tops all areas, making up over 20% of all papers.

      • Over 70% of submissions in reinforcement learning, alignment, and safety also involve LLMs.
      • Keywords like Latent Reasoning, Graph-of-Thought, RLHF, and Agentic AI dominate.
    2. AI for Science Boom — Explosive growth in physics, chemistry, biology, and neuroscience applications, marking the rise of “AI as the 4th scientific paradigm.”

    3. The Second RL Revolution — RL has shifted from games to aligning LLMs (DPO, GRPO, RLV-R), becoming the backbone for model reasoning and control.


    📊 Field Distribution: The Empire of Foundation Models

    Rank Primary Area Papers %
    1 Foundation / Frontier Models (incl. LLMs) 3,962 20.15%
    2 CV / Audio / Multimodal Applications 3,458 17.59%
    3 Generative Models 1,841 9.36%
    4 Alignment, Fairness, Safety 1,512 7.69%
    5 Datasets & Benchmarks 1,496 7.61%
    6 Reinforcement Learning 1,291 6.57%
    7 Representation Learning 1,028 5.23%
    8 Optimization 884 4.50%
    9 Explainable AI 689 3.50%
    10 Physics & Science Applications 462 2.35%
    ... ... ... ...

    This paints a crystal-clear picture: LLMs are not just a topic, they are the gravitational field reshaping every discipline. For the complete analysis, see here.


    🧩 Hot Themes and Frontiers

    🧠 Reasoning Renaissance

    Chain-of-Thought (CoT) evolves into Tree-of-Thought and Graph-of-Thought structures; “Latent Reasoning” is the new darling, aiming for implicit, efficient thought processes.

    🧍‍♂️ Agentic AI

    LLM Agents are no longer toys: they’re autonomous planners and tool users. Multi-agent collaboration and debate models explore social and evolutionary intelligence.

    🎞️ Multimodal Fusion

    Diffusion Transformers dominate video generation, while Text-to-3D and Gaussian Splatting lead the 3D frontier. “Unified multimodal models” (Omni-modal AI) become the next big wave.

    🧰 Efficiency Revolution

    Compression, quantization (even 1-bit!), and speculative decoding are now mandatory for deployment. Data-centric AI gains momentum: quality > quantity.


    🧮 Review Pressure: Two Weeks, Binary Scores, and 20,000 Dreams

    The ICLR 2026 review cycle has drawn controversy for its “extreme two-week” timeline — reviewers have until Oct 31, 2025 (AoE) to submit all feedback.

    ⏱️ Two Weeks to Rule Them All

    • 19,797 papers
    • ~100,000 reviewers
    • Only 14 days to read, judge, and rate
    • Deadline for initial results: Nov 12, 2025, 21:00

    ⚖️ The “Even-Only” Scoring System

    Gone is the classic 1–5 scale. The new 2–4–6–8–10 system removes the middle ground, forcing reviewers into decisive opinions.

    Score Meaning
    0 Strong Reject
    2 Reject
    4 Weak Reject
    6 Weak Accept
    8 Accept
    10 Strong Accept

    The new system acts as an “opinion amplifier”, making positive and negative evaluations more polarized: a double-edged sword for borderline papers.

    💬 Reviewer Insight

    The new formula reviewers adopt is:
    Novelty × Depth × Presentation × Reproducibility ≈ Weak Accept+

    And yes — LLMs like ChatGPT are allowed, but only for grammar correction, not for content drafting.


    🧩 Spotlight: SAM 3 & Mamba-3 — and more?

    Among nearly 20K submissions, two papers dominate the buzz — SAM 3 in vision and Mamba-3 in sequence modeling — each redefining its domain.

    🔍 SAM 3 — From Segmentation to Understanding

    • Paper: “SAM 3: Segment Anything with Concepts” (ICLR 2026 submission, Meta AI)
    • Core Idea: Expands SAM 1/2 into Promptable Concept Segmentation (PCS) — segmenting all instances of a concept from text or example images.
    • Highlights: Dual Transformer with Presence Token, 52 M masks / 4 M noun phrases across 15 domains.
    • Performance: Zero-shot AP 47.0 on LVIS; SAM 3 Agent (w/ Gemini-2.5 + Llama 3.2) reaches gIoU 73.8 on ReasonSeg.

    A leap from “segmenting objects” to “comprehending concepts.”

    ⚙️ Mamba-3 — Beyond Transformers

    • Paper: “Mamba-3: Improved Sequence Modeling using State Space Principles”
    • Core Idea: Advances the Mamba family with more stable, expressive State Space Models (SSMs) for efficient long-context modeling.
    • Highlights: New discretization schemes, complex dynamics, and MIMO updates; sub-quadratic inference with Transformer-level accuracy.

    A shift from scaling attention to engineering efficiency.


    🧠 Deeper Reflection: ICLR as a Mirror of AI Maturity

    Trend ICLR 2026 Signal Long-term Implication
    From scaling to specializing LLM fine-tuning, domain models Efficiency, interpretability focus
    AI for Science Physics, bio, neuro papers ↑ “AI as a discovery tool”
    Trustworthy AI Safety & alignment 7.7% Security becomes default design
    Data-centric ML Benchmark & dataset 7.6% Data quality over quantity
    Interdisciplinary ML Cognitive + social modeling “Human–AI co-evolution”

    ICLR 2026 is the turning point where machine learning enters its post-youth stage — moving from curiosity to responsibility, from scaling to systematization.


    🤖 The Reviewer’s Dilemma — and Co-Pilot Tools

    With tens of thousands of submissions and binary scores, review overload is inevitable.
    This is where reviewer-assist platforms like cspaper.org step in — not as replacements, but as meta-evaluators offering summarized signals, topic clustering, and citation context to help reviewers prioritize intelligently.

    Future reviews may evolve into a co-pilot model, where humans handle interpretation and ethical judgment, and AI assists in reproducibility checks and bias detection.


    🌍 The Big Picture: Beyond Acceptance, Toward Accountability

    ICLR 2026 isn’t just a conference — it’s a snapshot of the AI civilization process:

    • Foundation models as universal infrastructure
    • Science as the next AI frontier
    • Agents as emergent entities
    • Review as collective cognition

    “The competition is no longer about size — it’s about efficiency, reliability, and responsibility.”

    As we enter 2026, ICLR stands as both a scientific milestone and a stress test for how the research community governs its own exponential growth.

    1 Reply Last reply
    0
    Reply
    • Reply as topic
    Log in to reply
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes


    • Login

    • Don't have an account? Register

    • Login or register to search.
    © 2025 CSPaper.org Sidekick of Peer Reviews
    Debating the highs and lows of peer review in computer science.
    • First post
      Last post
    0
    • Categories
    • CSPaper Review
    • Recent
    • Tags
    • Popular
    • Paper Copilot
    • OpenReview.net
    • Deadlines
    • CSRanking