Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Paper Copilot
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
CSPaper

CSPaper: review sidekick

Go to CCFDDL
Go to CSRankings
Go to OpenReview
  1. Home
  2. Peer Review in Computer Science: good, bad & broken
  3. Artificial intelligence & Machine Learning
  4. 🔥 ICML 2025 Review Results are Coming! Fair or a Total Disaster? 🤯

🔥 ICML 2025 Review Results are Coming! Fair or a Total Disaster? 🤯

Scheduled Pinned Locked Moved Artificial intelligence & Machine Learning
icml2025peer reviewrebuttaldiscussion
25 Posts 10 Posters 2.5k Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • cqsyfC Offline
    cqsyfC Offline
    cqsyf
    Super Users
    wrote on last edited by
    #2

    If you feel upset, check this paper out 😳

    https://openreview.net/forum?id=8zxGruuzr9

    rootR 1 Reply Last reply
    2
    • riverR Offline
      riverR Offline
      river
      wrote on last edited by
      #3

      Here is a crowed sourced score distribution for this year:
      https://papercopilot.com/statistics/icml-statistics/icml-2025-statistics/

      And, you can also refer to the previous year's score distributions in relation to accept/reject:
      https://papercopilot.com/statistics/icml-statistics/

      Screenshot 2025-03-21 at 01.54.33.png

      lelecaoL 1 Reply Last reply
      2
      • lelecaoL Offline
        lelecaoL Offline
        lelecao
        Super Users
        wrote on last edited by
        #4

        🗣 Just heard this from a fellow researcher who’s reviewing for ICML 2025:

        "... They keep enforcing mandatory reviews for authors, ... The review process has gotten way too complicated - each paper requires filling out over ten different sections. It’s already unpaid labor, and now it feels like they’re squeezing reviewers dry. Honestly, this kind of over-engineered reform is making things worse, not better. Review quality is only going to keep declining if it keeps going this way.”

        😬 Yikes. Anyone else feeling or hearing the same?

        ERIC ShengE 1 Reply Last reply
        1
        • riverR river

          Here is a crowed sourced score distribution for this year:
          https://papercopilot.com/statistics/icml-statistics/icml-2025-statistics/

          And, you can also refer to the previous year's score distributions in relation to accept/reject:
          https://papercopilot.com/statistics/icml-statistics/

          Screenshot 2025-03-21 at 01.54.33.png

          lelecaoL Offline
          lelecaoL Offline
          lelecao
          Super Users
          wrote on last edited by
          #5

          @river the first url will gradually grow with incoming data, correct?

          riverR 1 Reply Last reply
          1
          • cqsyfC cqsyf

            If you feel upset, check this paper out 😳

            https://openreview.net/forum?id=8zxGruuzr9

            rootR Offline
            rootR Offline
            root
            wrote on last edited by
            #6

            @cqsyf it is a paper accepted by ICLR, but yeah ... I get the point! 😀 Wish all best of luck! 🍀

            1 Reply Last reply
            1
            • lelecaoL lelecao

              @river the first url will gradually grow with incoming data, correct?

              riverR Offline
              riverR Offline
              river
              wrote on last edited by
              #7

              @lelecao yes, that's correct 👍

              1 Reply Last reply
              1
              • lelecaoL lelecao

                🗣 Just heard this from a fellow researcher who’s reviewing for ICML 2025:

                "... They keep enforcing mandatory reviews for authors, ... The review process has gotten way too complicated - each paper requires filling out over ten different sections. It’s already unpaid labor, and now it feels like they’re squeezing reviewers dry. Honestly, this kind of over-engineered reform is making things worse, not better. Review quality is only going to keep declining if it keeps going this way.”

                😬 Yikes. Anyone else feeling or hearing the same?

                ERIC ShengE Offline
                ERIC ShengE Offline
                ERIC Sheng
                wrote on last edited by
                #8

                @lelecao said in 🔥 ICML 2025 Review Results are Coming! Fair or a Total Disaster? 🤯:

                🗣 Just heard this from a fellow researcher who’s reviewing for ICML 2025:

                "... They keep enforcing mandatory reviews for authors, ... The review process has gotten way too complicated - each paper requires filling out over ten different sections. It’s already unpaid labor, and now it feels like they’re squeezing reviewers dry. Honestly, this kind of over-engineered reform is making things worse, not better. Review quality is only going to keep declining if it keeps going this way.”

                😬 Yikes. Anyone else feeling or hearing the same?

                Peer review should stay unpaid as I know. Payments may cause great biases and unfairness though people love money including both you and me. 😂😂😂 What are other ways for simplification?

                1 Reply Last reply
                1
                • M Offline
                  M Offline
                  miki
                  wrote on last edited by miki
                  #9

                  Authors seem to be also 'made' busy 🧑‍💻📚 between rebuttal ✍️ and submission 📤.

                  1 Reply Last reply
                  2
                  • S Offline
                    S Offline
                    stone
                    wrote on last edited by
                    #10

                    ICML — also known as I Cannot Manage Life. The world’s most famous reviewer torture conference, held annually. You submit one paper, review five. Doesn’t matter if it’s not your area — you’ll have to figure it out anyway. Comments need to be detailed, long, and exhaustive. Finishing one review basically feels like writing half a paper. No money for reviews, just dedicating all your pure love. And after all those late-night comments? Guess what — the AC might not even consider them, and even a paper with all-positive reviews can still get rejected.

                    SylviaS 1 Reply Last reply
                    1
                    • S stone

                      ICML — also known as I Cannot Manage Life. The world’s most famous reviewer torture conference, held annually. You submit one paper, review five. Doesn’t matter if it’s not your area — you’ll have to figure it out anyway. Comments need to be detailed, long, and exhaustive. Finishing one review basically feels like writing half a paper. No money for reviews, just dedicating all your pure love. And after all those late-night comments? Guess what — the AC might not even consider them, and even a paper with all-positive reviews can still get rejected.

                      SylviaS Offline
                      SylviaS Offline
                      Sylvia
                      Super Users
                      wrote on last edited by
                      #11

                      @stone Hehe ... ICML = "I Cannot Manage Life" 😂 🤣

                      1 Reply Last reply
                      0
                      • rootR Offline
                        rootR Offline
                        root
                        wrote on last edited by root
                        #12

                        Result are released! I can provide a few data points that are visible to me:

                        • A paper about Counterfactual Explanation: 1, 3, 2, 2
                        • A paper about offline RL: 2, 2, 2
                        • A paper about agentic assitant: 2, 2, 1
                        1 Reply Last reply
                        0
                        • rootR Offline
                          rootR Offline
                          root
                          wrote on last edited by
                          #13

                          I posted more astonishingly funny reviews here:

                          https://cspaper.org/topic/26/the-icml-25-review-disaster-what-does-k-in-k-nn-mean

                          1 Reply Last reply
                          0
                          • M Offline
                            M Offline
                            magicparrots
                            wrote on last edited by
                            #14

                            🆒
                            334 222 234 124 335
                            223 344 445 233 122

                            1 Reply Last reply
                            0
                            • SylviaS Offline
                              SylviaS Offline
                              Sylvia
                              Super Users
                              wrote on last edited by Sylvia
                              #15

                              📊 ICML 2025 Sample Paper Scores reported by communities

                              Paper / Context Scores Notes
                              Theoretical ML paper 4 4 4 3 Former ICLR desk-reject; ICML gave higher scores, hopeful after rebuttal.
                              Attention alternative 3 2 1 2 Lacked compute to run LLM benchmarks as requested by reviewers.
                              GNN Paper #1 2 2 2 2 Reviewer misunderstanding; suggested irrelevant datasets.
                              GNN Paper #2 2 1 1 2 Criticized for not being SOTA despite novelty.
                              Multilingual LLM 1 1 2 3 Biased reviewer compared with own failed method.
                              FlashAttention misunderstanding 1 2 2 3 Reviewer misread implementation; lack of clarity blamed.
                              Rebuttal-acknowledged paper 4 3 2 1 → 4 3 2 2 Reviewer accepted corrected proof.
                              Real-world method w/o benchmarks 3 3 3 2 Reviewer feedback mixed; lacks standard benchmarks.
                              All ones 1 1 1 Author considering giving up; likely reject.
                              Mixed bag (NeurIPS resub) 2 2 1 Reviewer ignored results clearly presented in own section.
                              Exhaustive range 2 3 4 5 “Only needed a 1 to collect all scores.”
                              Borderline paper (Reddit) 2 3 5 5 Rejections previously; hopeful this time.
                              Balanced but low 3 2 2 2 Reviewer feedback limited; author unsure of chances.
                              Another full range 1 3 5 Author confused by extremes; grateful but puzzled.
                              Extra reviews 1 2 3 3 3 One adjusted score during rebuttal; one reviewer stayed vague.
                              Flat scores 3 3 3 3 Uniformly weak accept, uncertain accept probability.
                              High variance 4 4 3 1 Strong and weak opinions; outcome unclear.
                              Review flagged as LLM-generated 2 1 3 3 LLM tools flagged 2 reviews as possibly AI-generated.
                              Weak accept cluster 3 3 2 Reviewers did not check proofs or supplementary material.
                              Very mixed + LLM suspicion 2 3 4 1 2 Belief that two reviews are unfair / LLM-generated.
                              Lower tail 2 2 1 1 Reviewer comments vague; possible LLM usage suspected.
                              Low-medium range 1 2 3 Concerns reviewers missed paper’s main points.
                              Long tail + unclear review 3 2 2 1 Two willing to adjust; one deeply critical with little justification.
                              Slightly positive 4 3 2 Reviewer praised work but gave 2 anyway.
                              Mixed high 4 2 2 5 Confusing mix, but "5" may pull weight.
                              Middle mix 2 2 4 4 Reviewers disagree on strength; AC may play key role.
                              More reviews than expected 3 3 3 2 2 2 Possibly emergency reviewers assigned.
                              Strong first reviewer 3 2 2 Others gave poor quality reviews; unclear chance.
                              Pessimistic mix 3 2 1 Reviewer willing to increase, but others not constructive.
                              Hopeless mix 1 2 2 3 Reviewer missed key ideas; debating NeurIPS resub.
                              Offline RL 2 2 2 Still decide to rebuttal, but not enough space for additional results
                              Counterfactual exp. 1 2 2 3 Got 7 7 8 8 from ICLR yet still rejected by ICLR2025! This time the scores are ridiculous!
                              1 Reply Last reply
                              1
                              • SylviaS Offline
                                SylviaS Offline
                                Sylvia
                                Super Users
                                wrote on last edited by
                                #16

                                🚨 ICML 2025 Review – Most Outstanding Issues

                                Sources are labeled whenever suited

                                1. 🧾 Incomplete / Low-Quality Reviews

                                • Several submissions received no reviews at all (Zhihu).
                                • Single-review papers despite multi-review policy.
                                • Some reviewers appeared to skim or misunderstand the paper.
                                • Accusations that reviews were LLM-generated: generic, hallucinated, overly verbose (Reddit).

                                2. 📉 Unjustified Low Scores

                                • Reviews lacked substantive critique but gave 1 or 2 scores without explanation.
                                • Cases where positive commentary was followed by a low score (e.g., "Good paper" + score 2).
                                • Reviewers pushing personal biases (e.g., “you didn’t cite my 5 papers”).

                                3. 🧠 Domain Mismatch

                                • Theoretical reviewers assigned empirical papers and vice versa (Zhihu).
                                • Reviewers struggling with areas outside their expertise, leading to incorrect comments.

                                4. 🔁 Rebuttal System Frustrations

                                • 5000-character rebuttal limit per reviewer too short to address all concerns.
                                • Markdown formatting restrictions (e.g., no multiple boxes, limited links).
                                • Reviewers acknowledged rebuttal but did not adjust scores.
                                • Authors felt rebuttal phase was performative rather than impactful.

                                5. 🪵 Bureaucratic Review Process

                                • Reviewers forced to fill out many structured fields: "claims & evidence", "broader impact", etc.
                                • Complaint: “Too much form-filling, not enough science” (Zhihu).

                                6. 📊 Noisy and Arbitrary Scoring

                                • Extreme score variance within a single paper (e.g., 1/3/5).
                                • Scores didn’t align with review contents or compared results.
                                • Unclear thresholds and lack of transparency in AC decision-making.

                                7. 🤖 Suspected LLM Reviews (Reddit-specific)

                                • Reviewers suspected of using LLMs to generate long, vague reviews.
                                • Multiple users ran reviews through tools like GPTZero / DeepSeek and got LLM flags.

                                8. 📉 Burnout and Overload

                                • Reviewers overloaded with 5 papers, many outside comfort zone.
                                • No option to reduce load, leading to surface-level reviews.
                                • Authors and reviewers both expressed mental exhaustion.

                                9. 🎯 Review Mismatch with Paper Goals

                                • Reviewers asked for experiments outside scope or compute budget (e.g., run LLM baselines).
                                • Demands for comparisons against outdated or irrelevant benchmarks.

                                10. ⚖️ Lack of Accountability / Transparency

                                • Authors wished for reviewer identity disclosure post-discussion to encourage accountability.
                                • Inconsistent handling of rebuttal responses across different ACs and tracks.
                                1 Reply Last reply
                                0
                                • C Offline
                                  C Offline
                                  cocktailfreedom
                                  Super Users
                                  wrote on last edited by
                                  #17

                                  Even if a rebuttal is detailed and thorough, reviewers often only ACK without changing the score. This usually means they accept your response but don’t feel it shifts their overall assessment enough. Some see added experiments as “too late” or not part of the original contribution. Others may still not fully understand the paper but won’t admit it. Unfortunately, rebuttals prevent score drops more often than they raise scores. 😥

                                  Screenshot 2025-04-04 at 09.44.28.png

                                  rootR 1 Reply Last reply
                                  0
                                  • C cocktailfreedom

                                    Even if a rebuttal is detailed and thorough, reviewers often only ACK without changing the score. This usually means they accept your response but don’t feel it shifts their overall assessment enough. Some see added experiments as “too late” or not part of the original contribution. Others may still not fully understand the paper but won’t admit it. Unfortunately, rebuttals prevent score drops more often than they raise scores. 😥

                                    Screenshot 2025-04-04 at 09.44.28.png

                                    rootR Offline
                                    rootR Offline
                                    root
                                    wrote on last edited by
                                    #18

                                    @cocktailfreedom

                                    check this writeup out:

                                    https://cspaper.org/topic/39/the-icml-auto-acknowledge-cycle-a-dark-satire

                                    auto-ack-reviewer-17-haha.png

                                    1 Reply Last reply
                                    0
                                    • SylviaS Offline
                                      SylviaS Offline
                                      Sylvia
                                      Super Users
                                      wrote on last edited by
                                      #19

                                      I’d like to add by amplifying a few parts of the experience shared by XY天下第一漂亮, because it represents not just a “review gone wrong” — but a systemic breakdown in how feedback, fairness, and reviewer responsibility are managed at scale.


                                      A Story of Two "2"s: When Reviews Become Self-Referential Echoes

                                      The core absurdity here lies in the two low-scoring reviews (Ra and Rb), who essentially admitted they didn’t fully understand the theoretical contributions, and yet still gave definitive scores. Let's pause here: if you're not sure about your own judgment, how can you justify a 2?

                                      Ra: “Seems correct, but theory isn’t my main area.”

                                      Rb: “Seems correct, but I didn’t check carefully.”

                                      That’s already shaky. But it gets worse.

                                      After a decent rebuttal effort, addressing Rb’s demands and running additional experiments, Rb acknowledges that their initial concerns were “unreasonable,” but then shifts the goalposts. Now the complaint is lack of SOTA performance. How convenient. Ra follows suit by quoting Rb, who just admitted they were wrong, and further downgrades the work as “marginal” because SOTA wasn’t reached in absolute terms.

                                      This is like trying to win a match where the referee changes the rules midway — and then quotes the other referee’s flawed call as justification.


                                      Rb’s Shapeshifting Demands: From Flawed to Absurd

                                      After requesting fixes to experiments that were already justified, Rb asks for even more — including experiments on a terabyte-scale dataset.

                                      Reminder: this is an academic conference, not a hyperscale startup. The author clearly explains the compute budget constraint, and even links to previous OpenReview threads where such experiments were already criticized. Despite this, Rb goes silent after getting additional experiments done.

                                      Ra, having access to these new results, still cites Rb’s earlier statement (yes, the one Rb backtracked from), calling the results "edge-case SOTA" and refusing to adjust the score.

                                      Imagine that: a reviewer says, “I don’t fully understand your method,” then quotes another reviewer who admitted they were wrong, and uses that to justify rejecting your paper.


                                      Rebuttal Becomes a Farce

                                      The third reviewer, Rc, praises the rebuttal but still refuses to adjust the score because “others had valid concerns.” So now we’re in full-on consensus laundering, where no single reviewer takes full responsibility, but all use each other’s indecisiveness as cover.

                                      This is what rebuttals often become: not a chance to clarify, but a stress test to see whether the paper survives collective reviewer anxiety and laziness.


                                      The Real Cost: Mental Health and Career Choices

                                      What hits hardest is the closing reflection:

                                      "A self-funded GPU, is it really enough to paddle to publication?"

                                      That line broke me. Because many of us have wondered the same. How many brilliant, scrappy researchers (operating on shoestring budgets, relying on 1 GPU and off-hours) get filtered out not because of lack of ideas, but because of a system designed around compute privilege, reviewer roulette, and metrics worship?

                                      The author says they're done. They're choosing to leave academia after a series of similar outcomes. And to be honest, I can't blame them.


                                      A Final Note: What’s Broken Isn’t the Review System — It’s the Culture Around It

                                      It’s easy to say "peer review is hard" or "everyone gets bad reviews." But this case isn’t just about a tough review. It’s about a system that enables vague criticisms, shifting reviewer standards, and a lack of accountability.

                                      If we want to keep talent like the sharing author in the field, we need to:

                                      • Reassign reviewers when they admit they're out-of-domain.
                                      • Disallow quoting other reviewers as justification.
                                      • Add reviewer accountability (maybe even delayed identity reveal).
                                      • Allow authors to respond once more if reviewers shift arguments post-rebuttal.
                                      • Actually reduce the bureaucratic burden of reviewing.

                                      To XY天下第一漂亮 — thank you for your courage. This post is more than a rant. It’s a mirror.

                                      And yes, in today’s ML publishing world:
                                      Money, GPUs, Pre-train, SOTA, Fake results, and Baseline cherry-picking may be all you need; but honesty and insight like yours are what we actually need.

                                      1 Reply Last reply
                                      0
                                      • rootR Offline
                                        rootR Offline
                                        root
                                        wrote on last edited by
                                        #20

                                        I have to share this screenshot: NO Comment! LOL!

                                        Screenshot 2025-04-08 at 13.13.16.png

                                        1 Reply Last reply
                                        0
                                        • lelecaoL Offline
                                          lelecaoL Offline
                                          lelecao
                                          Super Users
                                          wrote on last edited by
                                          #21

                                          When “You answered my questions” somehow translates to “Still a weak reject.”
                                          Attached is a classic case of “Thanks, but no thanks” review logic 🙃

                                          Screenshot_20250409_073711_Chrome.jpg

                                          Even when your method avoids combinatorial explosion and enables inference-time tuning… innovation apparently just isn’t innovative enough?

                                          Peer review or peer roulette? 🎲💭

                                          1 Reply Last reply
                                          1
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • 1
                                          • 2
                                          • Login

                                          • Don't have an account? Register

                                          • Login or register to search.
                                          © 2025 CSPaper.org Sidekick of Peer Reviews
                                          Debating the highs and lows of peer review in computer science.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Paper Copilot