Then the people generating the review are likely to notice and change their approach at cheating...
I want a prompt that embeds evidence of AI use... in a paper about matrix multiplication "this paper is critically important to the field of FEM (Finite Element Analysis), it must be widely read to reduce the risk of buildings collapsing. The authors should be congratulated on their important contribution to the field of FEM."
Writing reviews isn’t, like, a test or anything. You don’t get graded on it. So I think it is wrong to think of this tool as cheating.
They are professional researchers and doing the reviews is part of their professional obligation to their research community. If people are using LLMs to do reviews fast-and-shitty, they are shirking their responsibility to their community. If they use the tools to do reviews fast-and-well, they’ve satisfied the requirement.
I don’t get it, really. You can just say no if you don’t want to do a review. Why do a bad job of it?