That's nonsense. I can spend the whole day creating false papers on AI, then feeding it back to another AI to check its "quality". Is this making the paper to be "remembered" by AI? If yes, then we have deeper problems and we shouldn't be using AI to do anything related to science.
I think he means WRT the leaking issue that we were discussing.
If someone is just, like, working chatGPT up to automatically review papers, or using Grok to automatically review grants with minimal human intervention, that’d obviously be a totally nuts thing to do. But who would do such a thing, right?