That’s not actually the abstract; it’s a piece from the discussion that someone pasted nicely with the first page in order to name and shame the authors. I looked at it in depth when I saw this circulate a little while ago.
Ah, that makes more sense. I looked up the original abstract and indeed it looks more like what you’d expect (hard to comprehend for someone that’s not in the field).
Though to clarify (for others reading this) they still did use generative AI to (help?) write the paper, which is only part of why it was withdrawn.
I’ve recently been watching a lot of videos on prominent cases of fraud and malpractice like Francesca Gino, Claudine Gay, Hwang Woo-suk, etc., which prompted me to start reading more into meta-research as well, and now I’m basically paranoid about every paper I read. There’s so much shady shit going on…
Ooh I mean not help. It’ll make it much worse. Particularly with the social sciences. Which were already pretty fuX0r3d anyway due to the whole “your emotions equal this number” thing.
Many journals are absolute garbage that will accept anything. Keep that in mind the next time someone links a study to prove a point. You have to actually read the thing and judge the methodology to know if their conclusions have any merits.
Full disclosure: I don’t intend to be condescending.
Research Methods during my graduate studies forever changed the way I interpret just about any claim, fact, or statement. I’m obnoxiously skeptical and probably cynical, to be honest. It annoys the hell out of my wife but it beats buying into sensationalist headlines and miracle research. Then you get into the real world and see how data gets massaged and thrown around haphazardly…believe very little of what you see.
I have this problem too. My wife gets so annoyed at things because I question things I notice as biases or statistical irregularities instead of just accepting that they knee what they were doing. I have tried to explain it to her. Skepticism is not dismissal and it is not saying I am smarter than them, it is recognizing that they are human and that I may be more proficient in one spot they made a mistake than they were.
I will acknowledge that the lay need to stop trying to argue with scientists because “they did their own research”, but the actually informed and educated need to do a better job of calling each other out.
We are in top dystopia mode right now. Students have AI write articles that are proofread and edited by AI, submitted to automated systems that are AI vetted for publishing, then posted to platforms where no one ever reads the articles posted but AI is used to scrape them to find answers or train all the other AIs.
The search giant doesn’t just face new competition from ChatGPT and other upstarts. It also has to keep AI-powered SEO from damaging its results.
More or less the same phenomenon of signal pollution:
“Google is shifting its responsibility for maintaining the quality of results to moderators on Reddit, which is dangerous,” says Ray of Amsive. Search for “kidney stone pain” and you’ll see Quora and Reddit ranking in the top three positions alongside sites like the Mayo Clinic and the National Kidney Foundation. Quora and Reddit use community moderators to manually remove link spam. But with Reddit’s traffic growing exponentially, is a human line of defense sustainable against a generative AI bot army?
We’ll end up using year 2022 as a threshold for reference criteria. Maybe not entirely blocked, but like a ratio… you must have 90% pre-2022 and 10% post-2022.
Perhaps this will spur some culture shift to publish all the data, all the notes, everything - which will be great to train more AI on. Or we’ll get to some type of anti-AI or anti-crawler medium.
Hold up. That actually got through to publishing??
It’s because nobody was there to highlight the text for them.
The entire abstract is AI. Even without the explicit mention in one sentence, the rest of the text should’ve been rejected as nonspecific nonsense.
That’s not actually the abstract; it’s a piece from the discussion that someone pasted nicely with the first page in order to name and shame the authors. I looked at it in depth when I saw this circulate a little while ago.
Ah, that makes more sense. I looked up the original abstract and indeed it looks more like what you’d expect (hard to comprehend for someone that’s not in the field).
Though to clarify (for others reading this) they still did use generative AI to (help?) write the paper, which is only part of why it was withdrawn.
Maybe a big red circle around the entire abstract would have helped
It’s Elsevier, so this probably isn’t even the lowest quality article they’ve published
yea lol
https://www.sciencedirect.com/science/article/pii/S1930043324004096
I’ve recently been watching a lot of videos on prominent cases of fraud and malpractice like Francesca Gino, Claudine Gay, Hwang Woo-suk, etc., which prompted me to start reading more into meta-research as well, and now I’m basically paranoid about every paper I read. There’s so much shady shit going on…
Yep. And AI will totally help.
Ooh I mean not help. It’ll make it much worse. Particularly with the social sciences. Which were already pretty fuX0r3d anyway due to the whole “your emotions equal this number” thing.
Many journals are absolute garbage that will accept anything. Keep that in mind the next time someone links a study to prove a point. You have to actually read the thing and judge the methodology to know if their conclusions have any merits.
Full disclosure: I don’t intend to be condescending.
Research Methods during my graduate studies forever changed the way I interpret just about any claim, fact, or statement. I’m obnoxiously skeptical and probably cynical, to be honest. It annoys the hell out of my wife but it beats buying into sensationalist headlines and miracle research. Then you get into the real world and see how data gets massaged and thrown around haphazardly…believe very little of what you see.
I have this problem too. My wife gets so annoyed at things because I question things I notice as biases or statistical irregularities instead of just accepting that they knee what they were doing. I have tried to explain it to her. Skepticism is not dismissal and it is not saying I am smarter than them, it is recognizing that they are human and that I may be more proficient in one spot they made a mistake than they were.
I will acknowledge that the lay need to stop trying to argue with scientists because “they did their own research”, but the actually informed and educated need to do a better job of calling each other out.
A good tactic, though not perfect, is to look at the journal impact factor.
https://en.m.wikipedia.org/wiki/Impact_factor
We are in top dystopia mode right now. Students have AI write articles that are proofread and edited by AI, submitted to automated systems that are AI vetted for publishing, then posted to platforms where no one ever reads the articles posted but AI is used to scrape them to find answers or train all the other AIs.
More or less the same phenomenon of signal pollution:
We’ll end up using year 2022 as a threshold for reference criteria. Maybe not entirely blocked, but like a ratio… you must have 90% pre-2022 and 10% post-2022.
Perhaps this will spur some culture shift to publish all the data, all the notes, everything - which will be great to train more AI on. Or we’ll get to some type of anti-AI or anti-crawler medium.