It doesn't worry me, so much as seem impractical. Writing a proper academic paper isn't the work of an afternoon.
Yeah, to be more precise in my response to
@Belen--I've worked in cultures where the main author writes the paper in word (etc.) and then sends it to everyone. And I've worked in ones where everyone collaborates at once in some kind of online environment. For me, the offline version worked better. Part of that is just more clarity and precision of thought. But some of it is because working online is a pain--you need an internet connection, and editing, especially figures, is slower. Even moving an image around on the page is a headache with word online. And if you want to write on a plane, or in a bar with no wifi (both things I do), then the connectivity is a problem.
And new authors? How do they get accepted?
Yeah that's a major issue. Relying on established reputation is going to favor people in more established institutions with all the issues for inclusivity and accessibility that implies.
Some quick searching indicates to me that ChatGPT currently hallucinates at a rate of 33% to 79% depending on the type of test.
In cases where you are depending on accurate presentation of information, hallucination is a failure.
Can you claim it does something well, when it fails 33% of the time? I mean, unless you are a baseball player?
I don't think that's a very nuanced approach to the topic. Like the chess example--we judge chess playing computers based on how good they are at chess. But LLMs are treated as this kind of everything engine, and judged by how well they perform at any task. That doesn't seem fair to me.
I get that, to some extent, that criticism is a response to hype and AI marketing cycles which claim LLMs are general intelligence or imply they are good at everything. And on those grounds, I'm sympathetic to that critique.
But I think overly focusing on that criticism can cause us to miss the very specific and structured ways where LLMs are useful. Maybe not useful enough to justify the expense and the environmental costs--I'm also sympathetic to points you've made in that regard. But, in the specific cases where they are beneficial--translation, programming, search, editing, brainstorming--I think it's premature to write off everything as slop.