125
submitted 1 year ago by floofloof@lemmy.ca to c/science@lemmy.ml
you are viewing a single comment's thread
view the rest of the comments
[-] henfredemars@infosec.pub 38 points 1 year ago

I've heard many papers are published to never be read by humans. It only makes sense that some portion of those papers aren't written by humans either.

I wonder what the overlap is between AI assisted papers and papers with few to no readers.

[-] PoisonedPrisonPanda@discuss.tchncs.de 24 points 1 year ago* (last edited 1 year ago)

The whole system should get ready for the 21st century.

Most of the scientists arent great writers. It does not make sense to still force them to be a good writer.

Let be fishes be good at swimming instead of climbing trees.

In a modern world where basically EVERYTHING is specialized and no generalist is alive anymore we should make use of language tools.

Hell Chatgpt writes an introduction which is fun to read instead or my overcomplicated bullshit that I would have brought up

Edit: the comment was not related to the OP but to a general chatgpt discussion.

[-] henfredemars@infosec.pub 8 points 1 year ago* (last edited 1 year ago)

It's interesting that you write this because the last place I worked focused on unspecializing by having almost everyone do every job.

In fact, they relocated across the country to save on building costs, and instead of hiring actual technical writers and office staff, they pushed the extra work down on their engineers because it's more profitable to bill for the engineering time.

I spent much of my job editing papers and I'm not even good at it while getting paid to do embedded design. It was weird. It was basically fraud but walking the fine line of technically legal.

I observed this happening multiple times throughout my career. Sometimes, inefficiency is the point in this case driven by capitalists and market forces.

[-] stockRot@lemmy.world 2 points 1 year ago

... Did you read the article? Language tools like grammarly and deepL are in use by scientists today. Copying+pasting the output of chatGPT without ever looking at it, or even using a language tool to publish thoughts that were never in your head to begin with, is the actual concern

Did you read the article?

I for sure didnt.

Thanks for highlighting that.

I was carried away by having the discussions at my university with my peers in mind.

Copying+pasting the output of chatGPT without ever looking at it, or even using a language tool to publish thoughts that were never in your head to begin with, is the actual concern

Nevertheless I dont understand why this is a concern.

The scientific standards existed decades if not already at least a century.

Those discussions are putting chatgpt in a bad light. However the fact that our scientific system was eroded and made a mockery of before the introduction of chatgpt is not highlighted.

There are still plagiarizations around and nobody cares. Mostly because of political sensitivity.

However science has failed to repel "bad actors" (intentional or unintentional) from the scene.

I dont know when. And why. But publisher have for sure something to do with it.

[-] monobot@lemmygrad.ml 0 points 1 year ago

I agree, I have no problem with people guiding chatgpt to help them write something they want and they checked it.

Generating bunch of articles even they didn't read is something else.

this post was submitted on 07 Oct 2023
125 points (96.3% liked)

Science

13216 readers
9 users here now

Subscribe to see new publications and popular science coverage of current research on your homepage


founded 5 years ago
MODERATORS