Guillaume Cabanac Last week, an environmental journal published a paper on the use of renewable energy in cleaning up contaminated land. To read it, you would have to pay 40 euros. But you still wo…
This is the key - it does not create, it can only copy. Which is good enough to fool us - there’s enough stuff to copy out there that you can spend your whole life copying other people and nobody will ever notice you’re not actually creating anything new. What’s more, you’ll probably come across as pretty clever. But you’re not creating anything new.
For me, this poses an existential threat to academia. It might halt development in the field without researchers even noticing: Their words look fine, as if they had thought it through, and they of course read it to make sure it’s logically consistent. However, the creative force is gone. Nothing new will come under the sun - the kind of new thoughts that can only be made by creative humans thinking new thoughts that have never been put on paper before.
If we give up that, what’s even the point of doing science in the first place.
Using ChatGPT to help write parts of the text in the same way you’d use a grammar- or spell-checker (e.g. if English isn’t your first language) after you’ve finished the experiments
Using ChatGPT to write a paper without even doing any experiments
Clearly the second is academic misconduct. The first one is a lot more defensible.
Yes, absolutely. But I still think it has its dangers.
Using it to write the introduction doesn’t change the substance of the paper, yet it does provide the framework for how the reader interprets it, and also often decides whether it’ll be read at all.
Maybe worse, I find that it’s oftem in the painful writing and rewriting of the introduction and conclusion that I truly understand my own contribution - I’ve done the analysis and all that, but in forcing myself to think about the relevance for the field and the reader I also bring myself to better understand what the paper means in a deeper sense. I believe this kind of deep thinking at the end of the process is incredibly valuable, and it’s what I’m afraid we might be losing with AI.
This is the key - it does not create, it can only copy.
I have asked ChatGPT to write poetry on subjects that I know with great certainty have never had poems written about them.
You can of course shuffle around the meanings of “create” and “copy” to try to accommodate that, but eventually you end up with a “copying” process that’s so flexible and malleable that it might as well be creativity. It’s not like what comes out of human brains isn’t based on stuff that went into them earlier either.
This is the key - it does not create, it can only copy. Which is good enough to fool us - there’s enough stuff to copy out there that you can spend your whole life copying other people and nobody will ever notice you’re not actually creating anything new. What’s more, you’ll probably come across as pretty clever. But you’re not creating anything new.
For me, this poses an existential threat to academia. It might halt development in the field without researchers even noticing: Their words look fine, as if they had thought it through, and they of course read it to make sure it’s logically consistent. However, the creative force is gone. Nothing new will come under the sun - the kind of new thoughts that can only be made by creative humans thinking new thoughts that have never been put on paper before.
If we give up that, what’s even the point of doing science in the first place.
There’s a difference between:
Using ChatGPT to help write parts of the text in the same way you’d use a grammar- or spell-checker (e.g. if English isn’t your first language) after you’ve finished the experiments
Using ChatGPT to write a paper without even doing any experiments
Clearly the second is academic misconduct. The first one is a lot more defensible.
Yes, absolutely. But I still think it has its dangers.
Using it to write the introduction doesn’t change the substance of the paper, yet it does provide the framework for how the reader interprets it, and also often decides whether it’ll be read at all.
Maybe worse, I find that it’s oftem in the painful writing and rewriting of the introduction and conclusion that I truly understand my own contribution - I’ve done the analysis and all that, but in forcing myself to think about the relevance for the field and the reader I also bring myself to better understand what the paper means in a deeper sense. I believe this kind of deep thinking at the end of the process is incredibly valuable, and it’s what I’m afraid we might be losing with AI.
I have asked ChatGPT to write poetry on subjects that I know with great certainty have never had poems written about them.
You can of course shuffle around the meanings of “create” and “copy” to try to accommodate that, but eventually you end up with a “copying” process that’s so flexible and malleable that it might as well be creativity. It’s not like what comes out of human brains isn’t based on stuff that went into them earlier either.