318
The New York Times sues OpenAI and Microsoft for copyright infringement
(edition.cnn.com)
This is a most excellent place for technology news and articles.
My question is how is an AI reading a bunch of articles any different from a human doing it. With this logic no one would legally be able to write an article as they are using bits of other peoples work they read that they learnt to write a good article with.
They are both making money with parts of other peoples work.
Let me ask you this: when have you ever seen ChatGPT cite its sources and give appropriate credit to the original author?
If I were to just read the NYT and make money by simply summarizing articles and posting those summaries on my own website without adding anything to it like my own commentary and without giving credit to the author, that would rightfully be considered plagiarism.
This is a really interesting conundrum though. I would argue that AI isn't capable of original thought the way that humans are and therefore AI creators must provide due compensation to the authors and artists whose data they used.
AI is only giving back some amalgamation of words and concepts that it has been trained on. You might say that humans do the same, but that isn't exactly true. The human brain is a funny thing. It can forget, it can misremember. It can manipulate. It can exaggerate. It can plan. It can have irrational or emotional responses. AI can't really do those things on its own. It's just mimicking human behavior at best.
Most importantly to me though, AI is not capable of spontaneous thought. It is only capable of providing information that it has been trained on and only when prompted.
There is evidence to suggest some LLM's have the ability to produce original outputs, such as DeepMind's solution to the cap set problem.
https://www.nature.com/articles/s41586-023-06924-6
On the other hand LLM's have some incredible text compression abilities
https://arxiv.org/abs/2308.07633
I'm pretty sure there is copyright infringement going on by the letter of the law. But I also think the world would be better off if copyright laws were a bit more loose. Not wild-west anything-goes libertarianism, but more open than the current state.
I tend to agree with your last point, especially because of the way the system has been bastardized over the years. What started out as well intentioned legislation to ensure that authors and artists maintain control over their work has become a contentious and litigious minefield that barely protects creators.
Bing chat now does that by default. Normally you have to prompt that manually.
No. It would be considered journalism. If you read the news a bit, you will find that they reference the output of other news corporations quite a bit. If your preferred news source does not do that, then they simply don't cite their sources.
Prompting for a source wouldn't satisfy me until I could trust that the AI wasn't hallucinating. After all, if GPT can make up facts about things like legal precedent or well documented events, why would I trust that its citations are legitimate?
And if the suggestion is that the person asking for the information double check the cited sources, maybe that's reasonable to request, but it somewhat defeats the original purpose.
Bing might be doing things differently though, so you might be right in your assessment on that front. I haven't played with their AI yet.
You did ask if ChatGPT had ever sighted sources. Bing uses it and besides, you can ask for that manually.
Whether it defeats the purpose depends on your original purpose.