I'd like to share my experiment with using ChapGPT to help write an article on English Wikipedia:
https://en.wikipedia.org/wiki/Artwork_title
You can see an explanation of the process here, your comments are welcome:
https://en.wikipedia.org/wiki/Talk:Artwork_title#Use_of_ChatGPT
Thanks, Richard (User:Pharos)
On Tue, Dec 20, 2022 at 1:45 PM ZhaoFJx zhaofjx@gmail.com wrote:
I'd be curious about copyright issues though, as it's licensed on Github https://github.com/acheong08/ChatGPT/blob/main/LICENSE. It may indeed be infringing copyright, since he may require attribution. However, I agree with The Cunctator that the community won't do anything about it.
Incidentally, I'm a little worried that he might grab broken content that wasn't rolled back. After all - the current artificial intelligence can not reach the level of human thinking
Sincerely, ZhaoFJx _______________________________________________ Wikimedia-l mailing list -- wikimedia-l@lists.wikimedia.org, guidelines at: https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and https://meta.wikimedia.org/wiki/Wikimedia-l
Anders Wennersten mail@anderswennersten.se 于2022年12月19日周一 08:52写道:
I think we should not underestimate what this could evolve into. We thrived because our readers find us "good enough" when it comes to finding facts, not the ultimate source.
And the software learns by reading, and can (and have done so) Wikipedia, Wikidata etc and represent our data in its own syntax och present it in a way so it is not a direct copy. Perhaps data will be a bit delayed to the actual content in Wikipedia, but so what - good enough?
Anders
Den 2022-12-19 kl. 14:26, skrev Gnangarra:
AI simply cant descriminate between good research and faked research, for any outcome it must provide all of its sources whether they are from Wikipedia, Wikidata, WikiCommons, WikiSource or some other place. Otherwise it will answer yes to some asking if the world is flat because it'll seek out that answer and find all the nonsense that has been produced.
On Mon, 19 Dec 2022 at 06:02, Erik Moeller eloquence@gmail.com wrote:
On Sun, Dec 11, 2022 at 5:55 AM Anders Wennersten mail@anderswennersten.se wrote:
ChatGPT is now making headlines more or less every day and I perceive them to try to position themself av the "next" google.
I suspect OpenAI will continue to focus on generative applications (images, code, text for purposes such as copywriting, eventually music/video) and won't attempt to compete with Google directly, but we'll see. Currently GPT-3.5 (which ChatGPT is based on) is very prone to generating nonsensical answers, citations to works that don't exist, etc. But it is pretty cool if you keep its limitations in mind--for example, it's quite good at bootstrapping small scripts in various programming languages (with mistakes and idiosyncrasies).
Google has one of the largest AI research programs on the planet, they just are extremely conservative about letting anyone try their models (due to reputational concerns, e.g., that generative AI will spit out racist output within about 30 seconds of people poking its guardrails). This blog post from September is instructive about the direction they're taking with what's called retrieval-augmented generation; see the paper linked from the post for details:
https://www.deepmind.com/blog/building-safer-dialogue-agents (DeepMind is part of Google)
That is likely to yield significantly more accurate answers than what ChatGPT is doing, and is difficult to replicate for folks like OpenAI without being dependent on the search APIs of big search companies. It's worth noting that Google has also started to incorporate language model tooling into how it's presenting search results (e.g., summarizing or highlighting different parts of a website to make the result snippet more useful).
A retrieval-augmented approach that leverages Wikidata could IMO be quite powerful and could be a useful research program for Wikimedia to pursue, be it independently or in partnership with others. The resulting technology should of course be fully open source.
Querying Wikidata via SPARQL is currently still a bit of wizardry (and the query builder is extremely limited). To pick a completely random example not at all inspired by current events, if I wanted to see a list of journalists with Mastodon accounts & a picture, I currently have to do this:
SELECT DISTINCT ?personLabel ?mastodonName ?pic WHERE { ?person wdt:P4033 ?mastodonName ; wdt:P106 ?occupation . OPTIONAL { ?person wdt:P18 ?pic . } ?occupation wdt:P279* wd:Q1930187 . SERVICE wikibase:label { bd:serviceParam wikibase:language "en" } }
Make a small mistake (a curly brace missing) and you'll get a red error message. Forgot the * after wdt:P279? A different response set in ways that are difficult to spot or reason about.
Why can't I type "list of journalists with their picture and Mastodon account" as a natural language query? (You can try it in ChatGPT and it'll get you started, but it'll generate nonsense P/Q numbers.) If such queries could be produced reliably, it could be a very useful tool for readers as well.
Warmly, Erik _______________________________________________ Wikimedia-l mailing list -- wikimedia-l@lists.wikimedia.org, guidelines at: https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and https://meta.wikimedia.org/wiki/Wikimedia-l Public archives at https://lists.wikimedia.org/hyperkitty/list/wikimedia-l@lists.wikimedia.org/... To unsubscribe send an email to wikimedia-l-leave@lists.wikimedia.org
-- Boodarwun Gnangarra 'ngany dabakarn koorliny arn boodjera dardoon ngalang Nyungar koortaboodjar'
Wikimedia-l mailing list -- wikimedia-l@lists.wikimedia.org, guidelines at: https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and https://meta.wikimedia.org/wiki/Wikimedia-l Public archives at https://lists.wikimedia.org/hyperkitty/list/wikimedia-l@lists.wikimedia.org/... To unsubscribe send an email to wikimedia-l-leave@lists.wikimedia.org
Wikimedia-l mailing list -- wikimedia-l@lists.wikimedia.org, guidelines at: https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and https://meta.wikimedia.org/wiki/Wikimedia-l Public archives at https://lists.wikimedia.org/hyperkitty/list/wikimedia-l@lists.wikimedia.org/... To unsubscribe send an email to wikimedia-l-leave@lists.wikimedia.org
-- Wikimedia-l mailing list -- wikimedia-l@lists.wikimedia.org, guidelines at: https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and https://meta.wikimedia.org/wiki/Wikimedia-l Public archives at https://lists.wikimedia.org/hyperkitty/list/wikimedia-l@lists.wikimedia.org/... To unsubscribe send an email to wikimedia-l-leave@lists.wikimedia.org
Wikimedia-l mailing list -- wikimedia-l@lists.wikimedia.org, guidelines at: https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and https://meta.wikimedia.org/wiki/Wikimedia-l Public archives at https://lists.wikimedia.org/hyperkitty/list/wikimedia-l@lists.wikimedia.org/... To unsubscribe send an email to wikimedia-l-leave@lists.wikimedia.org