tbh i somehow didnt even realize that wikipedia is one of the few super popular sites not trying to shove ai down my throat every 5 seconds
i'm grateful now
This is a most excellent place for technology news and articles.
tbh i somehow didnt even realize that wikipedia is one of the few super popular sites not trying to shove ai down my throat every 5 seconds
i'm grateful now
Christ, I miss when I could click on an article and not be asked to sign up for it.
Oh, right! Thanks for reminding me. I tried to archive it the last time but it took forever.
Edit. There ya' go: https://archive.is/oWcIr
As I have adblock mostly because of the abuse of trackers, I understand people trying to monetize their work.
Fuck AI
He is nobody to Wikipedia now. He also failed to create a news site and a micro SNS.
He can also stick AI inside his own ass
Sit down Jimmy. Wikipedia has enough problems already, it doesn't need more to be added by AI.
I will stop donating to Wikipedia if they use AI
What's funny is that for enormous big systems with network effects we are trying to use mechanisms intended for smaller businesses, like a hot dog kiosk.
IRL we have a thing for those, it's called democracy.
In the Internet it's either anarchy or monarchy, sometimes bureaucratic dictatorship, but in that area even Soviet-style collegial rule is something not yet present.
I'm recently read that McPherson article about Unix and racism, and how our whole perception of correct computing (modularity, encapsulation, object-orientation, all the KISS philosophy even) is based on that time's changes in the society and reaction to those. I mean, real world is continuous and you can quantize it into discrete elements in many ways. Some unfit for your task. All unfit for some task.
So - first, I like the Usenet model.
Second, cryptography is good.
Third, cryptographic ownership of a limited resource is ... fine, blockchains are maybe not so stupid. But not really necessary, because one can choose between a few versions of the same article retrieved, based on web of trust or whatever else. No need to have only one right version.
Fourth, we already have a way to turn sequence of interdependent actions into state information, it's called a filesystem.
Fifth, Unix with its hierarchies is really not the only thing in existence, there's BTRON, and even BeOS had a tagged filesystem.
Sixth, interop and transparency are possible with cryptography.
Seventh, all these also apply to a hypothetical service over global network.
Eighth, of course, is that the global network doesn't have to be globally visible\addressable to operate globally for spreading data, so even the Internet itself is not as much needed as the actual connectivity over which those change messages will propagate where needed and synchronize.
Ninth, for Wikipedia you don't need as much storage as for, say, Internet Archive.
And tenth - with all these one can make a Wikipedia-like decentralized system with democratic government, based on rather primitive principles, other than, of course, cryptography involved.
(Yes, Briar impressed me.)
EDIT: Oh, about democracy - I mean technical democracy. That an event (making any change) weren't valid if not processed correctly, by people eligible for signing it, for example, and they are made eligible by a signed appointment, and those signing it are made eligible by a democratic process (signed by majority of some body, signed in turn). That's that blockchain democracy people dreamed at some point. Maybe that's not a scam. Just haven't been done yet.
How do you prevent sybil attacks without making it overly expensive to vote?
How do you use Sybil attack for a system where the initial creator signs the initial voters, and then they collectively sign elections and acceptance of new members and all such stuff?
Doesn't seem to be a problem for a system with authorized voters.
Flood them with AI-generated applicants.
So why would they accept said AI-generated applicants?
If we are making a global system, then confirmation using some nation's ID can be done, with removing fakes found out later. Like with IRL nation states. Or "bring a friend and be responsible if they are a fake". Or both at the same time.
Would every participant get to see my government-issued ID?
One can elect a small group which will and will sign its connection to something intermediate. Then only they will.
How do we know if they're doing a good job without being able to review their work?
Wikipedia already has a decades operating cost of savings.
No they don't because they blast it on inflated exec wages.
Why don't they blast execs and reduce the expenses.
jimmy wales is also the president and co-founder of fandom
to give you an idea of who that guy is
I mean, the Wikipedia page does say it was sold in 2018. Not sure how it was before but it's not surprising that it enshittified by now.
I guess in his defense it wasn't too bad before 2018, as far as I can remember. Most of the enshittification of fandom I can remember has happened since.
Oh gross
Why?
Fandom (previously Wikia) is an extremely shitty service with low-quality wikis mostly consisting of content copied from independent wikis and a terrible layout that only exists to amplify their overwhelming advertising.
my one weird trick for using fandom.com is to disable javascript for that domain.
While this is true, the majority of the wikis are not at all low quality. Some are the only ones existing for a topic. The wikis are community-based, after all.
But its easy to vandalize and is highly profit-driven. The fandom wikis are filled with ads that absolutely destroy navigation. Infamous is the video ad that scrolls you up automatically in the middle of reading once it finishes. You have to pause it to read the article with no interruption.
they captured the "niche wiki" market as wikia, then rebranded and started serving shittons of ads. the vim wiki is unusable these days because it runs like ass and looks like a gamer rgb nightmare
There's an addon for that, Indie Wiki Buddy.
It tries to redirect you to non fandom/fextralife wikis if they exist, and if not, it proxies fandom wikis through BreezeWiki which just displays the content.
And I'll take this opportunity to plug Hohser and the uBlock AI blocklist as well.
Wales’s quote isn’t nearly as bad as the byline makes it out to be:
Wales explains that the article was originally rejected several years ago, then someone tried to improve it, resubmitted it, and got the same exact template rejection again.
“It's a form letter response that might as well be ‘Computer says no’ (that article's worth a read if you don't know the expression),” Wales said. “It wasn't a computer who says no, but a human using AFCH, a helper script [...] In order to try to help, I personally felt at a loss. I am not sure what the rejection referred to specifically. So I fed the page to ChatGPT to ask for advice. And I got what seems to me to be pretty good. And so I'm wondering if we might start to think about how a tool like AFCH might be improved so that instead of a generic template, a new editor gets actual advice. It would be better, obviously, if we had lovingly crafted human responses to every situation like this, but we all know that the volunteers who are dealing with a high volume of various situations can't reasonably have time to do it. The templates are helpful - an AI-written note could be even more helpful.”
That being said, it still reeks of “CEO Speak.” And trying to find a place to shove AI in.
More NLP could absolutely be useful to Wikipedia, especially for flagging spam and malicious edits for human editors to review. This is an excellent task for dirt cheap, small and open models, where an error rate isn’t super important. Cost, volume, and reducing stress on precious human editors is. It's a existential issue that needs work.
…Using an expensive, proprietary API to give error prone yet “pretty good” sounding suggestions to new editors is not.
Wasting dev time trying to make it work is not.
This is the problem. Not natural language processing itself, but the seemingly contagious compulsion among executives to find some place to shove it when the technical extent of their knowledge is occasionally typing something into ChatGPT.
It’s okay for them to not really understand it.
It’s not okay to push it differently than other technology because “AI” is somehow super special and trendy.
Honestly, translating the good articles from other languages would improve Wikipedia immensely.
For example, the Nanjing dialect article is pretty bare in English and very detailed in Mandarin
This is the goal of Abstract Wikipedia. https://meta.wikimedia.org/wiki/Special:MyLanguage/Abstract_Wikipedia
You can do that, that's fine. As long as you can verify it is an accurate translation, so you need to know the subject matter and the target language.
But you could probably also have used Google translate and then just fine tune the output yourself. Anyone could have done that at any point in the last 10 years.
As long as you can verify it is an accurate translation
Unless the process has changed in the last decade, article translations are a multi-step process, which includes translators and proof-readers. It's easier to get volunteer proof-readers than volunteer translators. Adding AI for the translation step, but keeping the proof-reading step should be a great help.
But you could probably also have used Google translate and then just fine tune the output yourself. Anyone could have done that at any point in the last 10 years.
Have you ever used Google translate? Putting an entire Wikipedia article through it and then "fine tuning" it would be more work than translating it from scratch. Absolutely no comparison between Google translate and AI translations.
Putting an entire Wikipedia article through it and then "fine tuning" it would be more work than translating it from scratch.
That depends on if you are capable of translating the language if you don't know the language then the translator will give you a good start.
Google translate is horrendously bad at Korean, especially with slang and accidental typos. Like nonsense bad.
Same in Hungarian, machine translation still often gives hilariously bad results. It's especially prone to mixing up formal and informal 'you' within the same paragraph, something which humans never do. At least it's easy to tell when a website is one of those 'auto-translated to 30 languages' content mill.
I recently have edited a small wiki page that was obviously written by someone that wasn’t proficient in English. I used AI to just reword what was already written and then I edited the output myself. It did a pretty good job. It was a page about some B-list Indonesian actress that I just stumbled upon and I didn’t want to put time and effort into it but the page really needed work done.
Wikipedia's translation tool for porting articles between languages currently uses google translate so I could see an LLM being an improvement but LLMs are also way way costlier than normal translation models like google translate. Would it be worth it? And also would the better LLM translations make editors less likely to reword the translation to make it's tone better?
You can use an LLM to reword the translation to make the tone better. It's literally what LLMs are designed to do