view the rest of the comments
Unpopular Opinion
Welcome to the Unpopular Opinion community!
How voting works:
Vote the opposite of the norm.
If you agree that the opinion is unpopular give it an arrow up. If it's something that's widely accepted, give it an arrow down.
Guidelines:
Tag your post, if possible (not required)
- If your post is a "General" unpopular opinion, start the subject with [GENERAL].
- If it is a Lemmy-specific unpopular opinion, start it with [LEMMY].
Rules:
1. NO POLITICS
Politics is everywhere. Let's make this about [general] and [lemmy] - specific topics, and keep politics out of it.
2. Be civil.
Disagreements happen, but that doesn’t provide the right to personally attack others. No racism/sexism/bigotry. Please also refrain from gatekeeping others' opinions.
3. No bots, spam or self-promotion.
Only approved bots, which follow the guidelines for bots set by the instance, are allowed.
4. Shitposts and memes are allowed but...
Only until they prove to be a problem. They can and will be removed at moderator discretion.
5. No trolling.
This shouldn't need an explanation. If your post or comment is made just to get a rise with no real value, it will be removed. You do this too often, you will get a vacation to touch grass, away from this community for 1 or more days. Repeat offenses will result in a perma-ban.
6. Defend your opinion
This is a bit of a mix of rules 4 and 5 to help foster higher quality posts. You are expected to defend your unpopular opinion in the post body. We don't expect a whole manifesto (please, no manifestos), but you should at least provide some details as to why you hold the position you do.
Instance-wide rules always apply. https://legal.lemmy.world/tos/
It's not only that 10% is wrong, it's knowing which 10% is wrong, which is more important than it seems at first glance. I feel strongly that AI is contributing to people's inability to really perceive reality. If you direct all your learning through a machine that lies 10% of the time, soon enough your entire world-view will be on shaky ground. How are you going to have a debate with someone when you don't know which parts of your knowledge are true? Will you automatically concede your knowledge to others, who may be more convincing and less careful about repeating what they've learned through AI?
I think all that AI really needs to do is translate natural language requests ("What factors led to WW2?") into normal destinations for further learning. Letting AI try to summarize those destinations seems like a bad idea (at least with where the technology is right now)