[HN Gopher] 'Contextualization Engines' can fight misinformation... ___________________________________________________________________ 'Contextualization Engines' can fight misinformation without censorship Author : laurex Score : 51 points Date : 2021-04-26 16:38 UTC (6 hours ago) (HTM) web link (aviv.medium.com) (TXT) w3m dump (aviv.medium.com) | standardUser wrote: | This already exists. Unless I am missing something, we've already | been seeing these on major platforms like Facebook and Instagram | (if not others) for at least a year now. I've noticed it mostly | on posts about the 2020 election and COVID-19 facts/news. | | I mostly like the approach. It seems like the least restrictive | and least objectionable solution (or partial solution) to an | incredibly tricky and important problem. | JI00912 wrote: | I maybe misunderstanding this post, but Facebook information | about covid and the last us election has been terrible. | standardUser wrote: | I'm talking specifically about the little info blurbs that | Facebook automatically attaches to posts about COVID. They | are links to factual information. I'm not talking about | information that other people share about COVID on Facebook. | the_lonely_road wrote: | I spent a good amount of time trying to build a contextualization | engine for a different reason (auto correct, sick to death of it | being so bad when I'm clearly talking in a very specific context | bubble using these same words that everyone else is using over | and over), and man can I say that is is extremely difficult to | get 5 people in a room to not define 5 different context buckets | for a specific piece of text in a given situation. | | In the end I decided that the real world phenomenon I was trying | to model was probably closer to a Google+ multiple identifies | than it was to a single channel like Facebook. | adamrezich wrote: | whenever I hear talk about "creating context" with regards to the | vast amounts of conflicting information on the Internet, I can't | help but recall the infamous scene near the end of 2001's Metal | Gear Solid 2: Sons of Liberty, wherein protagonist Raiden talks | to an AI that has kind of actually been behind everything up to | this point (keeping in mind that this game was released long | before the modern social media-oriented Web) (emphasis mine): | | --- | | AI: The mapping of the human genome was completed early this | century. As a result, the evolutionary log of the human race lay | open to us. We started with genetic engineering, and in the end, | we succeeded in digitizing life itself. But there are things not | covered by genetic information. [...] Human memories, ideas. | Culture. History. Genes don't contain any record of human | history. Is it something that should not be passed on? Should | that information be left at the mercy of nature? We've always | kept records of our lives. Through words, pictures, symbols... | from tablets to books... But not all the information was | inherited by later generations. A small percentage of the whole | was selected and processed, then passed on. Not unlike genes, | really. That's what history is [...] But in the current, | digitized world, trivial information is accumulating every | second, preserved in all its triteness. Never fading, always | accessible. Rumors about petty issues, misinterpretations, | slander... All this junk data preserved in an unfiltered state, | growing at an alarming rate. It will only slow down social | progress, reduce the rate of evolution. Raiden, you seem to think | that our plan is one of censorship. | | Raiden: Are you telling me it's not!? | | AI: You're being silly! _What we propose to do is not to control | content, but to create context._ | | Raiden: Create context? | | AI: The digital society furthers human flaws and selectively | rewards the development of convenient half-truths. Just look at | the strange juxtapositions of morality around you. Billions spent | on new weapons in order to humanely murder other humans. Rights | of criminals are given more respect than the privacy of their | victims. Although there are people suffering in poverty, huge | donations are made to protect endangered species. Everyone grows | up being told the same thing. "Be nice to other people." "But | beat out the competition!" "You're special." "Believe in yourself | and you will succeed." But it's obvious from the start that only | a few can succeed... You exercise your right to "freedom" and | this is the result. All rhetoric to avoid conflict and protect | each other from hurt. The untested truths spun by different | interests continue to churn and accumulate in the sandbox of | political correctness and value systems. Everyone withdraws into | their own small gated community, afraid of a larger forum. They | stay inside their little ponds, leaking whatever "truth" suits | them into the growing cesspool of society at large. The different | cardinal truths neither clash nor mesh. No one is invalidated, | but nobody is right. Not even natural selection can take place | here. The world is being engulfed in "truth." And this is the way | the world ends. Not with a bang, but a whimper. We're trying to | stop that from happening. It's our responsibility as rulers. Just | as in genetics, unnecessary information and memory must be | filtered out to stimulate the evolution of the species. | | Raiden: And you think you're qualified to decide what's necessary | and not? | | AI: Absolutely. Who else could wade through the sea of garbage | you people produce, retrieve valuable truths and even interpret | their meaning for later generations? _That 's what it means to | create context._ | | --- | | pretty prophetic for a 2001 video game | https://www.youtube.com/watch?v=C31XYgr8gp0 | marcodiego wrote: | There is something that can fight misinformation without | censorship: a mix a culture + education. Scientific method, | critical thing, logic, statistics coupled with a bunch of | historical examples of non-intuitive results and demonstrating | situations where correlation does not implies causation should be | teach on schools repeatedly and from early age. | | Most adults don't even know what a double blind study is. It is | not surprising that so many people are falling prey to covid | quackery. | keiferski wrote: | I don't think many people working in tech understand the nature | of the Internet when it comes to information. The cat is out of | the bag and there is no returning to a world of universal | "authoritative sources" (term used in the article.) | | Framed historically, we are experiencing a similar situation to | the invention of the printing press. All attempts to reel in | "misinformation" by appealing to legacy media outlets are doomed | to fail, precisely because these outlets have lost the trust they | once had. I don't see them regaining it in the near future, if at | all. | | Instead, the likely outcome is a constellation of reputation- | based media sources. "Joe Brown is always honest, so I'll see | what he says in his podcast" is probably the future source of | information for the average person. Fact checkers will never be | anything other than a biased opinion of a particular institution, | as been shown time and time again. | noofen wrote: | Has anyone considered that the "rise of misinformation" and | Qanon-type conspiracy discourse is just a symptom of elite and | institutional decadence? | WalterGR wrote: | _elite and institutional decadence_ | | What does that mean? I googled "institutional decadence" but | am still not clear on the definition. Could you give some | examples of elite and institutional decadence? | keiferski wrote: | Not the OP, but this just refers to the gradual erosion of | many institutions in American society. Universities, media, | political structures, etc. have all lost the trust of the | populace over the last ~30 years, if not longer. | | Whether this is from an actual lowering of standards or | just an inevitable consequence of new media is a different | question. I'd suggest reading works by a critic named John | Simon for the former opinion. | WalterGR wrote: | Hmm... When I think of decadence I think of Marie | Antoinette allegedly asking "Why don't they just eat | cake?" Like, actual decadence. In this context is | "decadence" being used as a loaded word? At university I | never saw professors or administrators popping Cristal | and making it rain hundos, for example. | | (Edit: For the record, I didn't downvote you. One can't | downvote a comment one is responding to.) | keiferski wrote: | I haven't really followed the Q-Anon thing well enough to | know much about it, but it definitely does strike me as a | consequence of institutional decadence. | | I'm not sure the content of the conspiracy is actually | relevant. Most people _want_ to trust the information they're | told, so the popularity of such things does seem like a | symptom and not a cause. | barbazoo wrote: | Education, to me, seems like the perfect way to "fight | misinformation without censorship". Teach our kids how to read, | how to understand, how to contextualize and think critically. We | don't need big tech to come up with a solution to a problem that | big tech caused to some degree. | potatoman22 wrote: | As much as I hated it as a kid, I think literary analysis and | semiotics is really beneficial in this manner. | narag wrote: | I agree but you also need to teach them that neither side of | the political divide is right on everything. Good luck getting | support for that. | barbazoo wrote: | Believing in there being 2 sides is probably a good indicator | that we've already failed. | narag wrote: | There are more than two in my country, I was just | "translating"... | amalcon wrote: | Let's suppose I am Google. I am in arguably the best position in | the world to build this system, because I have PageRank. PageRank | is (for the sake of argument) the closest thing yet invented to | an unbiased measure of authoritativeness (though still not _that_ | close). I also have the most comprehensive map of the Web ever | made, because that 's the input to PageRank. | | I could imagine building a contextualization machine out of this. | Oversimplified version: Look for cases where an authoritative | link frequently appears next to a link to the page in need of | context, and surface those authoritative links. | | The good news / problem is that you can already use google this | way. If you google a given URL, you'll get results of people | talking about it! E.g. the current top article on HN is: | | https://www.apple.com/newsroom/2021/04/ios-14-5-offers-unloc... | | If I google this, I get a Reddit thread discussing this, two | misses (MacRumors frontpage and a CNBC index page), and a Korean | page. The last might be fine, but I can't read it so I don't | know. Maybe the Apple announcement is just too new, but this | doesn't seem super useful. | | Of course, maybe there are refinements one could make for this | use case. Maybe this technique is just more useful for political | or science topics. I don't know. I tend to doubt it, though. | ergot_vacation wrote: | The problem of course is that a certain crowd runs most large | sites, and as seen on Youtube, if things they dislike become | too popular they will do whatever they can to suppress that | popularity. | | Put another way, there are two definitions of "Authoritative": | what does the raw data suggest, and what does the echo chamber | I've surrounded myself with and become convinced is reality | say. I think we all know which standard would be used if such a | system was implemented (and of course, it's already been | attempted in small ways). | motohagiography wrote: | Interesting that it's possible, but it's probably not ethical to | misdirect people about the intent of this. Some people experience | what we might see as misinformation as their culture and | identity, and pretending we aren't building technologies to | detect and exterminate it is pretty dodgy. Unless we think these | tools will only ever be in the hands of 'good' people? | | Being able to detect the tribal flavor of information so that you | can filter it out is what got us into this mess in the first | place because it indexed on creating outrage bubbles for clicks. | Growth comes from opportunity, which usually comes from something | being shitty, so militating and cleansing the world of badness | isn't doing anyone any favors except consolidating the role of | the cleaners. | | When you ask the question, "whose contextualization engine?" and | the whole premise falls apart, it becomes clear they have earned | any mockery these engines are designed to sheild them from. | lindy2021 wrote: | The real test of a fair system is to build your | "contextualization engine" and then hand it over to your | ideological enemies to run. | MayeulC wrote: | Would you hand the wheel over to your enemies? Agriculture? | Writing? | | Probably not, if they are assured to remain your enemies. | Wars have almost always been decided on technological | capabilities alone, with a few counter-examples. You asked an | interesting question, if only a bit narrow. | | Going further, most technologies can be directly used for | ill. | max-ibel wrote: | I cut -- you choose ? | | I like it ;) But it will probably never happen. | toss1 wrote: | When your "culture and identity" is based on relying on blatant | disinformation, whether it is that the earth is flat, the govt | is poisoning you with chemtrails, or NASA is fake, that is | already getting corrossive to society. | | When it extends to anti-vaccine disinformation, it is an active | threat not only to public health but national security. Many of | these types of disinformation are actively weaponized and | spread by adversaries. | | It is right to threat those kinds of threats to the society and | nation as we would treat any physical threat -- actively work | to neutralize it. | | Just because war is moving out of the physical and into the | information sphere, does not make it less serious. | motohagiography wrote: | The principle I'm appealing to is that we have to ask whether | it's appropriate or desirable to have that technology applied | to us(*). It's not disinformation if it's just outgroup | culture. | | It seems unwise to imagine oneself engaged in a war because | it just licenses your opponent. The idea of war is that it | the extension of policy, which in every other circumstance | can be negotiated, but if instead of negotiating policy, | these same people are engaged in a deception to exterminate | or subordinate an opposition, that's not policy, that's just | a power struggle for its own end, and it's not something to | be reconciled. | | If you want to treat fellow citizens as an enemy force, say | so, and they will come to the table as one. But if you want | to just decieve them, I'd wonder whether we were the baddies. | We probably don't want to make civilian technologists legit | targets in this war either. | | Surely the people developing or advocating these techs are | capable of enough self awareness to recognize how every | example of "blatant disinformation," has an equivalent worthy | of targeting, so introducing them as anything other than | abstractions doesn't really yield perspective. | visarga wrote: | > If you want to treat fellow citizens as an enemy force, | say so | | Dividing people is a bad idea because we have much more in | common than is different between us. We should always try | to be more inclusive and tone down the class wars and | cancellations. We probably won't make any social progress | while throwing stones into the opposing tribes instead of | working together. | | But you can't be inclusive of falsehoods, you can only open | to legitimate concerns raised by the various groups, | because any solution we find needs to be viable in the real | world. | | I don't trust revolutionaries - I've been burned before, | and also read history. Revolutionaries, social warriors and | activists love revolution more than peace itself. If we | want to succeed we need to win the peace, not the war. | freeflight wrote: | _> It 's not disinformation if it's just outgroup culture._ | | Just because wrong information can become outgroup culture | does not suddenly make the wrong information true. | | Particularly when it stipulates world views that vilify | whole other groups as allegedly being responsible for | everything wrong. | | The end result is not just a culture that might endorse | violence against others, but something that could probably | be described as a cargo cult where wrong information | becomes dogmatic tradition. | | Antisemitism has already taken such a shape, most if not | all of the theories voiced by some modern outgroups are | just reboots of allegations that are centuries old and have | justified violence for just as long. | | Case in point: The whole "elites abducting children to | sacrifice them" theory, which regained popularity trough | QAnon, is just a more modern take on the William of Norwich | blood libel [0]. | | That has been around for close to a millennium because it's | an idea that survives in certain religious outgroup | cultures. Do we really have to tolerate the intolerance to | be considered tolerant? Is there really no rough line that | can be drawn? | | [0] https://en.wikipedia.org/wiki/William_of_Norwich | motohagiography wrote: | I'd say that Popper's Paraphrase isn't a useful argument | when it reduces to just a first-accusers advantage. It's | a small mercy that he died before seeing what his work | would be reduced to. | | Tolerance is meaningful when it is mostly unbearable, | otherwise it's just preference. That idea is elaborated | well in SSC's "I can tolerate anything but the outgroup." | The relative truth of wrong information is a glass house | I don't think advocates of censorship are intellectually | equipped to defend. Their only tool is deception, and | inevitably force, and the points we can score on them | have no value. | | The deeper controversy is whether on the internet we will | accept a rules based order, or discretion. It's the old | "rule of law vs. rule of men" issue. What has changed in | the last 140 years or so is of course, total, instant | globalization, with the internet, but more subtly, that | the language used to reason about these things has itself | become unmoored. | | What I think we should all reflect on is how as | individuals we can become a bit less enthusiasitically | murderable to one another, before worrying about how we | can change what others think. It's the one problem we can | all make progress on. | gedy wrote: | This sounds "right" but then how would you feel applying this | things like native American tribal beliefs, e.g. regarding | genetics and migration. Those may be scientific facts but | many reject them due to fearing further dispossession of | their land, etc. | pessimizer wrote: | You're clearly an information warrior, but what if I think | that you're the one spreading misleading information? | | It's weird that you fully expect that you, or someone else | who agrees with you, will be in control of this. | toss1 wrote: | Whether you think that I'm spreading disinformation, or | whether it is someone else is easily determined by looking | to the external facts. | | The fact is that the earth is (nearly) round, contrails are | not chemical sprays, Capricorn One was a C-grade movie | while Apollo-11 went to the moon, and that anti-vax | movement was started by a fraud attempting to discredit | existing formulas to promote a formula in which the paper's | author had a vested interest (and which paper was | retracted, along with his MD and license) and that vaccines | do work. | | That is the point. A society cannot remain cohesive if | there are no shared truths. | | You may not want to take it seriously, but I'd suggest you | consider the following quotes, and why and how | disinformation can be weaponized. | | "Anti-intellectualism has been a constant thread winding | its way through our political and cultural life, nurtured | by the false notion that democracy means that 'my ignorance | is just as good as your knowledge.'" -- Isaac Asimov | | "You are entitled to your opinion. But you are not entitled | to your own facts.". -- Daniel Patrick Moynihan. | motohagiography wrote: | > A society cannot remain cohesive if there are no shared | truths. | | On this we agree. This truth needs to be on a scale of | one, some, or all truths that are shared. Let's | disqualify all, and simplfy to finding an example of one. | | The other thing about the truth is that it probably | shouldn't be falsifiable, as it's a bit of an all-in bet | on whatever that is. It shouldn't be a complicated | mystery because then you are back to the relative problem | of who has discovered more and how do you know. The | perfect example truth is the one that is not knowable | either way, but something that can be believed. | | We should probably recognize that destabilizing that | example shared truth is an attack on social cohesion as | well. Can't think of an example at the moment, but I | certainly agree it would be the One Important Thing you | would need to believe, or you're basically left with a | kind of hellscape. | BitwiseFool wrote: | Everyone needs to pick and choose which | information/disinformation battles they want to take part in. | Flat-Earthers, Chemtrail People, Moonlanding Hoaxers, these | people aren't really worth anyone's time. How does someone | else believing the world is flat impinge on your ability to | live your life? | | Now, I can understand wanting to combat vaccine hesitancy | because that does tangibly affect society at-large (recent | measles outbreaks). But most of the disinformation boogeymen | you listed are just people with kooky ideas who aren't | persuading society. | | Lastly, it's okay to be wrong about things. The body of human | knowledge is so vast that everyone is bound to be more wrong | than right. | toss1 wrote: | There is a difference between 1) simply being wrong by | mistake vs 2) deliberate and maintained ignorance vs. 3) | deliberately deceiving people and poisoning the agora. | | Mere ordinary wrongness is a big enough problem. And to the | degree that it is mere cultural ignorance, it can be merely | annoying. | | But intentionally spreading disinformation, fear porn, and | cultivating false outrage to inflame divisions in society | is much more serious, and should not be conflated with mere | buffoonery. | keiferski wrote: | This line of thought will both be extremely ineffective (and | further incentivize misinformation) and highly unjust. At | some point you need to stop and realize everything isn't a | conflict that necessarily has losers to be "neutralized." | toss1 wrote: | Agree, it only becomes a conflict when some parties are | deliberately weaponizing disinformation. | | Mere ignorance is something we must unfortunately tolerate. | | Active work to spread disinformation that will literally | kill people - see [1] or just search for it, where a mere | dozen people are responsible for 65% of the spread of | disinformation about the COVID-19 vaccines, and are working | to deliberately undermine vaccination efforts. | | This isn't mere ignorance, it will get people killed. And | before you say something like "well the fools that believe | that idiocy deserve what they get", consider that while | they may (sort of their choice), the family, friends, | neighbors, first responders, and healthcare workers do not | deserve to be exposed unnecessarily to those biohazards. | | [1] https://www.theguardian.com/world/2021/jan/06/facebook- | insta... | kodah wrote: | > Agree, it only becomes a conflict when some parties are | deliberately weaponizing disinformation. | | Agree, it only becomes a conflict when both major parties | are deliberately weaponizing disinformation. | | I think other commenters have warned about this. It's | fairly easy to tell what political allegiances you have. | Just understand that your parties spread a lot of | disinformation as well. As an independent, it makes me | cringe to watch liberals act like they aren't responsible | for loads of disinformation, hyperbole, fear mongering, | and misinformation too. Much less that it doesn't have | Russian origins. | keiferski wrote: | Every single revoking of civil liberties (e.g. | censorship) has been done in the name of safety. That | doesn't make it acceptable. | | If you want to combat misinformation, pick a better | solution. And as I said above, censorship is pretty much | impossible, both legally and technologically. It simply | cannot be done. | toss1 wrote: | Did you even read the article? | | The ENTIRE POINT of the idea is to NOT censor anything, | but to add context -- adding information, not censoring | it. Letting people make their own decisions. | | The problem here is that there is a massive asymmetry. | This has been noticed for centuries with the maxim "A Lie | Can Travel Halfway Around the World While the Truth Is | Putting On Its Shoes" (some interesting history on that | at [1]). | | This is because our brains are tuned to seek out novelty, | and to seek out information on threats. | | This enables those who want to manipulate us and | especially the masses -- creating fake news and fear porn | will get the attention of pretty much everyone, and | anyone who is not both actively better informed at the | time, and actively thinking, will be inflamed and | motivated. | | Voltaire pointed out that "Those who can make you believe | absurdities, can make you commit atrocities." | | The people spreading disinformation treat that as an | instruction instead of a warning. | | Adding this kind of contextual information seems a far | better solution than censoring. | | [1] https://quoteinvestigator.com/2014/07/13/truth/ | keiferski wrote: | Yes I read the article, but I was replying to your | comment, not the article's argument. I replied to that | directly in another comment. | | You seemed to be arguing for censorship, which is what I | was replying to. | toss1 wrote: | Please read more carefully | | I am absolutely not arguing for censorship, especially | since censorship creates it's own blowback, which is | often even more helpful to the disinformation. | | I am merely arguing that disinformation needs to be | fought, especially actively weaponized disinformation. | | It looks like this contextualization could be a very good | method. | andrewjl wrote: | > Being able to detect the tribal flavor of information so that | you can filter it out is what got us into this mess in the | first place because it indexed on creating outrage bubbles for | clicks. | | What you're referring to as outrage bubbles has come about due | to a lack of contextualization. One topic this is commonly seen | is discussion on GMOs, where both pro-GMO and anti-GMO | activists provide misleading facts as supporting evidence in | their discussion. In this example and also when it comes to | politics, a contextualization engine would be very effective in | spurring dialogue and getting people to see multiple sides of | an issue. | | > When you ask the question, "whose contextualization engine?" | and the whole premise falls apart, it becomes clear they have | earned any mockery these engines are designed to sheild them | from. | | Where I personally differ from some of the social networks on | this, is I think these should be default on but also come with | an off switch. Meaning the user always has the ability to turn | them off, but they need to take an active step in order to do | so. IMO it's a good compromise. | gweinberg wrote: | There are pro-GMO activists? | groby_b wrote: | Can we stop saying it's due to "lack of contextualization" as | if it's something that just oddly happened? | | The reason we lack context is that the world moved from | chronological feeds to "engagement" based feeds. Anything | that could have served as context gets optimized out, because | more clicks. We're not naturally polarized - we have created | an environment that favored the more outrageous claims, | slowly removing the middle ground as if it didn't exist. | | What we need isn't a "contextualizion engine", it's getting | rid of active efforts to decontextualize. | wmf wrote: | Yeah, people won't react well when the fact checker tells them | that _everything they read_ is fake news. | endymi0n wrote: | "If conspiracy theorists become convinced that their facts | are fake, they will not abandon conspiracy theories. They | will reject the authority of the fact checkers." | | - Loosely paraphrased after David Frum | | https://www.reddit.com/r/LeopardsAteMyFace/comments/mmy39q/o. | .. | JI00912 wrote: | It would be great if we could all just turn to an objective | entity with all the facts. But there's no such person. | Certainly not the average journalist. There's just no such | thing as a fact checker. | sofixa wrote: | Everyone has their biases, but most are able to look over | them in the face of facts. And the good thing about facts | is that they're objective and indisputable. | | Moon landing ? Fact. Trump being US president? Fact. | France winning the 2018 World Cup? Fact. | | So no, fact checkers do exist and are very important. | generalizations wrote: | As long as the fact checkers stick with facts that are as | obvious as the ones in your examples, then sure: they're | objective and indisputable. | | However, people do dispute facts, and anything someone | disputes is no longer in the purview of the fact checker, | _by your definition_. | argvargc wrote: | All of those examples could be invalidated with new | information. | | As inconvenient as it may be, truth is transitory and | progress toward it is frequently made through | invalidation. It's more prudent and useful to assume | facts don't exist. | | As for "fact-checkers", who fact-checks _them_? (Etc...) | The notion of visiting a webpage authored by God-knows- | who to check whether what 's on another webpage authored | by God-knows-who is correct or not is absurd. | | And we haven't even gotten into subjectivity or nuance. | msla wrote: | Is the idea that Jews are literal monsters part of someone's | culture and identity? | | Is that culture and identity something the rest of the world | can live with? | lindy2021 wrote: | Yes. Jews have been treated as literal monsters for millennia | and the rest of the world has lived with it. | | https://en.wikipedia.org/wiki/Expulsions_and_exoduses_of_Jew. | .. | giantg2 wrote: | "Is the idea that Jews are literal monsters part of someone's | culture and identity?" | | _The idea that some people may believe that_ is central to | some cultural identities. For example the Hasidic community, | or the defense /military posture of Israel. | ketzo wrote: | For such a long piece, this is disappointingly light on any | details or discussion about _how_ one would go about building a | "Contextualization Engine." | | I mean, we all saw what happened with Microsoft's Tay Bot [1] -- | any system that attempts to factor in user input is ripe for | abuse. Who would control what context is valid, and what context | is discarded? If you decide to factor in "all" content, how do | you rank it? Is it simply a factor of popularity, or are trusted | sources given more weight? Who is a trusted source? | | I don't mean to sound overly pessimistic: I agree with the author | that context is the biggest thing lacking from the internet | today, and is a valuable tool for fighting misinformation. But | any discussion like this needs to pretty immediately grapple with | the intricacies of how to implement context -- and what context | really _is._ | | [1] https://en.wikipedia.org/wiki/Tay_(bot) | throwawayboise wrote: | Sounded like it was mainly reliant on some pre-selected | whitelist of approved sources. Of course that raises the | question of who approved those sources, and what were their | motivations? | giantg2 wrote: | "The contextualization engine compares the content being shared | with that from authoritative sources and provides articles or | other media results that are sufficiently related." | | How are these authoritative sources chosen? This sounds like it | would be rife for manipulation. Plus, just relying on an appeal | to authority as the validation of the information will not weed | out mistakes or errors in the content. | | What are the standards for these other media results it will | provide? News articles, blogs, studies, etc. Some might be better | than others depending on the topic. The news and journalists are | supposed to be the old school contextual engine. How do we see an | automated version being any better quality? | Fellshard wrote: | It already is. Watch the Twitter 'trending' pane over time, and | you'll notice that they add context in exactly one worldview | direction, incurring massive editorialized anchor bias. | ergot_vacation wrote: | Exactly. The fundamental problem with the phrase | "misinformation" is that what is or isn't "misinformation" is | _subjective_. Truth and reality aren 't subjective, but | people's beliefs about the nature of them are, very much so. | It's impossible to solve the problem by asking "The Experts" to | adjudicate what is and isn't true, because The Experts have | become too mired in corruption, greed, political intimidation | and so on. They no longer represent anything even close to an | impartial party, and thus are not useful tools for discerning | reality anymore, at least in the traditional way of asking a | question, getting an answer and accepting it. | tryonenow wrote: | >Systems built specifically for contextualization might not only | support media literacy; they could also provide the data needed | for fact-checkers to determine what to focus on, and could even | help support the emotional literacy relevant to avoid harmful | reactions to misinformation (from lashing out at loved ones, to | terrorist radicalization). | | All I'm hearing is techno-authoritarians building a system to | further suppress dissenting views. Even "authoritative" sources | are supposed to disagree; that's how science works. | | I think the uncertain nature of reality is such that any | "authoritative" source is likely to eventually make mistakes, and | building rigid systems to hide alternative viewpoints is a form | of social disenfranchisement. | drngdds wrote: | Why would this be any more trusted than the misinformation | warnings Youtube and Twitter already put up automatically for | certain subjects? | | And also, I don't know how anyone can even float the possibility | of making something like this a legal requirement. The ways it | could be abused are really obvious. ___________________________________________________________________ (page generated 2021-04-26 23:02 UTC)