[HN Gopher] NeRF: An eventual successor for deepfakes? ___________________________________________________________________ NeRF: An eventual successor for deepfakes? Author : Hard_Space Score : 142 points Date : 2022-06-27 13:47 UTC (9 hours ago) (HTM) web link (metaphysic.ai) (TXT) w3m dump (metaphysic.ai) | 1970-01-01 wrote: | Neural Radiance Fields (NeRF) | Workaccount2 wrote: | I know there are a lot of groups working on how to prevent AI | disruptions to society, or how to mitigate their impact, but are | there any groups working on how to adapt society to a full blown | unchained AI world? | | Like throw out all the safeguards (which seems inevitable) and | how does society best operate in a world where no media can be | trusted as authentic? Or where "authentic" is cut from the same | cloth as "fake"? Is anyone working on this? | Agamus wrote: | One thing we should be doing is supporting critical thinking at | the high school and university level. Unfortunately, it seems | we have been dedicated to the opposite of this for about 50 | years, at least in the US. | Der_Einzige wrote: | Critical thinking is the most overrated skill ever. | | "Critical theorists" are the people who fetishize "critical | thinking" and all it got them was to embrace cultural | Marxism. | | Constructive thinking is far better than learning how to shit | on people, as the skill of critique teaches us... | gknoy wrote: | At the risk of falling for a joke, I'm not sure "critical | thinking" means what you think it means. It just means | thinking objectively about things before making judgments, | it has nothing to do with criticizing people. The things | one criticizes are our own beliefs and reasons for | believing them. | | What do I believe? Why do I believe that? Why do I feel | that evidence supports that belief, but not this one? For | example, I can explain in a fair bit of detail why I | believe that the Apollo landing was not faked. I wouldn't | normally bother to explain those reasons, but all of them | are based on beliefs and evidence that I've read about, and | most of those beliefs are subject to reversal should | counter-evidence surface. | whimsicalism wrote: | This is reasoning by word chaining. | Agamus wrote: | I think of critical thinking as the art of being critical | toward oneself when one is thinking. | | In other words, when I read something and hear myself | think, "oh yeah, that sounds right", there is another part | of my mind that thinks, "maybe not". | | Critical thinking is precisely what could have spared us | from all of that 'cultural marxism' you mentioned, or at | least, to do it in a way that is... constructive. | strohwueste wrote: | What about nft-based camera recording? | irjustin wrote: | This has been discussed many times and it doesn't work. | | Simple answer is I can just record a deep fake and get it | cryptographically signed. | andruby wrote: | If you trust a person (or source) and they have a private | key that they can properly secure, they could always sign | their material with that key. That would prove that the | source provided that material. | | A blockchain could be a way to store and publish that | signature & hash. | | It can't say "this is real", it can only say "that | signature belongs to source X". | blamestross wrote: | > A blockchain could be a way to store and publish that | signature & hash. | | yes but it would be a bad one. We have multiple key | distribution mechanism that are better for the use case. | intrasight wrote: | I disagree. The key alone is not sufficient nor secure. | We will need crowdsourced validity data as well. We need | a zero-trust model - and I too believe that blockchains | will play a role. | notfed wrote: | If a video is already cryptographcally signed, then you | can safely distribute the signature on an untrusted | channel. | | Adding Blockchain into the mix is superfluous, and | destroys scalability. | intrasight wrote: | We don't watch signature keys - we watch videos. | | The TV will have to match every short segment - perhaps 5 | seconds of video - against a blockchain which scores the | validity of that segment - and of course looking back to | it's original source. Signing the whole video is | necessary but not sufficient. | | But yes, this is going to be resource-intensive. | blamestross wrote: | > against a blockchain which scores the validity of that | segment | | Why not just allow a cert with that information to be | delivered alongside the video? Where would the "score" | come from? | blamestross wrote: | Zero trust models don't exist and the laws of physics | (probably) don't provide for them. (Materialism is a real | problem in physics nowadays) | endtime wrote: | I think most of those people believe that humans will no longer | exist in a "full blown unchained AI world". | cameronh90 wrote: | I suspect we'll need to return to the idea of getting our news | from trusted sources, rather than being able to rely on videos | on social media being trustworthy. | | Technically, we could try and build a trusted computing-like | system to ensure trust from the sensor all the way to a signed | video file output, but keeping that perfectly secure is likely | to be virtually impossible except in narrow situations, such as | CCTV installations. I believe Apple may be attempting to do | things like this with how Face ID is implemented on iPhone, but | I suspect we'll always find ways to trick any such device. | wongarsu wrote: | 80% of the problem could be solved with a reliable signature | scheme that allows some remixing of video content. So if CNN | publishes a video, signed with their key so it's verifiably | CNN, we need the ability to take a 20 second bit of it and | still have a valid key attached that verifies that the source | is CNN (without trusting the editor). Then you can share | clips, remix it, etc, and have integration in social media | that attests the source. | _tom_ wrote: | Once you remix it, it's no longer reliable. So, you don't | want it signed if it's modified. | intrasight wrote: | My plan to solve this "20 second bit of it" is that it's | done at the analog hole. Whatever is painting those pixels, | a smart TV for instance, will be coordinating with cloud | services to fingerprint at a relatively high temporal | resolution - maybe 5 seconds. The video itself is the | signature. But we will need either trusted analog hole | vendors or some trusted non-profit organization - or likely | both. I think that "viewing" will be delayed by perhaps 30 | seconds to allow for that signature analysis. These smart | TVs will overlay a scorecard for all displayed content, and | owners will be able to set device scorecard thresholds such | that low-scoring content will be fuzzed out. | dTal wrote: | I sincerely hope this dystopian vision of the future is | satire, but it's already a worrying sign of the times | that I'm not sure. | wazoox wrote: | There are more dangereous AI than deepfakes. Blackrock 10 | trillions investments are driven by an AI, Aladdin. It was also | sold to some other investors, and controls about 21$ trillion | globally. It has basically the power to drive markets worldwide. | It's a systemic problem nobody talks about... | astrange wrote: | Blackrock sells passive index funds. There's no room for an AI | to make decisions there, so it probably doesn't do anything | interesting. | smiddereens wrote: | nix0n wrote: | That's not an AI problem, it's a concentration of wealth | problem. | | Giving that power to a person or group of people would be | almost as bad. | kobalsky wrote: | why do websites feel the need to hijack the browser's scrolling | logic? | | this is very annoying to browse on chrome, but it works well on | firefox. | sigspec wrote: | Agree. It's a jolt to scrolling expectations | arky527 wrote: | Very much agreed. It just results in a terrible UX when 98% of | other websites have a standard scrolling mechanism | [deleted] | jdthedisciple wrote: | The bad news: This can and obviously will be abused - be it by | the secret services or hackers. | | The good news, I suppose: As fast and scarily as the tech to fake | things is evolving, so is presumably the tech that detects fakes. | _tom_ wrote: | Technology to detect fakes necessarily lags a bit from | technology to create fakes. | | In general, you need to have examples of a type of fake to | detect it. | RubyRidgeRandy wrote: | Something I've wondered lately is what will life be like in a | post-truth society? we already see examples of this now where a | large number of people get their news from fake memes on | facebook. There are huge swathes of people who live in their own | make-believe world, like those believing wholeheartedly that the | 2020 election was stolen. | | What will life be like when you can't trust any video or | interview you see because it could be completely fake? How long | before someone uses this technology to frame someone for a crime? | Could the FBI create a deepfake of a cartel leader meeting with | them and leak it so they think he's a snitch? | | I don't think we'll have the ability to handle this kind of tech | responsibly. | hutzlibu wrote: | "I don't think we'll have the ability to handle this kind of | tech responsibly." | | I do not think so either, but so far we survived 75+years with | nukes around. | | But you can argue, it was mainly by chance. Technological | progress _is_ awesome, but our societies cannot keep up yet. | They will have to do heavy transition anyway, or perish. Or | rather, we are in the process of transition. 20 years ago most | people did not really know, what the internet is, now most are | always online. Data mining, personalised algorithms for ad | exposing, .. | | So deepfakes are a concern, but not my biggest. Rather the | contrary, when people see how easy it is to fake things, they | might start developing a healthy sceptism to illuminating | youtube videos. | munificent wrote: | I think we'll solve it the same way we solved similar | transitions when text and image faking became easy: provenance. | | For many years now, most have understood that you can't take | text and images as truth because they can easily be simulated | or modified. In other words, the media itself is not self- | verifying. Instead, we rely on knowing where a piece of media | came from, and we associate truth value with those | institutions. (Of course, people disagree on which institutions | to trust, but that's a separate issue.) | kbenson wrote: | In other words, the same way we dealt with information before | photographs and videos were invented. The answer to how to we | deal with the fact that images and videos can't be trusted is | to look at what we did before we relied on them. If we're | smart about it we'll try to pick out the good things that | worked and try to build in safeguards (as much as possible) | against the things that didn't, but I won't hold my breath. | We're already heading back towards some of the more | problematic behavior, such as popularity or celebrity | equating to trust. | pfisherman wrote: | I think that people will adapt. Humans are very clever and have | been evolutionarily successful because of the ability to adapt | to a wide range of environments. | | Think about how devastatingly effective print, radio, and | television propaganda at the time each medium was widely | adopted compared to how effective they are now. They still | work, but for the most part people have caught on to the game | and adjusted their cognitive filters. | | My guess is that we will see a bifurcation of society into | those who are able to successfully weed out bullshit from those | who can't. The people who are able to process information and | build better internal models of the world will be more | successful, and eventually people will start imitating what | they do. | | Edit: I do think that these tools coupled with widespread | surveillance and persuasion tech (aka ad networks), have set up | conditions where the few can rule over the many min a way that | was not possible before. I do think some of the decentralized / | autonomous organization tech - scaling bottom up decision | making to make it more transparent and efficient - is a | possible counter. Imo, this struggle between technological | mediated centralization and top down enforcement and control vs | decentralization and bottom up consensus will be the defining | ideological struggle of our time. | azinman2 wrote: | I think you're underestimating the power of cognitive biases | and violence from those who only believe the information that | want to hear. | pfisherman wrote: | I think there are quite a few recent historical examples of | this - WW2, US invasion of Iraq, Russian Invasion of | Ukraine, etc. | | However there is a price to pay for operating on beliefs | that do not align with reality. It's why almost all | organizations engage in some form of intelligence | gathering. Those who are at an information disadvantage get | weeded out. | | Philip K Dick has a great quote "Reality is that which, | when you stop believing in it, doesn't go away." | adhesive_wombat wrote: | A skeptic (i.e. someone who cares to verify) not being able to | trust media because it might be fake is only a minor problem as | long as you have at least one trusted channel. | | The president, say, can just release the statement on that | channel and it can be verified there (including | cryptographically, say by signing the file or even using | HTTPS). | | If you lose that channel, then you're pretty much screwed | because you'll never know which one is the real president. But | there are physical access controls on some channels, say the | Emergency Alert System, which can be used to bootstrap a trust | chain. | | What will be much more possible is that someone who will not | check the veracity of the message will take it at face value | without bothering to validate it. This is your news-via- | Facebook crowd. | | At that point, it's less a technical issue than simply people | don't want to know the truth. No amount of fact-checking and | secure tamper-proofing of information chains of custody will | help that. | toss1 wrote: | Agree, and it's even worse than that | | An incredibly small minority of people even understand your | phrase with any actual fidelity and depth of meaning: | | >>it can be verified there (including cryptographically, say | by signing the file or even using HTTPS) | | Even fewer of that microscopic minority have and understand | how to use the tools required to verify the video | cryptographically, AND even fewer know how to fully validate | that the tools themselves are valie (e.g., not compromised by | a bogus cert). | | Worse yet, even in the good case where everyone is properly | skeptical, and 90+% of us figure out that no source is | trustworthy, the criminals have won. | | The goal of disinformation is _not_ to get people to believe | your lie (although the few useful idiots who do may be a nice | bonus). | | The goal of disinformation is to get people to give up an | even seeking the truth - just give up and say "we can't know | who's right or what's real" -- that is the opening that | authoritarians need to take over governments and end | democracies. | | So yes, this is next-level-screwed material. | adhesive_wombat wrote: | > AND even fewer know how to fully validate that the tools | themselves are valie (e.g., not compromised by a bogus | cert). | | Kind of, but once you have _a_ single verifiable channel | back to the source (in this case, some statement by the | president) it 's now possible for anyone to construct a web | of trust that leads back to that source. For example, | multiple trustworthy independent outlets reporting on the | same statement in the same way, providing some way to | locate the original source. This is why new articles that | do not link to (on the web) or otherwise unambiguously | identify a source wind me up. "Scientists say" is a common | one. It's so hard to find the original source from such | things. | | This falls over in two ways: sources become non-independent | and/or non-trustworthy as an ensemble. Then you can't use | them as an information proxy. This is what is often claimed | about "the mainstream media" _and_ the "non-mainstream | media" by the adherents if the other. All the fact checks | in the world are worthless if they are immediately written | of y those they are aimed at as lies-from-the-system. | | The second way is that people simply do not care. It was | said, it sounds plausible, and they want to believe it. | | So I would say that actually the risks here are social, not | technological. Granted, perhaps a deepfake-2'd video might | convince more people than a Photoshopped photo. The core | issue isn't the quality of the fake, it's that a | significant number of people simply wouldn't care if it | _were_ fake. | | Doesn't mean we're not screwed, just not specifically and | proximally because of falsification technology, that's | accelerant but not the fuel. | toss1 wrote: | >>that's accelerant but not the fuel. | | Yes, indeed! Which is why I'm having so much trouble with | ppl proposing technological solutions - technically it | might solve the problem in some situations, but the | bigger problem is indeed some combination of general | confusion, highly adversarial information environment | laden with disinformation, and people's all-too-frequent | love of confirmation bias and willingly believing BS and | overlooking warning signs. | | I hope we can sort it... | efrbwrh wrote: | gadders wrote: | >> There are huge swathes of people who live in their own make- | believe world, like those believing wholeheartedly that the | 2020 election was stolen | | There are also those that wholeheartedly believe Trump colluded | with Russia to win the 2016 election, or that the Steele | dossier was factual. | dTal wrote: | These are not equivalent. Russia _did_ interfere, there | _were_ links between Trump and Russia, therefore there is | circumstantial evidence that collusion occurred, sufficient | to trigger a widely publicized investigation. The allegations | of election fraud in 2020 however are 100% alternate universe | yarns spun for political gain with no basis in fact | whatsoever. | decafmomma wrote: | Here's the thing though: in theory, we should already be | skeptical of video and audio evidence on its own. | | Most of our institutions, in theory, do not focus on single | mediums for assessing veracity of truth. The strength of claims | and our ability to split the difference between noise and truth | comes down to corroboration. How many other sources strength | and work consistently with a claim? That's, in theory, how law | enforcement, intelligence, and reporting should work. | | In practice, there are massive gaps here and people's attention | -> decision is lower than ever. | | I don't think it's impossible for us to handle deep fakes, but | I sense the same fear you have. I think ultimately it is more | about our attention spans, and the "urgency" we feel to act | quickly, that will be more of our down fall than the ability to | produce fakes more easily. | | You don't in fact need a convincing fake to create a powerful | conspiracy theory. Honestly you only need an emotional | provocation, maybe even some green text on an anonymous web | form. | berdon wrote: | Reminds me of Stephenson's "Fall; Or Dodge in Hell" where all | digital media is signed by their anonymous author and public | keys become synonymous with identities. An entire industry of | media curation existed in the book to handle bucketing media as | spam, "fake", "true", interesting, etc. | narrator wrote: | Surprise Plot Twist: Maybe we're already living in a post-truth | society and you are still sure you know what the truth is. How | would you even know that what you were ferociously defending as | the truth wasn't a lie? What makes you think you're not smart | enough to not fall for lies? | | Largely, I think most people's means of finding the truth is | just to take a vote of the information sources they find | credible and go with whatever they say. I was talking with some | friends about the California propositions a while back. Some of | them were not clear cut which way we should vote on them. | Instead of discussing the actual issue, people just wanted to | know what various authority figures thought. These were not | dumb people I was talking to, and I used to remember an era in | the 90s maybe where you could actually have a reasoned debate | and come to the truth that way. It seems that's obsolete these | days since nobody seems to agree on the basic facts about | anything. | cwkoss wrote: | Disinformation is very common in traditional news media. This | technology just democratizes this tool and allows anyone to | engage in it. | | There will probably be a net increase in disinformation, but | citizens will likely also get better at being skeptical of | currently unquestioned modes of disinformation. | corrral wrote: | > There will probably be a net increase in disinformation, | but citizens will likely also get better at being skeptical | of currently unquestioned modes of disinformation. | | Russia seems to be farther along this path than we are and | every account I've read of their experience of disinfo | isn't that they got better at seeking the truth, but | instead just assume everything's a lie & nothing's | trustworthy, and disappear into apathy. | wildmanx wrote: | > I don't think we'll have the ability to handle this kind of | tech responsibly. | | Makes you also think whether anybody from the Hacker News crowd | working on any contributing tech is acting ethically | responsibly. For myself, I have answered this question with | "no", which rules out many jobs for me, but at least my kids | won't eventually look me in the eye and ask "how could you?" | | Sure it's cool tech. But so was what eventually brought us | nuclear war heads. | brightball wrote: | Honestly, when people have gone so far as to redefine common | words it makes it really difficult to have conversations with | people. | | 1. Hate going from one of the most visceral and obsessive | emotions that exists to being tossed around at everything | | 2. The advent of your truth instead of the truth | | 3. Constant injections of "x" into every existing word | apparently? | | "Womxn in Mathematics at Berkeley" - https://wim.berkeley.edu/ | | This is all before we get people to understand that having a | discussion where some of their points might not be a strong as | they think they are...somehow means you're attacking. | | The world that we have created for ourselves over the last 20 | years is weird. | jkaptur wrote: | Wasn't this bridge crossed when Photoshop became popular? | bberrry wrote: | It takes some level of skill to produce a convincing | Photoshopped image. | BoorishBears wrote: | Does that matter when they stakes are as high as these | arguments always claim? | | If we're doomsaying about a "post-truth society", we're | talking about high-stakes society-scaled skullduggery. | | If you're aiming for that level of disruption, easy | deepfakes vs hard video/photo editing is not an issue, | getting people to trust your made up chain of custody is. | | - | | This is like when people worry about general AI becoming | self-aware and enslaving mankind... the "boring version" of | the danger is already happening: ML models being trained on | biased data are getting embedded in products that are core | to our society (policing, credit ratings, etc.), and that's | really dangerous. | | Likewise, people worry about being able to easily make fake | news, when the real danger is people not being equipped to | evaluate the trustworthiness of a source... and that's | already happening. | | You don't even need a deepfake, you tweet that so and so | said X, write a fake article saying they said X, amplify it | all with some bots, and suddenly millions of people believe | you. | kache_ wrote: | The one unwavering thing about technology is that it doesn't | stop advancing, and we can't use it responsibly. | | The good news is that we've been going through rapid, rapid | tech advancements the past 50 years and we're still here. | mckirk wrote: | The thing I don't like about these 'well people have been | complaining about this forever' arguments is that, it's | entirely possible to have a) people pointing at an issue for | a long time and b) still have that issue get progressively, | objectively worse over time. | | There's that example of people pointing out smartphones might | be bad for children, then someone counters with 'well thirty | years ago people complained about children reading too much | instead of playing outside', with the implication being: | adults of all ages will find some fault with newer | generations, and not to worry so much. | | But just because it is true that adults will probably always | worry about 'new, evil things' corrupting the youth, this | does not mean that the 'new, evil things' aren't getting | _objectively more dangerous_ over time. Today adults would be | happy if children still had the attention span and motivation | necessary to read a book. They'd be happy if they themselves | still had it, actually. | | Graphing the progress of a sinking ship and pointing out that | the downwards gradient has been stable for a while now and we | should therefore be okay is generally not a useful | extrapolation, I would say. | cupofpython wrote: | >Graphing the progress of a sinking ship and pointing out | that the downwards gradient has been stable for a while now | and we should therefore be okay is generally not a useful | extrapolation, | | I like this analogy. I've had similar thoughts for a while | too. Granted I also saw some research that society has been | objectively _getting better_ in a lot of areas people | _think_ is getting worse (like violence, specifically | police abuse) compared to the past. theoretically this is | because we have a lot more information now than before, so | smaller occurrences are generating a larger impression. | | that said, I still very much agree with your point and that | it is very applicable to _specific_ individualized issues. | Saying that people have been concerned for a while and | nothing bad has happened yet is accurate for the situation | where nothing bad will happen, AND the situation that it | was bad then and is worse now, AND the situation where we | are approaching a tipping point / threshold where the bad | will start. | andruby wrote: | > The one unwavering thing about technology is that it | doesn't stop advancing, and we can't use it responsibly. | | While I think that is true in general, I am optimistic that | we've seen at least one technology where we were able to | constraint ourselves from self-destruction: nuclear weapons. | | Of course, nuclear weapons tech is not in reach of | individuals or corporations, which means there are only a | handful of players in this game-theory setting. | sitkack wrote: | > rapid tech advancements the past 50 years and we're still | here | | This is a tautology. At some point the music stops and you | aren't here to make the argument that we are still here. | cupofpython wrote: | not entirely tautological. the probability that something | bad happens tomorrow if we do X today for the first time is | very different than the probability that something bad | happens tomorrow if we do X today GIVEN weve been doing X | every day for 50 years. | | It is still insufficient to say nothing bad will happen, of | course | sitkack wrote: | Thats not the argument. The one you are making is the | same one people make when they conflate weather and | climate. | cupofpython wrote: | Conditional probability applies to many things | beisner wrote: | So there are information theoretic ways to certify that media | is genuine, if you assume trust at least somewhere in the | process. Basically just cryptographic signing. | | For instance, a camera sensor could be designed such that every | image that is captured on the sensor gets signed by the sensor | at the hardware level, with a certificate that is embedded by | the manufacturer. Then any video released could be verified | against a certificate provided by the manufacturer. Of course, | you have to trust the manufacturer, but that's an easier pill | to swallow (and better supported by our legal framework) than | having to try and authenticate each video you watch | independently. | | There are issues that can arise (what if I put a screen in | front of a real camera??, what if the CIA compromises the | supply chain???), but at the end of the day it makes attacks | much more challenging than just running some deepfake software. | So there are things that can be done, we're not destined for a | post truth world where we can't trust any media we see. | notfed wrote: | > a camera sensor could be designed such that every image | that is captured on the sensor gets signed by the sensor at | the hardware level | | A hardware-based private key like this will inevitably be | leaked. | beisner wrote: | Each sensor could have a unique cert | [deleted] | pahn wrote: | I'll probably get downvoted into oblivion for mentioning | blockchain tech, but this might at least help, maybe not in | its current form but... I did not follow that project, but | there do exist some concepts in this direction, e.g. this: | https://techcrunch.com/2021/01/12/numbers-protocols- | blockcha... | beisner wrote: | The idea of having a public record that attests to _when_ | an event happened is interesting, although not sure it has | to be blockchain for it to be useful. | bogwog wrote: | That's helpful for the legal system, but it's not going to | help for attacks designed to cause mass panic/unrest/revolts. | If another US president wants to attempt a coup, it'll be | much more successful if they're competent and determined | enough to produce deepfakes that support their narrative. | | The only way to prevent stuff like that is to educate the | public and teach people how important it is to be skeptical | of anything they see on the internet. Even then, human | emotions are a hell of a drug so idk how much it'd help. | notahacker wrote: | US Presidents have had the ability to make false claims | based on video of something completely different, create | material using actors and/or compromised communications, | stage events or use testimony that information has been | obtained via secret channels from appointees heading up | agencies whose job it is to obtain information via secret | channels for a long time now. | | If anything, recent events suggests the opposite: deepfakes | can't be _that_ much of a game changer when an election | candidate doesn 't even have to _try_ to manufacture | evidence to get half the people who voted for him to | believe his most outlandish claims. | [deleted] | Zababa wrote: | > What will life be like when you can't trust any video or | interview you see because it could be completely fake? | | I don't understand your point. This has been the case for a | while. People were editing photos to remove people in the time | of Stalin. And even before that, you can lie, write false | records, destroy them. | tomgraham wrote: | The good news is that public awareness of potentially | manipulated media is on the rise. Coupled with good laws, good | detection tech - public awareness and media literacy is | important. At Metaphysic.ai, we created the @DeepTomCruise | account on Tiktok to raise awareness. | | We also created www.Everyany.one to help regular people claim | their hyperreal ID and protect their biometric face and voice | data. We think that the metaverse of the future will feature | the hyperreal likenesses of regular people - so we all have to | work hard today to empower people to be in control of their | identity. | SantalBlush wrote: | Creating yet another product to monetize is not a solution, | it's just more of the problem. It incentivizes a perpetual | arms race between fabrication and verification at the cost of | everyday users. No thanks. | tomgraham wrote: | It is free. Protecting individuals' right is more important | that making money! | random-human wrote: | Free but collecting and storing peoples biometric data on | your servers (per the FAQ). How do I know it's not a | clearview ai clone but with easier data gathering? What | is that saying about what the real product is if | something is free? | fxtentacle wrote: | I believe we'll go back to trusting local experts that you can | meet in person to confirm that they are not a bot. | | Because anything online will be known to be untrustworthy. Most | blogs, chat groups and social media posts will be spam bots. | And it'll be impossible for the average person to tell the | difference between chatting with a bot and chatting with a | human. But humans crave social connections and intimate | physical contact. So people will get used to the fact that | whoever you meet online is likely fake and so they'll start | meeting people in the real world again. | | I also predict that some advanced AIs will be classified as | drugs, because people get so hooked on them that it destroys | their life. We've already banned abusive loot box gambling | mechanics in some EU countries, and I think abusive AI systems | are next. We'll probably also age-limit generative AI models | like DALL-E, due to their ability to generate naughty and/or | disturbing images. | | But overall, I believe we will just starting to treat | everything online as fake, except in the rare case that you | message a person which you have previously met in real life (to | confirm their human-ness). | newswasboring wrote: | Your second paragraph is very intriguing. I never really | thought about this. I wonder if people will actually be able | to restrict usage though. Its software, and historically it | has been hard to restrict it. Of course cloud based systems | have two advantages, software is hidden behind the API and | they have really powerful systems. But the former requires a | single lapse in security to leak and latter just requires | time till consumer hardware can catch up. If I use past data | to predict future (which might be a bad idea in this case), | it might be almost impossible to restrict AI software. | formerkrogemp wrote: | I've heard this for years, but software will eventually | face its own regulation and barriers to entry much as | healthcare and accounting have theirs. | userabchn wrote: | I suspect that many chat groups (such as Facebook groups), | even small niche ones, already have GPT-3-like bots posting | messages that seem to fit into the group but that are trained | to provide opinions on certain topics that align with the | message that the organisation/country controlling them wishes | to push, or to nudge conversations in that direction. | fxtentacle wrote: | Aww, that reminds me of the good old IRC days where | everyone would start their visit with !l to get a XDCC bot | listing. | fartcannon wrote: | I want to agree with you, deeply, but the number of people | who fall for simple PR/advertising in today's world suggests | otherwise. | | I think we'd have a chance if they taught PR tricks in | schools starting at a young age. Or at minimum, if websites | that aggregate news would identify sources that financially | benefit from you believing what they're saying. | corrral wrote: | I've long thought that high school should require _at | least_ one course that I like to call "defense against the | dark arts" (kids still dig Harry Potter, right? Hahaha). | | The curriculum would mostly be reasoning, how to spot | people lying with graphs and statistics, some rhetoric, and | extensive coverage of Cialdini's _Influence_. The entire | focus would be studying, and then learning to spot and | resist, tricks, liars, and scam artists. | jayd16 wrote: | When you say "everything online" do you mean every untrusted | source? Surely the genie is out of the bottle on | communication over the web. That local source will have a | website. Because of that I feel like we'll always just have | to be vigilant, just like we always should have been. After | all, local scams still exist. Real humans are behind the | bots. | fxtentacle wrote: | > Real humans are behind the bots. | | Yes, but those humans are usually anonymous and on the | other side of the planet which makes them feel safe. And | that allows them to be evil without repercussions. | | Back in the days, I went to LAN parties. If someone spotted | a cheater, they would gang up with their friends and | literally throw the offender out of the building. That was | a pretty reliable deterrent. But now with all games being | played online, cheating is rampant. | | Similarly, imagine if those Indian call centers that scam | old ladies out of their life savings were located just a | quick drive away from their victims' families. I'm pretty | sure they would have enough painful family visits such that | nobody would want to work there. | | Accordingly, I'm pretty sure the local expert would have | strong incentives to behave better than an anonymous online | expert would. | jayd16 wrote: | To argue that scams didn't exist or weren't a problem | before the internet is pretty indefensible, no matter the | anecdotes. | fxtentacle wrote: | I was merely trying to argue that scams within a local | community would be less severe than scams between | strangers, because they are easier to punish and/or | deter. | wongarsu wrote: | I'm not sure the experts have to be local. I can't be sure | that a random twitter account isn't a bot, but I can be | pretty sure that tweets from @nasa are reasonably | trustworthy. People will form webs-of-trust: they trust one | source, the people viewed as trustworthy by them, etc. Anyone | outside of that will be untrustworthy. | | That's not too dissimilar from what we do today, after all | people have always been able to lie. The problem is just that | if you start trusting one wrong person this quickly sucks you | into a world of misinformation. | | I find your point about regulating AI interesting. We already | see some of this, with good recommendation systems being | harmful to vulnerable people (and to a lesser degree most of | us). This will probably explode once we get chatbots that can | provide a strong personal connection, replacing real human | relationships for people. | freetinker wrote: | HN is a good example (or precursor) of webs-of-trust. Nice | phrase. | mrshadowgoose wrote: | The concerns in your second paragraph can be mostly mititgated | using a combination of trusted timestamping, PKI, | cryptographically chained logs and trusted hardware. Recordings | from normal hardware will increasingly approach complete | untrustworthiness as time goes on. | | The concerns raised in the first paragraph however... the next | few decades are going to be a wild ride. Hopefully humanity | eventually reaches an AI-supported utopic state where people | can wrap themselves in their own realities, without it | meaningfuly affecting anyone else. Perception of reality is | already highly subjective, most of the fundamental issues are | due to resource scarcity/inequality. Most other issues | evaporate once that's solved. | FastMonkey wrote: | I think you can technically mitigate some concerns for the | people who understand that, but practically it's going to be | a very different story. People will believe who/what they | believe, and an expert opinion on trustworthiness is unlikely | to change that. | | I think being in the real world and meeting real people is | the only way to create a real, functional society. Allowing | people to drift away into their own AI supported worlds would | eventually make cooperation very difficult. I think it would | just accelerate the tendency we've seen with social media, | creating ever more extreme positions and ideologies. | thedorkknight wrote: | I don't think it'll be way too much different than it has been | for most of human history. We really only had a brief blip of | having video, which was generally trustable, but keep in mind | that before that for thousands of years it was just as hard to | know truth. | | Someone told you stuff about the outside world, and you either | had the skepticism to take it with a grain of salt, or you | didn't. | gernb wrote: | I was listening to "This American Life" and they had a segment | on someone who setup an site to give you a random number to | call in Russia where you were supposed to give them info about | what's happening in the Ukraine. It was someone shocking to | hear their side of the story, that Russia is a hero for helping | oppresed Russians in Ukraine. | | But then I stepped back and wondered, I'm assuming that the | story I've been told is also 100% correct. What proof do I have | it is? I get my news from sources who've been wrong before or | who have a record of reporting only the official line. My gut | still tells me the story I'm being told in the west is correct, | but still, the bigger picture is how do I know who to trust? | | I see this all over the news. I think/assume the news I get | about Ukraine in the west is correct but then I see so much | spinning on every other topic that it's hard to know how much | spinning is going on here too. | RobertRoberts wrote: | I was asked "What are we going to do about Ukraine!?" And I | said, "It's a civil war that's been going on for almost 10 | years, what is different now?" and their response was "what? | I'd never heard that." And I added, "In 2014 there was an | overthrow of an elected president there and it started the | war." Blank stares. | | I have a friend who traveled to Europe regularly for tech | training, including Ukraine, and he was surprised about how | little people know what is going because people's news | sources are so limited. (mostly by choice I assume) | | No special tech needed to manipulate people, just lack of | multiple information sources? | mcphage wrote: | > what is different now? | | Um. Is this a serious question? | RobertRoberts wrote: | Yes, I didn't know there was an invasion when asked about | Ukraine, but I knew about the past history. (some at | least) | HyperSane wrote: | The president that lost power in 2014, Viktor Yanukovych, | was a Russian puppet who refused to sign the European | Union-Ukraine Association Agreement in favor of closer ties | to Russa. The Ukrainian parliament voted to remove him from | office by 328 to 0. He then fled to Russia. | synu wrote: | It's hard to fathom believing there's nothing new or | relevant happening with the 2022 invasion, or why if | there was a lead-in to the conflict that would be on its | own a reason to conclude that there's nothing to be done | now. | RobertRoberts wrote: | See this is the problem. While I follow plenty of | international news, I didn't know this. | | There is often times just too much to know to fully | understand a situation. So how can anyone form a valid | opinion? | | As a follow up, was the election of Victor Yanukovych | lawful? If not, then why not point out he was a puppet | from a manipulated election? That would be worth a coupe, | but not because you disagree with his politics, that's | just insanity. Look what Trump believed and supported, we | didn't start civil war because Trump wouldn't sign a | treaty and he was accused of being a Russian puppet too. | There is just more to this story than you are letting on. | idontwantthis wrote: | This doesn't bother me that much because evidence isn't | required to convince millions of people that a lie is true. We | already know this. Why make fake evidence that could be | debunked when you can just have no evidence instead? | [deleted] | fleetwoodsnack wrote: | Different instruments can be used to capture different | segments of the population. You're right there are gullible | people who are more likely to believe things with limited or | no evidence. But it isn't necessarily about the most | impressionable people, nor is it about installing sincerely | held beliefs. | | Instead, what may be a cause for concern is simply the | installation of doubt in an otherwise reasonable person | because of the perceived validity of contrived evidence. Not | so much that it becomes a sincerely held belief, but but just | enough that it paralyzes action and encourages indifference | due to ambiguity. | astrange wrote: | This discussion isn't useful because you're assuming people | actually care if something is true before they "believe" | it, which they don't, so they don't need evidence. | "Believing" doesn't even mean people actually hold beliefs. | It means they're willing to agree with something in public, | and that's just tribal affiliation. | [deleted] | fleetwoodsnack wrote: | >you're assuming people actually care if something is | true before they "believe" it, which they don't | | This seems like an assumption too. I know there are | instances like you've described but they're not absolute | nor universal and I accounted for that in my original | comment. | idontwantthis wrote: | Think about how few people believe in Bigfoot when video, | photographs, footprints, eye witness testimony all exist. | | Think about how many people believe in Jesus without any of | that physical evidence. | | If anything, the physical evidence turns most people off. | And I'd argue that most Bigfooters don't even believe in | the physical evidence, but use it as a tool to hopelessly | attempt to convince other people to believe in what they | already believe is true. | mgkimsal wrote: | > You're right there are gullible people who are more | likely to believe things with limited or no evidence | | Often the lack of evidence _is the proof_ of whatever is | being peddled. "No evidence for $foo? OF COURSE NOT! | Because 'they' scrubbed it so you wouldn't be any wiser! | But _I_ have the 'truth' here... just sign up for my | newsletter..." | mc32 wrote: | I think a bigger question is whether reputable sources --those | people trust for whatever reason, would use this technology to | prop up ideas and or to create narratives. | | I don't think it's far-fetched. We've already seen where videos | are misattributed[1] to stoke fear or to promote narratives | --by widely trusted news sources. | | [1] This was foreshadowed with "Wag the Dog" but happens often | enough in the media today that I don't think use of "deepfake" | technology is beyond the pale for any of them. | ketralnis wrote: | It almost doesn't matter now that people have fractured on | which sources they consider reputable. Trump called a CNN | reporter "fake news" and presumably his followers think of | them the same way I think of Fox. I absolutely think that Fox | would use technology to lie, and I'm sure Fox fans think that | "the liberal media" would. So people are going to think that | reporting is fake whether or not it is | mc32 wrote: | Wasn't 'The Ghost of Kiev' almost entirely fake but the | news carried it as real? | ketralnis wrote: | I don't know. How would you "prove" it? Google it, and | look for what other people that agree with you think? | mc32 wrote: | Well... | https://www.washingtonpost.com/world/2022/05/01/ghost-of- | kyi... admitted by Ukrainians themselves... | ketralnis wrote: | According to an article you found online. That's exactly | my point, if we can't trust news sources then we can't | really know anything. Because of my aforementioned | distrust of Fox News, if it were written by them I'd | dismiss that article out of hand placing no truth value | on it either way. I'd expect somebody that distrusts CNN | to do the same if it were written by them. | | "It's confirmed!", "They admitted it!", and other | unprovable internet turns of phrase in comments sections | are really just "I believe it because somebody I trust | said so" and that only has weight as long as trust has | weight. | mc32 wrote: | If the accused admit to something it's more believable | than the alternative (that they were forced into false | admission). | | So in this case if the Ukrainian government admit to | making things up then I would think it's believable that | they made something up for the sake of propaganda. We can | also check more independent sources --read Japanese news, | or Indian news sources, etc. | ketralnis wrote: | We don't know that the accused admitted to anything. We | know that the Washington Post says that they did. The | world becomes very solipsistic when you lose trust in | reporting. | GuB-42 wrote: | There has never been a truth society. | | This tech will certainly be used to frame someone for a crime, | like I am sure Photoshop was used in such a way, and thousands | of other techniques. And modern technology offers counters. It | is an arms race but because of the sheer amount of data that is | collected, I think that truth is more accessible than ever. The | more data you have, the harder it is to fake and keep | consistent. | fxtentacle wrote: | "Cheerleader's mom created deepfake videos to allegedly | harass her daughter's rivals" | | "The girl and her coaches were sent photos that appeared to | depict her naked, drinking and smoking a vape pen, according | to police" | | https://abcnews.go.com/US/cheerleaders-mom-created- | deepfake-... | jl6 wrote: | I don't know, it seems like the existence of widespread, easy | photo/video/audio faking technology could be a really strong | argument for dismissing any purported photo/video/audio | evidence. | | Wouldn't it be funny if deepfakes destroyed the blackmail | industry? | kleer001 wrote: | Thankfully things that a real are very cheap to follow up on. | Questioning some security footage? No worries there's 100+ | hours of it to cross check with the footage and three other | cameras too. | | IMHO, consilience and parsimony will save us. | VanillaCafe wrote: | The real problem isn't the veracity of the information, but the | consensus protocol we use to agree on what's true. Before the | internet, we were more likely to debate with their neighbors to | come to an understanding. Now, with the large bubbles we can | find ourselves in, afforded by the internet social media, we | can find a community to agree on anything, true or not. It's | that lack of challenge that allows false information to | flourish and is the real problem we need to solve. | whimsicalism wrote: | I would be curious if false information is actually more | common now. It seems like people regularly believed all sorts | of false things not too long ago. | redox99 wrote: | People with some degree of knowledge already know that any | photo could be photoshopped. People that don't care will | blindly trust a picture of someone with a quote or caption | saying whatever, as long as it fits their narrative. | | This has been the case for photos for almost 2 decades. The | fact that you can now do it with video or audio doesn't change | that much IMO. | hutrdvnj wrote: | I think it does, because while you obviously couldn't trust | images since two decades or so, you could resort to video | which wasn't easy to believably deep fake until recently. But | if everything online could be a deep fake, how can you find | out the truth? | whimsicalism wrote: | Videos can be faked too, it is just cheaper now. | makapuf wrote: | It's called special fx, is more than a century old | (people are now aware it's fake but remember the word is | that the train coming in la ciotat movie made people run | out of the movie theatre). | micromacrofoot wrote: | Photos have been altered for much longer than 2 decades. | Think of airbrushing models in magazines (used to be literal | airbrushes painting over photos). This has had a serious | impact on our perception of beauty and reality. | MarcoZavala wrote: | WanderPanda wrote: | I think society will adapt within 1 generation. The tech is | already there (signing messages with asymmetric encryption) | toss1 wrote: | And how many use or will use the tech, and how many of those | will use it competently, and how many of those are competent | to validate and know that their checking technology has not | been compromised (e.g., hacking or distributing bad | authenticity checkers and/or certs like hacking or | distributing bad crypto-wallets)? | bee_rider wrote: | End-to-end encryption was a giant pain in the butt that | required dinking around with PGP or whatever, but now it is | a pretty mainstream feature for chat apps (once they | figured out how to monetize despite it). Tech takes a while | to trickle down to mainstream applications, but it'll get | there if the problem becomes well known enough. | toss1 wrote: | I agree that e2e encryption is becoming more widespread | and "user friendly". | | However, the friendliness seems inversely proportional | with the ability of the users to detect that their tool | is failing/corrupted/hacked/etc. So while we might have | more widespread tools, we also have a more widespread | ability to give a _false_ sense of security. | nathias wrote: | you could never trust it, now you'll know you can't trust it | EGreg wrote: | Is this what Matterport 2 app uses? | | When you take many photos of a scene or indoor place and it's | stitched together? | | Can this be used for metaverses? | | ALSO why not synchronize videos of the same event and make an | animated 3d movie from 2d ones !!! Great as a "disney ride | metaverse" | | Who is doing that in our space? | getcrunk wrote: | I just started playing cyberpunk 2077. Spoilers: | | The idea of the "black wall" to keep out bad ai comes to mind. | Not arguing for it but just acknowledging that maybe one day will | all have to live in walled gardens to stay safe from rouge ai's | or rather rouge actors using powerful ai's | henriquecm8 wrote: | One thing I've always wondered about that, they explained where | those are running, but how they still have autonomy? Are they | producing they own power? How about when they need to replace | hardware? | | I am not saying it's impossible, but I would like to see that | part being explored, even if it's in the tie-ins comics. | echelon wrote: | What's the difference in NeRF from a classical photogrammetry | pointcloud workflow? It seems like the representation and outputs | are identical. | | Why would you prefer NeRF to photogrammetry? Or vice versa? | randyrand wrote: | Nerfs can represent reflection, speculars, refractions. | | They also are proving to be faster, more accurate, etc | | The input data is the same. Nerfs have the chance of requiring | less input data. | flor1s wrote: | Neural Radiance Fields are a technique from the neural | rendering research field, while photogrammetry is a research | field on its own. However these are just turf wars and in | practice there is a lot of overlap between both fields. | | For example, most NeRF implementations recommend the use of | COLMAP (traditionally a photogrammetry tool) to obtain camera | positions/rotations that are used alongside their images. So | this multi-view stereo step is shared between both NeRF (except | a few research works that also optimize for camera | positions/rotations through a neural network) and | photogrammetry. | | After the multi-view stereo step in NeRF you train a neural | renderer, while in photogrammetry you would run a multi-view | geometry step/package that uses more traditional optimization | algorithms. | | The expected output of both techniques is slightly different. | NeRF produces renderings and can optionally export a mesh | (using the marching cubes algorithm). Photogrammetry produces | meshes and in the process might render the scene for editting | purposes. | natly wrote: | I was initially annoyed by this title but now I'm gonna switch my | perspective to being happy that ideas like this are floating | around since it acts as a really cheap signal to tell if someone | knows what they're talking about or not when it comes to ML. | nathias wrote: | can't wait until deepfakes completely revolutionizes people's | relation to information ___________________________________________________________________ (page generated 2022-06-27 23:01 UTC)