[HN Gopher] Google, Meta, others will have to explain algorithms...
       ___________________________________________________________________
        
       Google, Meta, others will have to explain algorithms under new EU
       legislation
        
       Author : niklasmtj
       Score  : 255 points
       Date   : 2022-04-23 09:56 UTC (13 hours ago)
        
 (HTM) web link (www.theverge.com)
 (TXT) w3m dump (www.theverge.com)
        
       | jasfi wrote:
       | If you read the article they're asking for more than explaining
       | algorithms. Overall they want the the tech providers to be
       | responsible.
       | 
       | Explaining algorithms could, in theory, give away a competitive
       | advantage. However fairness to users seems to be a priority in
       | this decision.
        
         | omegalulw wrote:
         | I would love to first see a technical definition of fairness
         | from EU that can be used to evaluate algorithms. That is a non-
         | trivial detail often overlooked from these discussions.
        
           | rtsil wrote:
           | Not technical, but fairness is the opposite of "our algorithm
           | is so complicated that we can't prevent it from penalizing
           | you even if you are not at fault. Unless you reach the top of
           | HN, in which case we will manually intervene to fix things."
        
         | pwdisswordfish9 wrote:
         | > Explaining algorithms could, in theory, give away a
         | competitive advantage.
         | 
         | Which is good. We could use some more competition on the
         | market.
        
         | otherotherchris wrote:
         | Things like "fairness" aren't defined in the legislation and
         | will be determined in smoke filled rooms by shadowy moneyed
         | interests.
         | 
         | Ordinary users will get censored. By the courts, by unelected
         | regulators, and by Big Tech AI zealously nuking content to
         | avoid arbitrary fines. It's content ID on steroids.
        
           | jasfi wrote:
           | I agree that it could get out of hand. We'll have to wait and
           | see how it turns out. Since this is an EU law I wonder if it
           | applies to content hosted on EU servers only, or any content
           | that shows up in their users' results.
        
             | otherotherchris wrote:
             | Platforms are responsible for everything shown to a user
             | inside the EU.
             | 
             | I suspect that Google and Facebook will not offer country
             | specific blocklists like they do for Nazi content in
             | Germany. If Hungary bans LGBTQIA content, it'll disappear
             | in France. Europe can then have an argument about how they
             | "really really not really" believe in free speech.
        
               | anothernewdude wrote:
               | _If_ they do business in the EU. Otherwise this is
               | without teeth entirely.
        
               | barrucadu wrote:
               | I mean, yes? That seems obvious?
               | 
               | EU law applies to companies which operate in the EU.
        
               | jasfi wrote:
               | I am worried about the term "disinformation" since that
               | can be really subjective. On the other hand anti-vax
               | content is harmful, to me, so there's no easy answer.
        
               | [deleted]
        
         | ethbr0 wrote:
         | >> _" Large online platforms like Facebook will have to make
         | the working of their recommender algorithms (e.g. used for
         | sorting content on the News Feed or suggesting TV shows on
         | Netflix) transparent to users. Users should also be offered a
         | recommender system "not based on profiling.""_
         | 
         | Both of those seem like good ideas and progress. The non-
         | profiled recommender system option especially!
         | 
         | It's also really bothered me that tech companies of sufficient
         | size can discriminate against legally-protected classes because
         | "algorithms are complicated" and government regulators haven't
         | pushed.
         | 
         | I'm not a fan of regulating design or use, but I'm a huge
         | proponent of requiring transparency and detail on demand.
         | 
         | We'll see how willing the EU is to levy fines for breaches.
         | 
         | It's no doubt a consequence of most huge tech companies being
         | American, but it's been refreshing to see the repeated "We have
         | a law; You clearly broke it; Here's your fine" follow-through
         | thus far from EU enforcement.
        
           | dmitriid wrote:
           | > We'll see how willing the EU is to levy fines for breaches.
           | 
           | It has been very slow with GDPR, I expect it to be even
           | slower here.
        
             | ohgodplsno wrote:
             | Google is rolling out one-click cookie rejection as a
             | result of gigantic fines threatened by the French CNIL.
             | Having already been slapped with 90M and 60M, seems like
             | there's not much of a need for fines. They know Europe
             | isn't playing around.
        
             | ethbr0 wrote:
             | https://www.privacyaffairs.com/gdpr-fines/
        
         | pmoriarty wrote:
         | _" Explaining algorithms could, in theory, give away a
         | competitive advantage."_
         | 
         | Why should anyone care if they have a competitive advantage?
         | 
         | If anything I want them to have a disadvantage, lose money, and
         | go out of business.
        
         | vmception wrote:
         | Nevada Gaming Control Board requires source code of all the
         | casino games
         | 
         | Easy to see this concept expanding
        
           | leksak wrote:
           | This deserves more attention as it does set a good precedent
        
       | Irishsteve wrote:
       | The new regulation requires a hand-wavey style explanation i.e
       | build a retrieval / ranking / matching algorithm that learns from
       | customer clicks and considers blah blah.
       | 
       | There will be no explanation of the actual algorithm.
        
         | postsantum wrote:
         | No explanation - no market. If you don't like it, you can run
         | your business in another contract jurisdiction. I hope these
         | companies will taste their own medicine, we desperately need
         | competition
        
           | johnny22 wrote:
           | the person you're replying to is saying that their
           | explanation will be trivial and thus useless in determining
           | the responses.
           | 
           | I don't know if that's true or not myself though, since I
           | haven't read myself.
        
         | rm_-rf_slash wrote:
         | Probability ranking principle will never not be the foundation
         | for recommender systems. Any statistical model flowing from PRP
         | will be more or less the same regardless of whether the
         | architecture is neural, boosted trees, etc.
         | 
         | However, if regulation required companies to disclose all of
         | the data goes into those models, how they acquire it (tracking
         | browser/app behavior, purchase from 3rd parties), and so on,
         | that would be the real game changer for consumer privacy and
         | protection.
        
       | throwaway4323r wrote:
       | unless you can reproduce this is not going to cut. I dont think
       | this is going to help. It only creates more useless software jobs
        
       | FpUser wrote:
       | I am curious how would they "explain" AI algorithms where it is
       | impossible to explain how / why the decision has been made.
        
         | kazamaloo wrote:
         | Not that the algorithms are impossible to explain but in some
         | cases the real explanations might require explanations, too.
         | But I think companies will probably get away with hand-wavy
         | explanations like you get this recommendation because you
         | watched this movie neglecting all the
         | sourcing/ranking/filtering workflows.
        
       | neatze wrote:
       | If I had to guess it probably similar to SEC or NYSE required
       | explanation when you do suspicious trades.
        
       | narrator wrote:
       | "We run it through our deep learning model. Here's 50 gigabytes
       | of neural net weights."
        
         | mnau wrote:
         | > Here's 50 gigabytes of neural net weights.
         | 
         | .. from few months ago. Weights change daily, most likely
         | updated by another NN.
         | 
         | I guess it's nice that lawmakers understand that at some point
         | these companies used algorithms to search or sort stuff, but
         | industry has already moved to another level. We might be able
         | to explain specific result of neural networks (Shapely values
         | or something like that), but the actual algorithm (=NN)... no
         | way.
        
       | ernirulez wrote:
       | EU is becoming more like an authoritarian state. They put
       | constraints on companies but allow governments to have full
       | control and surveillance over their citizens. It's so hopocrit
        
         | naoqj wrote:
         | Not trying to defend the EU here, but isn't this true of every
         | government?
        
         | smrtinsert wrote:
         | You're saying governments shouldn't have the ability to govern
         | its people, or inquire as to how global scale companies impact
         | their citizens? That doesn't make any sense whatsoever.
        
         | frereubu wrote:
         | "... allow government to have full control and surveillance
         | over their citizens."
         | 
         | This is hyperbole.
         | 
         | This suggestion is the logical next step of the part of GDPR
         | where it says that citizens should be able to understand how
         | automated decisions are made about theme and their data. This
         | is about transparency for citizens, not governments dictating
         | how algorithms should work.
        
         | rtsil wrote:
         | As a EU citizen, I fully expect my country and the EU to "put
         | constraints" (i.e. regulate) on companies, so everything seems
         | to work as expected.
        
         | Epa095 wrote:
         | I don't see how it's hypocritical to give democratically
         | elected governments different possibilities that random private
         | companies.
         | 
         | It's government's job to put constraints on companies, stopping
         | them from becoming the absolute assholes they become if they
         | have no limitations. That does not make them authoritarian.
        
           | mediascreen wrote:
           | Sure. On the other hand you can usually chose which companies
           | to interact with. When it comes to governments the
           | relationship is not optional. Your government usually have
           | more ways to affekt your life than most private companies has
           | (like no fly lists).
           | 
           | A few years ago the true extent of the Swedish program for
           | tracking left wing sympathisers became known. It ran from the
           | sixties up until 1998. For example, if your car was seen
           | outside of a left wing publication you could end up on a list
           | somewhere. That caused you to be automatically excluded from
           | 5-10% of all jobs without you never finding out about it
           | until 20-30 years afterwards. Imagine wanting to become a
           | police officer, a pilot or and engineer and never
           | understanding that the reason you didn't get an interview was
           | because you had parked in the wrong spot one day years
           | before. Or that your sister briefly dated a left wing
           | journalist at some point.
        
           | netizen-936824 wrote:
           | Some people are sad that they can't set up businesses which
           | exploit the populace as easily in the US I guess?
        
           | rafale wrote:
           | The tyranny of majority is a real threat. Power should be
           | "shared" (or under contention). Companies just want to take
           | your cash, governments can take your freedom.
        
       | arnvald wrote:
       | > The greater the size, the greater the responsibilities of
       | online platforms
       | 
       | > as a rule, cancelling subscriptions should be as easy as
       | signing up for them
       | 
       | Overall I like these principles, but we'll see in a few years how
       | they're enforced in practice. It's been 4-5 years since we've had
       | GDPR and I still see sites that require tens of clicks to disable
       | all advertising cookies (and the most I've seen was 300+ clicks).
       | Even Google only this week announced they'll add "reject all"
       | button to their cookie banners.
       | 
       | I expect it'll be similar in this case, companies will do bare
       | minimum to try to stay compliant with the regulation, and it will
       | take a few years to see real differences, but I hope it's at
       | least a step in the right direction.
        
         | mijamo wrote:
         | 4-5 years is nothing for the law. You have murders from 15
         | years ago still in process in courts. But eventually things
         | settle. It just takes time. It's a bit like ents
        
         | FollowingTheDao wrote:
         | > as a rule, cancelling subscriptions should be as easy as
         | signing up for them
         | 
         | Before I sign out for any service this is the first thing I
         | check.
        
       | wave-creator wrote:
       | Do you think the EU will enforce the law for non-US and non-EU
       | companies like TikTok will disclose them? It will be interesting
       | to see if they will uphold the law equally to all.
        
         | manquer wrote:
         | All the big guys roughly know what kind of pipelines every one
         | has , they hire from each other etc.
         | 
         | The level of disclosure is not going to break a lot of
         | competitive advantage.
         | 
         | basically need to say what input sources and feedback they use
         | and modular blocks on what different steps go into the pipe,
         | nobody is asking them to expose the actual weights of billion
         | parameter ml model they all probably have .
         | 
         | Even if hypothetically they did expose that level of detail it
         | is useless for regulators as they don't have resources to run
         | the model , and testing a model for side effects in depth is
         | hard .
        
       | jdrc wrote:
       | Requiring an alternative to algorithmic sorting (chronological)
       | is good even though most sites do it already. "Explaining the
       | algorithms" sounds like an impossible-to-implement, feel-good
       | clause.
       | 
       | Requiring transparency for bans and censorship though will
       | probably have a major effect if people start asking nosy
       | questions and exposing corporate and government abuses of power.
       | Many EU governments will regret that users can expose them , that
       | will be fun to watch. It will also make it very hard for
       | companies like reddit to function: could reddit be legally liable
       | for actions of its moderators?
       | 
       | the other clauses are the typical wishful thinking by EU
       | legislators who think that you can legislate the solution to
       | unsolved or unsolvable tech problems
        
       | frereubu wrote:
       | IANAL, so happy to be corrected, but my understanding is that EU
       | and US law work in quite different ways. EU law sets general
       | rules, and law courts decide what that means with reference to
       | existing legal precendents. US law is very, very specific about
       | what each clause means and how it should be interpreted.
       | 
       | Every time I see these kinds of discussions I wonder if quite a
       | few of the disagreements are due to e.g. US commenters worried by
       | the relative lack of specific details.
        
         | krastanov wrote:
         | Did you flip EU and US in your comment? My understanding is the
         | exact opposite of what you wrote:
         | 
         | - US, common law, https://en.wikipedia.org/wiki/Common_law
         | 
         | - EU, civil law,
         | https://en.wikipedia.org/wiki/Civil_law_(legal_system)
         | 
         | Citing: Civil law is a legal system originating in mainland
         | Europe and adopted in much of the world. The civil law system
         | is intellectualized within the framework of Roman law, and with
         | core principles codified into a referable system, which serves
         | as the primary source of law. The civil law system is often
         | contrasted with the common law system, which originated in
         | medieval England, whose intellectual framework historically
         | came from uncodified judge-made case law, and gives
         | precedential authority to prior court decisions.
        
       | dontblink wrote:
       | Regarding the rules surrounding fake information, I wonder why
       | the be EU hasn't taken a similar stance against Fox News
       | equivalencies?
        
         | dehrmann wrote:
         | Rules around fake information scare me because they're a limit
         | on speech, and as Russia has recently shown, fake information
         | is anything a dictator doesn't approve of.
        
           | Broken_Hippo wrote:
           | _Russia has recently shown, fake information is anything a
           | dictator doesn 't approve of._
           | 
           | This isn't an issue of "limits on speech", but rather,
           | another reminder that one shouldn't enable folks to become
           | dictators. Not having some reasonable limits on actual
           | misinformation makes us all less free, however, because we
           | cannot put our trust in some organizations.
        
             | dehrmann wrote:
             | > Not having some reasonable limits on actual
             | misinformation makes us all less free
             | 
             | This is a step towards "freedom is slavery."
        
               | hnbad wrote:
               | If I had a penny for every time someone referenced George
               | Orwell without understanding his politics or even having
               | read his books, I would have a lot of pennies.
               | 
               | I take it you believe the tolerance paradox also gives
               | off 1984 vibes?
        
         | walkhour wrote:
         | Because we wouldn't have any media whatsoever to consume.
        
       | vampiretooth1 wrote:
       | What actually constitutes a full explanation of the algorithm?
       | Article doesn't get into this enough, it mentions a high level
       | overview is required but not much else. I can imagine that it's
       | not going to require sharing the codebase or IP, of course.
        
       | somewhereoutth wrote:
       | A good start. However let's go further, simply ban personal
       | tracking and personalized algorithmic feeds. This would combat
       | the echo chamber effect and social media can become a broad
       | community experience, like TV and newspapers. It would also
       | cripple tech advertising revenues, thus redressing the balance
       | with traditional media.
        
         | tester89 wrote:
         | > personalized algorithmic feeds
         | 
         | But there are good uses, like for music. I can't really think
         | of a downside for music tbh, it's not like music tends to
         | spread extremism, and on the upside lesser known artists have a
         | better shot at being discovered through the algorithm.
        
         | vimsee wrote:
         | I usually have few reasons for being concerned about the
         | future.
         | 
         | I think wars (even with the on-going war that Russia started),
         | climate issues (even with the high consumption present today)
         | and poverty (even with many countries still in it) will all
         | have a trend of declining. However, this echo chamber fueled
         | with miss-information is one of the things I care for.
         | 
         | I am so happy the EU has power and will to make good changes
         | that gives mutual benefit to everyone when other parts of the
         | world does not.
        
         | marcinzm wrote:
         | So should I be prevented from following only anti-capitalistic
         | people on twitter or only following right wing subreddits?
         | Should people also be banned from subscribing only to a single
         | newspaper versus a mix of newspapers with different political
         | leanings? What about looking at only socialist web pages that
         | link to other such web pages. Should web pages we forced to
         | link to pages with other political leanings?
        
           | somewhereoutth wrote:
           | No, that's not what I meant. More that it should be a
           | conscious decision to knowingly consume content with a
           | particular bias - as you do when you pick up a certain
           | newspaper or turn over to a specific tv channel, as opposed
           | to being algorithmically presented with a stream of content
           | that may veer in a direction without you being aware.
        
         | hairofadog wrote:
         | One of the things that frustrates me about discussions of
         | censorship here on HN is that there's a _lot_ of intense focus
         | on censorship via deleting a tweet or Facebook post, but no
         | focus given to the more insidious problem of censorship by
         | algorithm.
         | 
         | I am wholeheartedly in favor of a free marketplace of ideas
         | where (we would hope) good ideas win out over bad, but as it
         | is, once you're deemed by an algorithm to be susceptible to a
         | certain category of extremist information, that's all you're
         | ever going to see again; the competing ideas are never going to
         | have a chance.
         | 
         | Algorithmic distribution of ideas is sorta like distributing
         | ideas via gasoline-powered leaf blower directly to the face. I
         | am free to speak my competing ideas, and so technically I
         | haven't been censored, but no audience is going to hear me over
         | the leaf blower.
        
           | visarga wrote:
           | We need to get some level of control over the criteria for
           | ranking and filtering. A third one is the UI - it is the
           | place where all sort of dark patterns hide.
           | 
           | I'd like to see the browser put in a sandbox and its
           | inputs/outputs sanitised and de-biased before being presented
           | to the user. Could also protect privacy more. We need more
           | browser innovation. A neural net should be in every browser
           | ready to apply semantic rules.
        
         | pessimizer wrote:
         | I don't think people should be forced into the public square by
         | law. If you want to live in an echo chamber, you should be able
         | to. We don't forcibly close convents. If I want to choose the
         | _" Smart Feed(tm),_ although I can choose not to, that should
         | be my choice to make.
         | 
         | I don't know TikTok, but people seem to like its choices.
        
           | somewhereoutth wrote:
           | I think the distinction is that generally it should be a
           | _conscious_ choice to be in the echo chamber - and not the
           | easy unknowing default choice (if you even have a choice) for
           | smart feeds.
        
         | ben_w wrote:
         | I've seen the difference between what YouTube presents to me
         | when I'm logged in v.s. when I'm on a clean computer it can't
         | associate with me, and I do value the personalisation -- when
         | not logged in it shows me a hundred duds for every one thing I
         | care about, and logged in it's about 50:50.
         | 
         | How much of this improvement is a mysterious machine learning
         | algorithm and how much is it just looking for new things from
         | my subscription list, I'm not sure, and that's important: being
         | trapped in a torrent of self-reinforcing falsehoods is
         | something I fell for in my teenage-goth-New-Age phase, which
         | Carl Sagen condemned in _The Demon-Haunted World_ , and which
         | people in general have been falling for with every sychophant
         | and propagandist from soothsayers to tabloids telling them what
         | they want to be so.
        
           | PolygonSheep wrote:
           | > I'm not sure, and that's important: being trapped in a
           | torrent of self-reinforcing falsehoods is something I fell
           | for in my teenage-goth-New-Age phase, which Carl Sagen
           | condemned in The Demon-Haunted World.
           | 
           | Genuinely curious here: how can you tell you've escaped one
           | set of self-reinforcing falsehoods while being sure you
           | haven't fallen into another, different set?
        
             | ben_w wrote:
             | You can't be certain of not falling into a different set of
             | false beliefs, but you can look for inconsistencies in your
             | beliefs and be less and less wrong.
             | 
             | Spotting inconsistencies in my beliefs is what pushed me
             | out of New Age mode, and ironically what pushed me into it
             | in the first place (from Catholicism).
        
           | EnderShadow8 wrote:
           | An alternative might be for personalisation to be opt-in
           | rather than opt-out even when signed in with an account
           | (which shouldn't even be necessary for many services anyway)
        
         | cush wrote:
         | I think it's fair to use personal data collected on the same
         | site. Without it, most sites would be rendered useless.
        
           | somewhereoutth wrote:
           | Indeed - that would be a legitimate use of cookies etc. In
           | fact, if that were enforced, we can get rid of the annoying
           | cookie warnings.
        
         | BeFlatXIII wrote:
         | How does a ban on personalized algorithmic feeds work if each
         | user is subscribed to a different set of others?
        
           | somewhereoutth wrote:
           | "news from friends" might be ok if it is presented without
           | algorithmic curation - i.e. strictly on time order or
           | similar.
        
       | tjbiddle wrote:
       | > "Dark patterns" -- confusing or deceptive user interfaces
       | designed to steer users into making certain choices -- will be
       | prohibited. The EU says that, as a rule, cancelling subscriptions
       | should be as easy as signing up for them.
       | 
       | This is an excellent addition.
        
         | eli wrote:
         | It's a great idea but my understanding is that they have not
         | yet defined the term. And that sounds very hard to define.
        
           | aaaaaaaaaaab wrote:
           | "I know it when I see it."
        
             | bryanrasmussen wrote:
             | in common law systems I know it when I see it is good
             | enough, but I believe most of EU is under a Napoleonic
             | system where you should define what you mean.
        
               | [deleted]
        
               | jamesrr39 wrote:
               | Nitpick: most of it falls under the Civil law system,
               | some of which is Napoleonic. Wikipedia has a pretty nice
               | map of the breakdown: https://en.wikipedia.org/wiki/Civil
               | _law_(legal_system)#/medi...
        
               | XorNot wrote:
               | The requirement of "subscribe is as easy as unsubscribe"
               | is a metric which you could argue about in court, but
               | would be very hard to game.
               | 
               | i.e. if signup is "email and credit card number" then
               | you're going to be hard pressed to explain why a similar
               | option to cancel does not exist and isn't accessible in
               | as many clicks, with equivalent screen real-estate usage.
        
               | q-big wrote:
               | > i.e. if signup is "email and credit card number" then
               | you're going to be hard pressed to explain why a similar
               | option to cancel does not exist
               | 
               | So you argue that to cancel a subscription, you should
               | have to provide your credit card number again. If a check
               | on the credit card fails for some obscure reason, you
               | cannot cancel your subscription.
               | 
               | This is what "subscribe is _as easy as_ unsubscribe "
               | also means.
        
               | kaba0 wrote:
               | I mean, that is precisely the court's job to interpret
               | the law. Your take is just a deliberately twisted one, it
               | wouldn't stand a chance in a court setting the same way
               | as a willful offense can't be defined that precisely, yet
               | there is generally no problem with it.
        
               | XorNot wrote:
               | Sure: but if it fails, then the card is invalid, and the
               | card no longer can be billed. Again - you wouldn't get
               | away with saying "well we couldn't verify the number" as
               | standard practice - you'd just get sued and then
               | punitively fined if it was found to be a lie.
        
             | eli wrote:
             | I'm just not a big fan of laws where just about everyone
             | could arguably be breaking them in some small way. That's a
             | lot of faith to put in regulators to always act honorably.
             | 
             | Vaguely worded laws can also lead conservative corporate
             | counsels to make decisions like geoblocking all of the EU
        
         | ugjka wrote:
         | I hope this makes Google Pay app subscription cancellations
         | actually cancel them instead of postponing them for 3 months or
         | so
        
         | bratwurst3000 wrote:
         | Windows is fuckinh the worst at this. The whole system
         | experience is at some point ,, please login to the mircrosoft
         | produkt you never signed for" or whatever new noninteresting
         | feature they have. That there isnt something like a windows
         | version striped off that stuff is a shame .
         | 
         | Oh and firewall or defender that puts a big !! Everywhere so it
         | seems that my system will explode anytime
         | 
         | Are they aware that people use it for working?
        
         | bgro wrote:
         | Remember when Xbox let you sign up for Live online, but you had
         | to do a 3 hour interrogation on the phone to cancel? And calls
         | would cost like 25 cents a minute or something crazy.
         | 
         | Or the auto renewing subscriptions that either cancel your
         | service immediately the second you turn off auto renew, even if
         | you paid for the current time allotment, or they just prevent
         | or ignore your request to not renew.
         | 
         | I feel like reverse charging didn't exist back then.
         | 
         | There's also entitled devs that say your email domain or VOIP
         | number isn't good enough when signing up for their service.
         | There's no reason for anybody to use an email from their
         | perfect in test whitelist of gmail or Microsoft domains... And
         | why would anybody ever have a voip number unless they were a
         | terrorist?
        
           | Terretta wrote:
           | Or DirecTV recently:
           | 
           |  _"Hey we couldn't process your card due to a temporary error
           | so we went ahead and cancelled your $59 for AllTheThings plan
           | you had for the last 10 years as a loyal customer. We're very
           | much not at all sorry that plan isn't available any more. Now
           | AllTheThings costs $129, but don't worry, just click to
           | reactivate, we'll try your card again." ... "AllTheThings
           | processed successfully for $129, thank you for your custom."_
        
           | SantalBlush wrote:
           | There's also this classic attempt at cancelling Comcast
           | service. [1] What a nightmare.
           | 
           | [1] https://m.youtube.com/watch?v=yYUvpYE99vg
        
         | jpz wrote:
         | Sounds extremely subjective. How do you measure it and where do
         | you draw the line? All marketing is somewhat coercive.
         | 
         | How do you get economic and business growth (things which are
         | good for people - jobs and employment) without marketing and
         | advertising?
        
           | nathias wrote:
           | It's objective its just very widespread. Amazon is probably
           | the greatest offender, but most of the platforms and BigTech
           | is just dark patterns all the way down.
        
         | einszwei wrote:
         | I am very tired of the cookie/tracking popups on many websites
         | that don't have option to "reject all" but just "accept all"
         | and "customise". Main example being Google Search.
         | 
         | Looking at this, I am hopeful but not too optimistic.
        
           | bratwurst3000 wrote:
           | The << i dont care about coockies >> plugin for firefox is
           | superb at geting rid of that problem
        
             | layer8 wrote:
             | I wouldn't count on sites not tracking you until you
             | actually saved your "custom preferences".
        
               | alwayslikethis wrote:
               | Of course you can't.
               | 
               | My recommendation:
               | 
               | 1. Install "I don't care about cookies"
               | 
               | 2. Install "Temporary containers"
               | 
               | This requires that you use special containers for things
               | you do wish to have cookies for such as HN for the login.
               | Other than that, you can safely click accept for all
               | websites, since it won't persist anyways.
        
           | nottorp wrote:
           | > I am very tired of the cookie/tracking popups on many
           | websites that don't have option to "reject all" but just
           | "accept all" and "customise". Main example being Google
           | Search.
           | 
           | And The Verge on this very article :)
        
             | andrei_says_ wrote:
             | The people writing the articles are different from the MBAs
             | forcing the financial and technological decisions.
             | 
             | "Integrity" has different meanings for each group. For the
             | latter, the meaning is likely closer to "bring in enough
             | revenue to keep the publication running." Applying dark
             | patterns does not conflict with this.
        
               | nottorp wrote:
               | Well it's fine with me. I only open Verge links when
               | they're on HN and the title feels interesting. Which is
               | pretty rare.
        
           | adam0c wrote:
           | use a vpn, any incognito browser, stop using google. simple.
        
           | jsnell wrote:
           | From a couple of days ago: https://blog.google/around-the-
           | globe/google-europe/new-cooki...
        
             | einszwei wrote:
             | Wow, best UX change from Google in sometime now.
        
               | hnbad wrote:
               | Don't let the prose fool you. They're doing this because
               | what they did before was in violation and the walls were
               | closing in.
               | 
               | This reminds me of supermarkets in Germany loudly
               | announcing that they would abandon plastic bags to save
               | the environment ... a few weeks before legislation came
               | into effect banning them from selling plastic bags.
               | 
               | Why wait until you're potentially facing fines if you can
               | move slightly ahead and sell it as a voluntary good thing
               | you do for your users/customers?
        
         | lin83 wrote:
         | Amazon's attempts to get users to unwittingly sign up to Prime
         | is one of the most egregious examples I encounter on a regular
         | basis. As a European I cannot wait to see it gone.
        
           | psyc wrote:
           | After the last iOS update, Apple nagged the shit out of me to
           | setup Apple Pay, for two days. No way to say 'fuck off' -
           | only 'remind me'. No obvious way to stop the nagging. Finally
           | I gave them just the tip, and then pulled out before the
           | money shot, and that seems to have shut them up for now.
        
             | cyral wrote:
             | Apple Pay is legitimately useful though. You can use it to
             | pay at physical businesses if you forget your wallet
             | (double click and face ID to turn your phone into a "tap to
             | pay" card basically). There are also lots of apps/sites
             | that support it so you don't have to type in your card
             | number or even your shipping info sometimes.
        
             | wrboyce wrote:
             | How is that even relevant? Your phone was trying to onboard
             | you to a _free to use_ feature. If you can't see the
             | difference here, then I suspect there probably was a button
             | labelled "fuck off" and you didn't see that either.
             | Honestly.
        
               | joe_guy wrote:
               | Apple makes money off the interchange fee. It might be
               | "free" to the end user, but the corporate motivation is
               | the same as Prime's -- money.
               | 
               | And please don't ad hominem attack people you're
               | responding to.
        
               | wrboyce wrote:
               | Yeah you're right, there was no need for the last bit.
               | I'm still struggling to see the relevance though, trying
               | to get me to buy things is very different from trying to
               | get me to use a feature you profit from (in my opinion).
               | You also have to bare in mind that HN represents the more
               | technical users, plenty of people probably do need the
               | popups to discover these features. Saying that, a "no
               | thanks, don't remind me again" button would be a nice
               | inclusion - perhaps with a secondary confirmation.
        
               | tjoff wrote:
               | Feature you profit from?
               | 
               | You feel you profit from facebook tracking as well?
               | 
               | Regardless, these dark patterns are truly disgusting and
               | how some can defend them so mindlessly just because they
               | apparently found a use for a product is quite disturbing.
        
               | wrboyce wrote:
               | "You" in that context was the entity pushing the feature.
               | I'm not sure what point you think you're making about
               | Facebook tracking tbh, but I don't use Facebook so you're
               | asking the wrong person; to claim I am mindlessly
               | defending dark patterns is nonsense.
        
               | psyc wrote:
               | Windows regularly tries to nag me into free features I
               | don't want too.
               | 
               | At no time has the term 'dark pattern' ever been
               | necessarily dependent on getting you to pay money.
               | 
               | Your argument is that I sound stupid, so I must be wrong?
               | 
               | There's no button.
               | 
               | https://www.cultofmac.com/538999/apple-under-fire-apple-
               | pay-...
               | 
               | https://www.wsj.com/articles/apple-insists-iphone-users-
               | enro...
               | 
               | My other peeve is when streaming apps put a button in the
               | bottom-right of an ad, same size and style as the 'skip'
               | button one reflexively clicks. Except it turns out to be
               | an 'engage even moar' button.
        
               | wrboyce wrote:
               | Apologies for the implication you're stupid, I didn't
               | really mean that and it was uncalled for at any rate.
               | 
               | I don't disagree regards dark patterns, your example just
               | felt a bit irrelevant to the specific topic being
               | discussed (Amazon pushing a paid for product / cancelling
               | a paid subscription).
        
               | psyc wrote:
               | I can understand why you would make the distinction.
               | Making distinctions is good, in general. However from my
               | perspective as a frustrated user being antagonized by
               | 'my' devices, it's all the same battle to me.
        
               | KMag wrote:
               | "Free to use", but presumably comes with a user agreement
               | that opens you to some financial liability. There's a
               | (granted small) chance that a bug, security incident, or
               | fraud lands you in a Kafkaesque debt nightmare.
               | 
               | I had a bit of a nightmare where one of the credit
               | reporting agencies was convinced my residential address
               | was inside my bank. Their online system referred me to
               | their phone system or sending them mail. Their phone
               | system referred me to their online system or sending them
               | mail. I sent them mail 3 times and got no reply. An
               | online cheat guide for getting to an actual human through
               | their phone system didn't work, and I eventually just
               | started hitting random keys in their phone system and got
               | to a human who was able to sort it out.
               | 
               | You can't even get a secured credit card (backed by a
               | cash deposit) without a credit check (I looked into it),
               | which is going to fail if your residential address is
               | wrong.
               | 
               | Opening a financial account that might misreport
               | something to a credit agency shouldn't be taken lightly.
        
           | pessimizer wrote:
           | I feel like I remember it being pretty easy to cancel Prime,
           | though. Have things changed?
        
             | fmx wrote:
             | Yes: https://www.msn.com/en-us/autos/news/yes-it-seems-
             | amazon-did...
        
             | [deleted]
        
             | tjoff wrote:
             | Doesn't excuse tricking people into it.
        
           | bcrosby95 wrote:
           | If it's easy to accidentally sign up for something, does that
           | mean it has to be easy to accidentally cancel something?
           | Because that would be hilarious.
        
             | haswell wrote:
             | I'm imagining a scenario where you're about to check out,
             | and decide not to finish the transaction because you wanted
             | to add something else to your very first.
             | 
             | So you click the cancel button.
             | 
             | Only you find out you've cancelled Prime.
        
       | jumpifzero wrote:
       | In principle looks good but lots of potential for going wrong.
       | 
       | Just hope this doesn't backfire. The cookie law was also a thing
       | the EU created with good intentions after some politicians
       | decided "omg cookies are bad" and we ended up still using cookies
       | but pop-ups in every single website basically forcing you to
       | accept the use of cookies.
        
       | FollowingTheDao wrote:
       | "They have electrolytes."
        
       | LightG wrote:
       | Good. More.
        
       | TavsiE9s wrote:
       | That's going to be interesting to see if the whole credit scoring
       | industry will have to disclose their algorithms as well.
       | 
       | Looking at you, Schufa.
        
         | daniel-cussen wrote:
        
           | tester89 wrote:
           | I didn't think I would see the day with a troll on HN. I
           | doubt it's an AI too because of the coherency of what they're
           | saying.
        
             | daniel-cussen wrote:
             | Trolling whom? Trolling whom?
             | 
             | These guys send letters every month about artificial debt
             | they self-inflicted and demanding payment under threat of
             | theft, and I'm the troll?
             | 
             | International debt was under 1% before Ukraine, these dudes
             | were asking 20% a year, like the sixteenth power of what
             | the international market was asking. They can perfectly
             | well get credit at under 1%, they can perfectly well
             | function with lower rates.
             | 
             | They just love usury.
             | 
             | They love money.
        
       | [deleted]
        
       | MeteorMarc wrote:
       | It will be fun to see Google's algorithm for ranking search
       | results.
        
         | otherotherchris wrote:
         | "Here's 100PiB of unlabeled neural net weights. Knock
         | yourselves out."
        
           | jasfi wrote:
           | They want to know how the algorithms work, not the data
           | itself.
        
             | otherotherchris wrote:
             | Google doesn't know what the algorithm is anymore. The
             | whole site is a black box.
        
               | notacoward wrote:
               | Same at FB as far as I could tell while I was there. "The
               | algorithm" is a misnomer, popularized by the press but
               | really kind of silly. There are really thousands of
               | pipelines and models developed by different people
               | running on different slices of the data available. Some
               | are reasonably transparent. Others, based on training,
               | are utterly opaque to humans. Then the weights are all
               | combined to yield what users see. And it all changes
               | _every day_ if not every hour. Even if it could all be
               | explained in a useful way, that explanation would be out
               | of date as soon as it was received.
               | 
               | I'm not saying that to defend anyone BTW. This complexity
               | and opacity (which is transitive in the sense that a
               | combined result including even one opaque part itself
               | becomes opaque) is very much _the problem_. What I 'm
               | saying is that it's likely impossible for the companies
               | to comply without making fundamental changes ... which
               | might well be the intent, but if that's the case it
               | should be more explicit.
        
               | manquer wrote:
               | What needs to be shared is a high level arch not nuts and
               | bolts.
               | 
               | At a broad level:
               | 
               | what are the input sources like IP address , clicks on
               | other websites etc you use to feed the model.
               | 
               | What is the overall system optimized for , like some
               | combination of engagement , view time etc, just listing
               | them if possible in a order of preference is good enough
               | 
               | Alternatively what does your human management measure and
               | monitor as the business metrics of success .
               | 
               | I want to know what behaviors (not necessarily how ) are
               | used , I want to know what is feed trying to optimize for
               | , more engagement, more view time to etc
               | 
               | This is not adversarial, knowing this helps as modify
               | user behavior to make the model work better.
               | 
               | Users already have some sense of this and work around it
               | blindly , for example YouTube has heavy emphasis on
               | resent views and search . I (and am sure others) would
               | use signed out user to see content way outside my
               | interest area so my feed isn't polluted with poor
               | recommendations. I may have watched 1000's hours of
               | educational content but google would still think some how
               | to video I watched once means I need to only see that
               | kind of content.
               | 
               | Google knows it is me sure even am signed out, but they
               | don't use it change my feed that's the important part and
               | knowing that can help improve my user experience
        
               | jasfi wrote:
               | I doubt it, they should know what the various algorithms
               | are, especially the most important ones that drive most
               | of the ranking. But their competitive advantage would be
               | on the line.
        
               | [deleted]
        
               | ekianjo wrote:
               | > Google doesn't know what the algorithm is anymore
               | 
               | You are an insider?
        
               | [deleted]
        
               | tyingq wrote:
               | They haven't talked much detail since Matt Cutts left,
               | but over time they did sort of outline the basics. That
               | the core ranking is still some evolution of PageRank,
               | weighting scoring of page attributes/metadata and flowing
               | it down/through inbound links as well. But then altered
               | via various waves of ML, like Vince (authority/brand
               | power), Panda (inbound link quality), Penguin (content
               | quality), and many others that targeted other attributes
               | (page layout, ad placement, etc).
               | 
               | Even if some of that is off, the premise of a chain of
               | some ML, and some not ML, processors means they probably
               | can't really tell you exactly why anything ranks where it
               | does.
        
               | zo1 wrote:
               | Those sound like awesome potential features. Allow users
               | to assign 0-100% weights for each of those scoring
               | adjustments during search,and show them the calcs (if you
               | can).
        
               | tyingq wrote:
               | Supposedly there's thousands of different features that
               | are scored, and those are just the rolled-up categories
               | that needed their own separate ML pipeline step.
               | 
               | Like, maybe, for example, a feature is _" this site has a
               | favicon.ico that is unique and not used elsewhere"_ (page
               | quality). Or _" this page has ads, but they are below the
               | fold"_ (page layout). Or _" this site has > X amount of
               | inbound links from a hand curated list of 'legitimate
               | branded sites'"_ (page/site authority).
               | 
               | Google then picks a starting weight for all these things,
               | and has human reviewers score the quality of the results,
               | order of ranking, etc, based on a Google written how-to-
               | score document. Then tweaks the weights, re-runs the ML
               | pipeline, and has the humans score again, in some
               | iterative loop until they seem good.
               | 
               | There's a never-acted-on FTC report[1] that describes how
               | they used this system to rank their competition
               | (comparison shopping sites) lower in the search results.
               | 
               | [1] http://graphics.wsj.com/google-ftc-report/
               | 
               | Edit: Note that a lot of detail is missing here. Like
               | topic relevance, where a site may rank well for some
               | niche category it specializes in. But that it wouldn't
               | necessarily rank well for a completely different topic,
               | even with good content, since it has no established
               | signals it should.
        
               | dehrmann wrote:
               | > and those are just the rolled-up categories that needed
               | their own separate ML pipeline step.
               | 
               | AKA ensemble models.
        
               | dehrmann wrote:
               | It's clear the public and lawmakers like the idea of
               | knowing how the algorithm works, but what you posted is
               | about as deep as people can reasonably understand at a
               | high level. I don't think they realize how complex a
               | system built over 20 years that's a trillion-dollar
               | company's raison d'etre can be.
        
             | amelius wrote:
             | "Bad programmers worry about the code. Good programmers
             | worry about data structures and their relationships."
             | 
             | -- Linus Torvalds
        
               | ohgodplsno wrote:
               | Linus never had to deal with hundreds of petabytes of
               | search data, nor ML black boxes, to be fair.
        
             | wetpaws wrote:
             | Data is already an algorithm
        
           | thedeadfish wrote:
           | Google manually adjusts its results for censorship reasons.
           | This is probably why google has gotten so much worse, they
           | don't want information to be freely accessible, they only
           | want things they approve of to be seen.
        
             | otherotherchris wrote:
             | I reckon you're right, but I doubt that it's manual or
             | under Google's control. Google is too important a tool of
             | control to be left in the hands of Silly Valley idealists.
             | 
             | I've always wondered why Sergey Brin and Larry Page retired
             | when they did, it coincides almost exactly with the
             | beginning of the SERP quality decline. Wonder what sort of
             | conversation they had with intelligence to quietly walk to
             | the door, cash out, and say nothing about the company
             | since.
        
               | blihp wrote:
               | What happened was they got what they wanted: full control
               | of running the business. Then they quickly learned that
               | was actually a lot of work and not very much fun, made
               | some fairly unpopular decisions (business, product and
               | policy) with a fair amount of public backlash, put Sundar
               | in charge and backed away.
        
           | simion314 wrote:
           | >"Here's 100PiB of unlabeled neural net weights. Knock
           | yourselves out."
           | 
           | You need to give the user an explanation on why you blocked
           | his account, but if Google is kind enough to add on top the
           | secret neural network then some people would be happy to have
           | a look at it and find even more garbage in it.
        
             | NullPrefix wrote:
             | >Violation of guidelines and community standards.
        
               | ohgodplsno wrote:
               | Which is itself detected by a 80PiB neural network, based
               | on the 60TB output of new rules that another neural
               | network spits out every week based on the temperature
               | outside of the corner office and the taste of Sundar's
               | coffee this morning.
        
             | [deleted]
        
       | swayvil wrote:
       | I see a vast technical writing documentation project in their
       | future.
        
         | jdrc wrote:
         | But they must be short and easy to understand by users. Like
         | this:
         | 
         | "Our algorithms use gradient descent. Data flows through our
         | connected tubes, slowly wiggling their size until the data
         | starts flows back and forth faster."
        
       | ivrrimum wrote:
        
       | Proven wrote:
        
       | anothernewdude wrote:
       | "Linear Algebra"
        
       ___________________________________________________________________
       (page generated 2022-04-23 23:00 UTC)