[HN Gopher] Dramatic growth in mental-health apps has created a ... ___________________________________________________________________ Dramatic growth in mental-health apps has created a risky industry Author : axiomdata316 Score : 24 points Date : 2021-12-20 22:05 UTC (54 minutes ago) (HTM) web link (www.economist.com) (TXT) w3m dump (www.economist.com) | rbartelme wrote: | https://archive.md/pog41 | Pigalowda wrote: | From the article: | | "No universal standards for storing "emotional data" exist. John | Torous of Harvard Medical School, who has reviewed 650 mental- | health apps, describes their privacy policies as abysmal. Some | share information with advertisers. "When I first joined | BetterHelp, I started to see targeted ads with words that I had | used on the app to describe my personal experiences," reports one | user" | | >I could drum up a dark fantastical setting of HR departments | buying this data and screening candidates with it. Hopefully this | patient data gets HIPAA treatment and is more protected. | ImaCake wrote: | The problem is that there is no requirement for these apps to | actually _work_. They might work, or might not. But that doesn't | really impact whether scientists get grants. They make the | funders /government look like they care about mental health and | these apps are _cheap_ compared to more traditional mental health | and psychiatric therapies. The scientists are desperate for the | next round of grants funding and are happy to produce something | that makes the government funding them look good. | | This kind of thing is pretty common in science. It can be hard to | see from the outside, but it is so obvious once you have been | involved in grant writing. | eganist wrote: | disclosure: I hold an equity position in an end to end encrypted, | trust on first use communications app, not really relevant to | this conversation so no value in naming them. | | I'm not terribly surprised at the number of services that have | popped up in the last few years, but considering the sensitivity | of the data (in many cases, info that most would consider to be | some of our deepest feelings on a topic that we wouldn't want | just anyone to hear about), I'm surprised at the lack of ventures | working to distance themselves from the content of the | conversations. It just feels like in this specific niche, | protection of data would be a much bigger selling point than most | others. | | I'd imagine any suitably well-funded chat service with end to end | encryption could fund a subsidiary venture to build out a | therapy-focused implementation of such, and especially in the | world of therapy, hand-written notes would probably be preferred | anyway, so there shouldn't be much in the way of storage of data | aside from directory data for individual users of the service and | probably card transaction pointers to a third party processor, | plus a handful of other bits that still shouldn't be anywhere | near as compromising as actual _contents of conversation._ | | Open to discussion on the topic. There's probably an angle I'm | not seeing. | rbartelme wrote: | I think it might depend on whether or not U.S. insurance | companies are involved. If so HIPAA is a real liability as a | company. For example, a former colleague worked at a high | profile private research facility that had their HIPAA firewall | breached and their entire department was fired. Assuming that | chat logs on these apps could fall into that same class of data | in the U.S., that's a huge amount of risk/liability as a | company. ___________________________________________________________________ (page generated 2021-12-20 23:00 UTC)