Despite their sensitive content, mental health and prayer apps are “worse than any other product” when it comes to user privacy and security, an analysis by browser firm Mozilla revealed on Monday.
“The vast majority of mental health and prayer apps are exceptionally creepy,” Mozilla’s Jen Caltrider, the primary creator of the firm’s “Privacy not Included” guide, which evaluated 32 such apps on their respect for users’ personal data, told the Verge on Monday.
Caltrider noted that the apps one might think would recognize the sensitive nature of their data instead “track, share, and capitalize on users’ most intimate personal thoughts and feelings, like moods, mental state, and biometric data.”
Privacy is assumed in such apps when compared with the real-life relationship between a therapist and patient, or a believer and religious institution, one might hope that app developers would at least attempt to replicate such safe spaces when taking the whole process online.
For instance, one popular app, Talkspace, stashes entire chat transcripts between user and therapist. An AI therapy chatbot called Woebot, collects information about users from third parties, then shares it for advertising purposes.
Mozillla researcher Misha Rykov referred to the apps his team analyzed as “data-drawing machines with a mental health app veneer,” or “a wolf in sheep’s clothing.” With real-life therapy increasingly expensive and the process of finding a therapist who 'matches' a patient hit or miss, the lure of a friendly voice just a click away is difficult to resist for some. But given the apps’ apparent core purpose of data mining and selling users’ deepest darkest secrets, it might be a better to hold one’s self until one can meet up with a qualified therapist or at least a trusted friend – ideally in real life.
-RT