800
Apple opts everyone into having their Photos analyzed by AI
(www.theregister.com)
Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.
In this community everyone is welcome to post links and discuss topics related to privacy.
much thanks to @gary_host_laptop for the logo design :)
Did they? Because it seems like everyone else is in a hype bubble and doesn't give a shit about how much this costs or how much money it makes.
Looks like they did "Brakerski-Fan-Vercauteren (BFV) HE scheme, which supports homomorphic operations that are well suited for computation (such as dot products or cosine similarity) on embedding vectors that are common to ML workflows" namely they use a scheme that is both secure and efficient specifically for the kind of compute they do here. https://machinelearning.apple.com/research/homomorphic-encryption
At least it's not going to be the overhyped LLM doing the analysis, it seems, considering the input is a photo data.
Here indeed I don't think so but other vision models, e.g. https://github.com/vikhyat/moondream are relying on LLM to generate the resulting description.
My gosh, what is with people's reliance on single thing
Well to be fair, and even though I did spend a bit of time to write about the broader AI hype BS cycle https://fabien.benetou.fr/Analysis/AgainstPoorArtificialIntelligencePractices LLMs are in itself not "bad". It's an interesting idea to rely on our ability to produce and use languages to describe a lot of useful things around us. So using statistics on it to try to match is actually pretty smart. Now... there are so many things that went badly for the last few years I won't even start (cf link) but the concept per se, makes sense to rely on it sometimes.