• 0 Posts
  • 38 Comments
Joined 3 months ago
cake
Cake day: March 3rd, 2024

help-circle


  • yeah i see that too. it seems like mostly a reactionary viewpoint. the reaction is understandable to a point since a lot of the “AI” features are half baked and forced on the user. to that point i don’t think GNOME etc should be scrambling to add copies of these features.

    what i would love to see is more engagement around additional pieces of software that are supplemental. for example, i would love if i could install a daemon that indexes my notes and allows me to do semantic search. or something similar with my images.

    the problems with AI features aren’t within the tech itself but in the surrounding politics. it’s become commonplace for “responsible” AI companies like OpenAI to not even produce papers around their tech (product announcement blogs that are vaguely scientific don’t count), much less source code, weights, and details on training data. and even when Meta releases their weights, they don’t specify their datasets. the rat race to see who can make a decent product with this amazing tech has made the whole industry a bunch of pearl clutching FOMO based tweakers. that sparks a comparison to blockchain, which is fair from the perspective of someone who hasn’t studied the tech or simply hasn’t seen a product that is relevant to them. but even those people will look at something fantastical like ChatGPT as if it’s pedestrian or unimpressive because when i asked it to write an implementation of the HTTP spec in the style of Fetty Wap it didn’t run perfectly the first time.


  • a lot of things are unknown.

    i’d be very surprised if it doesn’t have an opt out.

    a point i was trying to make is that a lot of this info already exists on their servers, and your trust in the privacy of that is what it is. if you don’t trust them that it’s run on per user virtualized compute, that it’s e2e encrypted, or that they’re using local models i don’t know what to tell you. the model isn’t hoovering up your messages and sending them back to Apple unencrypted. it doesn’t need to for these features.

    all that said, this is just what they’ve told us, and there aren’t many people who know exactly what the implementation details are.

    the privacy issue with Recall, as i said, is that it collects a ton of data passively, without explicit consent. if i open my KeePass database on a Recall enabled machine, i have little assurance that this bot doesn’t know my Gmail password. this bot uses existing data, in controlled systems. that’s the difference. sure maybe people see Apple as more trustworthy, but maybe sociology has something to do with your reaction to it as well.



  • people generally probably hate the iOS integration just because it’s another AI product, but they’re fundamentally different. the problem with Recall isn’t the AI, it’s the trove of extra data that gets collected that you normally wouldn’t save to disk whereas the iOS features are only accessing existing data that you give it access to.

    from my perspective this is a pretty good use case for “AI” and about as good as you can do privacy wise, if their claims pan out. most features use existing data that is user controlled and local models, and it’s pretty explicit about when it’s reaching out to the cloud.

    this data is already accessible by services on your phone or exists in iCloud. if you don’t trust that infrastructure already then of course you don’t want this feature. you know how you can search for pictures of people in Photos? that’s the terrifying cLoUD Ai looking through your pictures and classifying them. this feature actually moves a lot of that semantic search on device, which is inherently more private.

    of course it does make access to that data easier, so if someone could unlock your device they could potentially get access to sensitive data with simple prompts like “nudes plz”, but you should have layers of security on more sensitive stuff like bank or social accounts that would keep Siri from reading it. likely Siri won’t be able to get access to app data unless it’s specified via their API.


  • no need for Python. there’s a Google SDK, ML Kit, that will do the heavy lifting on this. if that’s not acceptable, TensorFlow, PyTorch, and ONNX support Android, albeit not as nicely integrated.

    your image processing pipeline will be imageSource -> RGB encoding -> OCR -> profit. your OCR just needs an RGB encoded image. doesn’t matter if that’s a JPEG or YUV video feed at the source.

    as for if there’s an app that fits OP’s exact use case, dunno.




  • used to be the Android team used Ubuntu, not sure if that’s still the case. Linux is pretty much the native environment for Android dev. i’d recommend at least 4GB of dedicated RAM if not 8. definitely at least 8 if you plan to use the emulator (which is itself a VM).

    Android Studio will get you 90% of the way there. it will help you install the SDK, emulators, etc, and provide UI front ends for the CLI tools, ie adb.

    there’s really not much to system level dependencies. if your distribution supports JDK 17 (probable) you’ll be fine with whatever.

    obligatory: i use Arch, btw


  • tbh this research has been ongoing for a while. this guy has been working on this problem for years in his homelab. it’s also known that this could be a step toward better efficiency.

    this definitely doesn’t spell the end of digital electronics. at the end of the day, we’re still going to want light switches, and it’s not practical to have a butter spreading robot that can experience an existential crisis. neural networks, both organic and artificial, perform more or less the same function: given some input, predict an output and attempt to learn from that outcome. the neat part is when you pile on a trillion of them, you get a being that can adapt to scenarios it’s not familiar with efficiently.

    you’ll notice they’re not advertising any experimental results with regard to prediction benchmarks. that’s because 1) this actually isn’t large scale enough to compete with state of the art ANNs, 2) the relatively low resolution (16 bit) means inputs and outputs will be simple, and 3) this is more of a SaaS product than an introduction to organic computing as a concept.

    it looks like a neat API if you want to start messing with these concepts without having to build a lab.



  • as big as the circle jerk is here against AI, i think it’s on the whole a good thing if we use it for what it’s actually good at: approximating an answer. but once companies start promising things like security that require 100% accuracy they totally lose me. as someone who has worked on recognition systems i will be opting out so fast to things like facial scan at PoS. it’s not AI because it’s not actually intelligent. you can’t reason with it or change its mind without rigorous training. write some shitty code for me to fix? fine. buy a TV with whatever bs contractor bid the lowest for the facial scanning job? gtfo. startup founders, executives, and managers will promise the moon when they’re so far up their own ass they’ve never even seen it.






  • IBM then. or, i don’t know, the British Royal Family?

    the reality of talking about extremist economics is no one knows how it would work out in the long term. but regardless, if it happened tomorrow we already have a Microsoft to deal with.

    “taxation is theft” “wage labour is exploitation”

    sometimes things are subtle and complicated and can’t be practically boiled down to absolutes.


  • “we don’t know how” != “it’s not possible”

    i think OpenAI more than anyone knows the challenges with scaling data and training. anyone working on AI knows the line: “a baby can learn to recognize elephants from a single instance”. reducing training data and time is fundamental to advancement. don’t get me wrong, it’s great to put numbers to these things. i just don’t think this paper is super groundbreaking or profound. a bit clickbaity and sensational for Computerphile