Actually, really liked the Apple Intelligence announcement. It must be a very exciting time at Apple as they layer AI on top of the entire OS. A few of the major themes.

Step 1 Multimodal I/O. Enable text/audio/image/video capability, both read and write. These are the native human APIs, so to speak.

Step 2 Agentic. Allow all parts of the OS and apps to inter-operate via “function calling”; kernel process LLM that can schedule and coordinate work across them given user queries.

Step 3 Frictionless. Fully integrate these features in a highly frictionless, fast, “always on”, and contextual way. No going around copy pasting information, prompt engineering, or etc. Adapt the UI accordingly.

Step 4 Initiative. Don’t perform a task given a prompt, anticipate the prompt, suggest, initiate.

Step 5 Delegation hierarchy. Move as much intelligence as you can on device (Apple Silicon very helpful and well-suited), but allow optional dispatch of work to cloud.

Step 6 Modularity. Allow the OS to access and support an entire and growing ecosystem of LLMs (e.g. ChatGPT announcement).

Step 7 Privacy. <3

We’re quickly heading into a world where you can open up your phone and just say stuff. It talks back and it knows you. And it just works. Super exciting and as a user, quite looking forward to it.

https://x.com/karpathy/status/1800242310116262150?s=46

  • c10l@lemmy.world
    link
    fedilink
    English
    arrow-up
    91
    arrow-down
    2
    ·
    22 days ago

    Founding member of company that stands to make fortunes through a product endorses said product.

    • Z4rK@lemmy.world
      cake
      OP
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      29
      ·
      22 days ago

      I mean, that’s fair, if you don’t believe in his integrity than this news have very little value to you.

    • reattach@lemmy.world
      link
      fedilink
      English
      arrow-up
      27
      arrow-down
      2
      ·
      22 days ago

      I thought the original post was satire - list all of the privacy issues, then throw in “Privacy <3” at the end. Seriously, almost every one of those points has a potential privacy issue.

      Guess I was being too generous.

    • Z4rK@lemmy.world
      cake
      OP
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      41
      ·
      22 days ago

      How so? Many people want to use AI in privacy, but it’s too hard for most people to set it up for themselves currently.

      Having AI tools on the OS level so you can use it in almost any app and that is guaranteed to be processed on device in privacy will be very useful if done right.

      • TheFriar@lemm.ee
        link
        fedilink
        English
        arrow-up
        41
        arrow-down
        7
        ·
        22 days ago

        You think your iPhone isn’t collecting data on you? Is that what you’re saying?

        • ji17br
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          15
          ·
          22 days ago

          The phone is, Apple isn’t. They outline everything in the keynote if you are interested.

        • Z4rK@lemmy.world
          cake
          OP
          link
          fedilink
          English
          arrow-up
          12
          arrow-down
          28
          ·
          22 days ago

          Unless you are designing and creating your own chips for processing, networking etc, then privacy today is about trust, not technology. There’s no escaping it. I know iPhone and Apple is collecting data about me. I currently trust them the most on how they use it.

          • MigratingtoLemmy@lemmy.world
            link
            fedilink
            English
            arrow-up
            21
            arrow-down
            3
            ·
            22 days ago

            Running FOSS and taking control of your network will do a far better trick of privacy vs convenience than most people can imagine

          • EngineerGaming@feddit.nl
            link
            fedilink
            English
            arrow-up
            11
            arrow-down
            1
            ·
            edit-2
            22 days ago

            There are degrees of trust though. You can trust the developers and people who audited the code if you have no skill/desire to audit it yourself, or you can trust just the developers.

            And even closed systems’ behavior can be monitored and analyzed.

            • Z4rK@lemmy.world
              cake
              OP
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              12
              ·
              22 days ago

              Yes definitely, Apple claimed that their privacy could be independently audited and verified; we will have to wait and see what’s actually behind that info.

      • Zoot@reddthat.com
        link
        fedilink
        English
        arrow-up
        19
        arrow-down
        2
        ·
        22 days ago

        Yeah just like Microsoft Recall right? An AI that has access to every single thing you do (and would also be recording, otherwise how does it know “you”) can never be private by design. Its literal design is to know everything about you, your actions, and your habits. I wouldn’t trust anyone to be able to create an actually secure piece of software that does the above. It will always be able to be stolen/sold/abused.

        • Z4rK@lemmy.world
          cake
          OP
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          13
          ·
          21 days ago

          macOS and Windows could already be doing this today behind your back regardless of any new AI technology. Don’t use an OS you don’t trust.

            • Z4rK@lemmy.world
              cake
              OP
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              12
              ·
              21 days ago

              That’s fair, but you are misunderstanding the technology if you’re bashing the AI from Apple for making macOS less secure. Most likely, it will be just as secure as for example their password functionality, although we don’t have details yet. You either trust the OS or not.

              Microsoft Recall was designed so badly, there’s no hope for it.

              • Zoot@reddthat.com
                link
                fedilink
                English
                arrow-up
                16
                arrow-down
                2
                ·
                21 days ago

                I simply don’t, and wouldn’t trust Apple. They will tell you they are all about privacy, and happily sell your data behind your back. Just like any other company.

  • LANIK2000@lemmy.world
    link
    fedilink
    English
    arrow-up
    54
    arrow-down
    3
    ·
    22 days ago

    The amount of corporate speak makes me sick. Especially the mix of buzzwords being mixed with shit like “KERNEL PROCESS”, shit’s cursed.

  • henfredemars@infosec.pub
    link
    fedilink
    English
    arrow-up
    53
    arrow-down
    3
    ·
    edit-2
    22 days ago

    Kernel process LLM

    God I hope not. That sounds extremely insecure. Definitely do not do this in the kernel.

    • Z4rK@lemmy.world
      cake
      OP
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      14
      ·
      22 days ago

      He sort of invented it, so you have to think he’s commenting on the concept here, not the implementation.

      I have tried a lot of medium and small models, and there it just no good replacement for the larger ones for natural text output. And they won’t run on device.

      Still, fine-tuning smaller models can do wonders, so my guess would be that Apple Intelligence is really 20+ small and fine tuned models that kick in based on which action you take.

      • gravitas_deficiency@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        3
        ·
        21 days ago

        An LLM has no comprehension of what it says. It’s just a puppy that is really good at performing for treats. This will always yield nonsense a meaningful proportion of the time.

        I don’t care how statistically good your model can be under certain constraints and inputs. At the end of the day, all you’ve done is classically condition your computer.

        • Z4rK@lemmy.world
          cake
          OP
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          13
          ·
          21 days ago

          It goes a tad bit beyond classical conditioning… LLM’a provides a much better semantic experience than any previous technology, and is great for relating input to meaningful content. Think of it as an improved search engine that gives you more relevant info / actions / tool-suggestions etc based on where and how you are using it.

          Here’s a great article that gives some insight into the knowledge features embedded into a larger model: https://transformer-circuits.pub/2024/scaling-monosemanticity/

          • gravitas_deficiency@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            edit-2
            21 days ago

            That’s great. But that’s not how it’s being marketed and sold to the public. It’s being sold as an oracle (as in crystal ball, not database). And it’s misleading and hurting people as a result.

            I’ll reiterate: An LLM has no comprehension of what it says.

            It’s a matter of engineering ethics, on multiple levels:

            • the training data in the vast majority of cases is outright stolen
            • it’s being sold as something that it’s not, and the result is causing real damage to people and society in a ton of ways we’re still discovering
            • most people deeply involved in developing LLMs, and basically all of the technical leadership, are categorically ignoring and abrogating any and all responsibility around this “magical” new system they’ve made. We’ve seen this before with social networking. We know where this road leads.

            I’m not saying the tech should be banned. That’s obviously idiotic. Neural nets can - and are - used for tons of fascinating and excellent applications. It’s just that my staunch opinion is that LLMs are a terrible application of that the tech at this stage of development, and it’s particularly terrible that OpenAI/Microsoft/etc are aggressively foisting this technology on the public, and simultaneously refusing to take any ethical responsibility for it.

            • Z4rK@lemmy.world
              cake
              OP
              link
              fedilink
              English
              arrow-up
              2
              ·
              21 days ago

              To be honest, I’m not sure what we’re arguing - we both seem to have a sound understanding of what LLM is and what it is not.

              I’m not trying to defend or market LLM, I’m just describing the usability of the current capabilities of typical LLMs.

              • gravitas_deficiency@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                1
                ·
                21 days ago

                I’m saying that I wish that more people involved with the core development of the technology took the ethical considerations seriously, and communicated those concerns as a first-order issue when they talk about applications like this.

                It’s fascinating tech, but the way it’s being employed these days is deeply irresponsible.

    • Z4rK@lemmy.world
      cake
      OP
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      13
      ·
      22 days ago

      Care to elaborate?

      The suspicious parts to me was that they didn’t show much of the private cloud stuff, how much it would cost, and that they still feel the need to promote ChatGPT .

      • Jo Miran
        link
        fedilink
        English
        arrow-up
        27
        ·
        22 days ago

        All of it sounds like marketing and I have serious doubt’s about their commitment to, or ability to respect privacy when one of their previous points is that they plan to integrate third party systems. So…I have doubts.

        • Z4rK@lemmy.world
          cake
          OP
          link
          fedilink
          English
          arrow-up
          11
          arrow-down
          5
          ·
          22 days ago

          I mean, that’s fair, I personally use Apple devices specifically because I trust them the most on privacy, but if you don’t trust Apple with privacy, which is a 100% valid take to have, then of course this mayor selling point of their marketing becomes moot.

          • umami_wasabi
            link
            fedilink
            English
            arrow-up
            14
            arrow-down
            1
            ·
            22 days ago

            I would not give the right of anyone deciding what is good for my privacy, including Apple. This should be a judgement made by myself.

            • Z4rK@lemmy.world
              cake
              OP
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              7
              ·
              22 days ago

              I do agree, but privacy in 2024 is sadly about trust, not technology, unless you yourself can design and create every chip used in your devices and in the network cells you connect to. No setting on your device on “do not allow…” have any meaning without trust in the creator.

              • umami_wasabi
                link
                fedilink
                English
                arrow-up
                4
                arrow-down
                1
                ·
                edit-2
                21 days ago

                I didn’t say trust no one, but whom and what I trust shall be decided by me. Yes, there are things we can’t just build in our garage, yet there are tools enables us to investigate, and people and organization working on it. Maybe Apple’s take on AI have better privacy then others, but that shall be investigated and proven upon after release, not automatically granted.

  • restingboredface@sh.itjust.works
    cake
    link
    fedilink
    English
    arrow-up
    15
    arrow-down
    3
    ·
    22 days ago

    I look forward to Apple Marketing coming up with their usual line of nonsense, like a meaningless name for an existing capability that they are claiming to have invented.

    • helenslunch@feddit.nl
      link
      fedilink
      English
      arrow-up
      7
      ·
      22 days ago

      I watched an abbreviated video. Pretty much everything they announced was available on other platforms 5+years ago

        • helenslunch@feddit.nl
          link
          fedilink
          English
          arrow-up
          1
          ·
          21 days ago

          I mean don’t get me wrong, these things are a huge QoL improvement. It’s just that they should be embarrassed to be lacking such basic functionality on such insanely expensive devices. And people should be embarrassed to carry them around. There are still tons of basic features missing.

    • Z4rK@lemmy.world
      cake
      OP
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      22 days ago

      Well they just name-grabbed all of AI with their stupid Apple Intelligence branding.

  • demonsword@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    arrow-down
    2
    ·
    edit-2
    21 days ago

    Andrej Karpathy endorses Apple Intelligence

    Who is this guy and why his opinion should mean anything to me?

    EDIT: nevermind, searched for it and its some guy who used to work at OpenAI.