• Lettuce eat lettuce
    link
    fedilink
    English
    arrow-up
    86
    arrow-down
    1
    ·
    4 months ago

    Capitalists: So you’re telling me I can build 1000x more AI data center infrastructure now?

    • werefreeatlast@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      3
      ·
      4 months ago

      Today, I started it like any other day… A good big boob gangbang video. Then to the serious stuff, the stuff AI is really here for!.. Hello Mr Chat GPT! Could you please write me a python program to predict the next lotto ticket numbers and winning locations?

      Oh and I do like that new EV. I see you would like me to purchase it since every page has it. But how could I tell AI that in the state of Washington we cannot have e-bikes?! Some idiot put a regulation that limits the power to 750w and the speed to 20mph. That makes going to the nearest store almost practical if I want to spend my day riding.

      AI could you please cause accidents that create the right environment for new politicians to increase that power limit to 3kw or something more practical? …eh okay Mr Chat, my cat just died on the window, could you please re-write the script so that my cat doesn’t die on the window and politicians decide to increase the power to 3kw?.. Ok Mr Chat, my cat now died in our microwave and the microwave is now 3kw. That’s very close to what I need but could you please not kill my cat? … Introducing the new KitchenAid 3kw horse discombobulation machine! Invented by the smartest persons in the entire planet! Bike? What bike?

      • xthexder@l.sw0.com
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        4 months ago

        I don’t really see the problem with restricting e-bike power. You can still go faster than 20mph if you pedal. I think what you really want is a motorcycle. They make those in electric form too.

  • A_A@lemmy.world
    link
    fedilink
    English
    arrow-up
    35
    ·
    edit-2
    4 months ago

    Experimental demonstration of magnetic tunnel junction-based computational random-access memory
    “In this work, a CRAM array based on magnetic tunnel junctions (MTJs) is experimentally demonstrated. First, basic memory operations, as well as 2-, 3-, and 5-input logic operations, are studied. Then, a 1-bit full adder with two different designs is demonstrated.”
    https://www.nature.com/articles/s44335-024-00003-3
    So, this is experimentally demonstrated, yet, only at small scale.

  • FaceDeer@fedia.io
    link
    fedilink
    arrow-up
    19
    ·
    4 months ago

    It probably doesn’t matter from a popular perception standpoint. The talking point that AI burns massive amounts of coal for each deepfake generated is now deeply ingrained, it’ll be brought up regularly for years after it’s no longer true.

    • palordrolap@kbin.run
      link
      fedilink
      arrow-up
      11
      ·
      edit-2
      4 months ago

      To stick with the analogy, this is like putting a small CPU inside the bottle, so the main CPU<->RAM bottleneck isn’t used as often. That said, any CPU, within RAM silicon or not, is still going to have to shift data around, so there will still be choke points, they’ll just be quicker. Theoretically.

      Thinking about it, this is kind of the counterpart to CPUs having an on-chip cache of memory.

      Edit: counterpoint to counterpart

  • HubertManne@moist.catsweat.com
    link
    fedilink
    arrow-up
    2
    ·
    4 months ago

    I hope this is true. ai has its uses but it can’t be way more inneficient. It would be great if it answering used no more energy than a standard web query

  • Ilovethebomb@lemm.ee
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    2
    ·
    4 months ago

    Arm’s CEO recently suggested that by 2030, AI may consume a quarter of all energy produced in the U.S.

    No way does AI produce enough value that they could afford this.