• werefreeatlast@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      arrow-down
      6
      ·
      4 months ago

      Today, I started it like any other day… A good big boob gangbang video. Then to the serious stuff, the stuff AI is really here for!.. Hello Mr Chat GPT! Could you please write me a python program to predict the next lotto ticket numbers and winning locations?

      Oh and I do like that new EV. I see you would like me to purchase it since every page has it. But how could I tell AI that in the state of Washington we cannot have e-bikes?! Some idiot put a regulation that limits the power to 750w and the speed to 20mph. That makes going to the nearest store almost practical if I want to spend my day riding.

      AI could you please cause accidents that create the right environment for new politicians to increase that power limit to 3kw or something more practical? …eh okay Mr Chat, my cat just died on the window, could you please re-write the script so that my cat doesn’t die on the window and politicians decide to increase the power to 3kw?.. Ok Mr Chat, my cat now died in our microwave and the microwave is now 3kw. That’s very close to what I need but could you please not kill my cat? … Introducing the new KitchenAid 3kw horse discombobulation machine! Invented by the smartest persons in the entire planet! Bike? What bike?

      • xthexder@l.sw0.com
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        4 months ago

        I don’t really see the problem with restricting e-bike power. You can still go faster than 20mph if you pedal. I think what you really want is a motorcycle. They make those in electric form too.

  • BlackLaZoR@kbin.run
    link
    fedilink
    arrow-up
    53
    arrow-down
    1
    ·
    4 months ago

    This constant shuttling of information back and forth is responsible for consuming as much as 200 times the energy used in the computation, according to this research.

    Press x to doubt. I know moving data costs more energy than computation itself, but that sounds like a pure BS.

    • oakey66@lemmy.world
      link
      fedilink
      English
      arrow-up
      39
      arrow-down
      3
      ·
      4 months ago

      Everyone is trying to cash in before it all collapses because big tech has just turned into what I dub as hype hopping since they have no good ideas.

      • Sanctus@lemmy.world
        link
        fedilink
        English
        arrow-up
        23
        ·
        4 months ago

        Where the fuck are the automatic dishwashers that put the dishes away? Dryers that fold your clothes for you? We’ve got a long fucken way to go and we haven’t even half pillaged the Jetsons. Let’s get on with the progress already! Fuck chasing profit!

          • TomSelleck@lemm.ee
            link
            fedilink
            English
            arrow-up
            3
            ·
            4 months ago

            Let me know when A.I. can make me a decent sandwich, and then I’ll start care.

          • Aniki 🌱🌿@lemmy.zip
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            4 months ago

            I’ve been toying with the idea that giving rocks logic or moving FTL is just not something we’re going to see cracked in our lifetimes. Certainly not while its capitalism or bust.

            We’ll need breakthroughs that involve the whole of humanity.

  • A_A@lemmy.world
    link
    fedilink
    English
    arrow-up
    39
    ·
    edit-2
    4 months ago

    Experimental demonstration of magnetic tunnel junction-based computational random-access memory
    “In this work, a CRAM array based on magnetic tunnel junctions (MTJs) is experimentally demonstrated. First, basic memory operations, as well as 2-, 3-, and 5-input logic operations, are studied. Then, a 1-bit full adder with two different designs is demonstrated.”
    https://www.nature.com/articles/s44335-024-00003-3
    So, this is experimentally demonstrated, yet, only at small scale.

  • FaceDeer@fedia.io
    link
    fedilink
    arrow-up
    19
    arrow-down
    1
    ·
    4 months ago

    It probably doesn’t matter from a popular perception standpoint. The talking point that AI burns massive amounts of coal for each deepfake generated is now deeply ingrained, it’ll be brought up regularly for years after it’s no longer true.

    • palordrolap@kbin.run
      link
      fedilink
      arrow-up
      12
      ·
      edit-2
      4 months ago

      To stick with the analogy, this is like putting a small CPU inside the bottle, so the main CPU<->RAM bottleneck isn’t used as often. That said, any CPU, within RAM silicon or not, is still going to have to shift data around, so there will still be choke points, they’ll just be quicker. Theoretically.

      Thinking about it, this is kind of the counterpart to CPUs having an on-chip cache of memory.

      Edit: counterpoint to counterpart

  • HubertManne@moist.catsweat.com
    link
    fedilink
    arrow-up
    2
    ·
    4 months ago

    I hope this is true. ai has its uses but it can’t be way more inneficient. It would be great if it answering used no more energy than a standard web query

  • Ilovethebomb@lemm.ee
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    2
    ·
    4 months ago

    Arm’s CEO recently suggested that by 2030, AI may consume a quarter of all energy produced in the U.S.

    No way does AI produce enough value that they could afford this.