• deranger@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    7
    ·
    1 year ago

    This only got bad with the most recent generation of CPUs. AMD 5xxx series is very efficient as demonstrated by Gamers Nexus. The Intel CPUs from 2500k to idk, 8xxx series? were efficient until they started slapping more cores and then cranking the power on them.

    • candyman337@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      Yes the second thing about cranking power and cores is what I’m talking about.

      Also, as far as gpus, the 2000 series was ridiculously power hungry at the time, and it looks downright reasonable now. It’s like the Overton window of power consumption lol.

      Image

      • deranger@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        1 year ago

        I dunno, I ran a 2080 on the same PSU that I used on a 2013 build, a 650W seasonic. Got some graphs? Power consumption didn’t seem to jump that bad until the latest gen.

        My current 3090 is a power hog though, that’s when I’d say it started for Nvidia (3000 series). For AMD, 7000 series CPUs, and I’m not really sure for Intel. 9900k was the last Intel CPU I ran, it seemed fine. I was running a 9900k/2080 on the same PSU as the 2500k/570 build.

        • candyman337@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          1 year ago

          As for as the 2080 goes, like I said, it was big FOR THE TIME, and power hungry FOR THE TIME. It’s still reasonable especially for today’s standards

          As for as the last two gens, 3000 and 4000 series, they are known to draw more than their rated power requirements, which, for their min recommended psu wattage, 3080 was 50 watts more than the 2080 (750w), and 4080 was 100 w more than that (850w)

          To add to that, both of these gens of cards, when doing graphics intensive things like gaming, can overdraw power and have been known to cause hard shutdowns in pcs with PSUs that are even slightly higher rated than their min rec. Before these last two gens you could get away with a slightly lower than rated wattage PSU and sacrifice a little performance but that is definitely no longer the case.

          And sure, the performance to watts used is better in the 3080, but they also run 10+ degrees hotter and the 4000 series even moreso.

          I just hope the 5000 series goes the way of power consumption refinement rather than smashing more chips onto a board or vram fuckery like with the 4060, like I’d be happy with similar performance on the 5000 series if it was less power hungry

    • Fermion@feddit.nl
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      The 7 series are more efficient than the 5 series. They just are programmed to go as fast as thermals allow. So the reviewers that had really powerful coolers on the cpus saw really high power draw. If instead you set a power cap, you get higher performance per watt than the previous generations.

      Having the clocks scale to a thermal limit is a nice feature to have, but I don’t think it should have been the default mode.

    • Dudewitbow
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Intel became less efficient because of how long they were stuck on 14nm. In order to compensate to beat amd in performance mindshare, they needed to push the clocks hard.

      Overtime, cpus have been sitting closer to max clock, defeating the purpose of overclocking to many, where adding 1GHz was not out of the ordinary. Now getting 0.5GHz is an acheivement.