• Grimy@lemmy.world
    link
    fedilink
    English
    arrow-up
    53
    ·
    5 months ago

    The oil industry must be so giddy to have found a new scape goat out of nowhere.

    Datacenters take a lot of energy because they serve a lot of people. The impact can be lessened with a proper grid centered around renewable.

    There are actual things that are fucking up the planet, individuals using AI, gaming or having a Google account aren’t the actual issue.

    • silence7@slrpnk.netOP
      link
      fedilink
      English
      arrow-up
      17
      ·
      5 months ago

      Data center demand has created huge backlog of gas turbine orders. They’re not planning on renewables for the next big expansion

    • Feyd@programming.dev
      link
      fedilink
      English
      arrow-up
      12
      ·
      5 months ago

      You are correct that renewable energy would help but if huge amounts of power are specifically being drawn for AI data centers that is part of the equation. Just like it’s reduce/reuse/recycle in that order for handling items, it should be reduce/renewable for power, and we should have to build the renewable infrastructure before building more data centers.

  • Optional@lemmy.world
    link
    fedilink
    English
    arrow-up
    49
    ·
    5 months ago

    An unholy amount.

    An amount guaranteed to spike climate targets a decade early.

    Stoopid much.

  • Xaphanos@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    5 months ago

    A problem is that the information is not in the hands of the company selling the AI. The actual hardware is often owned by service providers and independent data centers.

    • silence7@slrpnk.netOP
      link
      fedilink
      English
      arrow-up
      9
      ·
      5 months ago

      They know exactly what the power consumption of that hardware is though. This isnt tough to figure out just because you use a cloud provider

      • Xaphanos@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        5 months ago

        Well, I work at an AI hyperscaler. I can tell you how much my facility uses, and how much each rack uses, but don’t have any way to determine what the customer is doing on that server. Or even which servers a given customer is using. Is it being used heavily for queries? How many? Of what kind? We don’t know. Only what the rack/row/pod/hall is consuming.

        Also, does the network gear overhead count? How do you apportion that?

        We have no visibility into the customer workload. Some of our customers use our systems for scientific research. Drugs, etc. How do you tally that?

        I’m not saying that it is impossible, just that if the customer won’t pay for that report, we’re not going to spend money to build the systems to produce it.

        Do I agree? No. But I’m just a grunt.

        • Zeoic@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          5 months ago

          Im sure they can do the simple math of: we pay for x power, we have y customers. x / y would be a rough but probably pretty accurate number if we are talking tens of thousands to millions of customers.

        • silence7@slrpnk.netOP
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 months ago

          You can produce a remarkably good estimate by looking at CPU and GPU utilization out of procfs and profiling a handful of similar machines power use with similar utilization and workloads.

          Network is less than 5% of power use for non-GPU loads; probably less for GPU.

          • ddh@lemmy.sdf.org
            link
            fedilink
            English
            arrow-up
            2
            ·
            5 months ago

            Sure, you can do that at an aggregate level, but then how do you divide it by customer? And even then, some setups will be more efficient than others, so you’d only get that setup’s usage.

            And even if you do that and can narrow it down to a single user and a single prompt, you can still only roughly predict how long it will think and how long the response will be.

            • silence7@slrpnk.netOP
              link
              fedilink
              English
              arrow-up
              1
              ·
              5 months ago

              By customer is easy: they’re each renting specific resources. A fractional cloud instance (excepting the sma burst able ones) is tied to specific CPUs and GPUs. And there are records of who rented which one when being kept already.

              You might not be able to break out specific individual queries, but computing averages is completely straightforward

  • ieatpwns@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    5 months ago

    So much that now they want to turn nuclear reactors back on. Not because it’s green energy but because it’s free energy for them