The University of Rhode Island’s AI lab estimates that GPT-5 averages just over 18 Wh per query, so putting all of ChatGPT’s reported 2.5 billion requests a day through the model could see energy usage as high as 45 GWh.

A daily energy use of 45 GWh is enormous. A typical modern nuclear power plant produces between 1 and 1.6 GW of electricity per reactor per hour, so data centers running OpenAI’s GPT-5 at 18 Wh per query could require the power equivalent of two to three nuclear power reactors, an amount that could be enough to power a small country.

  • MangoCats@feddit.it
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    edit-2
    4 hours ago

    This figure is already not bad. 40 watt hours = 0.04kWh - you know kWh? That unit on your electric bill that is around $0.18 per kWh (and data centers tend to be in lower cost electric areas, closer to $0.11/kWh.) Still, 40Wh would register on your home electric bill at $0.0072, less than a penny. For comparison, an average suburban 4 ton AC unit draws 4kW - that 40Wh request? 1/100th of an hour of AC for your home, about 36 seconds of air conditioning. I don’t know that this article is making anybody “look bad” in terms of power used.

    • msage@programming.dev
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 hours ago

      What exactly do you get for that power though?

      The point is that it’s too much power for little gain in return.