We use ChatGPT, Claude, Copilot or Gemini to AI generate answers to our questions as if nothing were happening. After all, in many cases it is free, right? It is for users at the moment, yes, but all those responses—let alone images or videos generated by AI—cost these companies a lot of money. How much? None says it.
burning money. We have already talked about how both OpenAI and the rest of the companies that have opted for AI are burning money like there was no tomorrow. Sam Altman recently indicated that OpenAI will not be profitable until 2029but until then—if they arrive—they will continue to lose money. A lot. An analysis by some analysts reveals how the management and operation of AI models has an enormous cost, and for example OpenAI will have estimated losses of 5000 million dollars… only in 2024!
Difficult to find out the real cost of AI. The companies have hardly provided data on the internal cost of operating these services, although in The Washington Post they indicated in June 2023 something that is probably true: AI chatbots they lose money every time you use them. Some independent studies already pointed out in 2023 that the cost of a simple question on ChatGPT was 36 cents on the dollar, and that OpenAI was spent $700,000 a day to keep his business running LLM.
And now the models want to reason (and that costs more). The launch of o1, the OpenAI model that “reasons”has been the starting signal for an alternative that theoretically makes fewer mistakes and that, although it takes longer to give the answer, does so in a more complete and precise way. But also more expensive: these systems actually generate several responses, analyze and review them and keep the one they consider most appropriate to show to the user. And that is more expensive for the AI company and for the user.
That’s why the $200 a month subscription is just the beginning. That OpenAI has launched a subscription of 200 dollars a month to be able to use its reasoning model without limitations, o1, is a good sign of how expensive it is for the company to offer an “open bar of artificial intelligence.” Certain users and scenarios benefit intensively from AI chatbots, and it will be in these cases where companies will be able to advocate for offering subscriptions increasingly powerful… and expensive.
But for images and videos, the cost is even higher. Generating text is the least resource-demanding task, yet it is expensive. It is also difficult to estimate this increase in cost because there are many factors at play: efficiency of the servers, complexity and resolution of the image or video, etc. Even so, it is likely that generating an image will cost several orders of magnitude more than generating text, and generating a video will also cost several orders of magnitude more than generating an image.
Lots of energy… John Hennessy, one of the executives of Alphabet, explained in Reuters that the responses generated by an LLM consume 10 times more energy than a conventional Google search. The implications are enormous —also for the environment—, and in fact it is estimated that Energy consumption could triple by 2030. Above all, because of the data centers that are being built and that will be in charge of managing all these requests. Google and Microsoft already consume alone more than 100 different countries.
…and water. There is also a problem with the water needed to cool those data centers. According to a study published by Washington Postasking the AI to write us a 100-word email translates into a consumption of 519 milliliters of water.
Seeking efficiency. Of course the hope is that these training and inference costs will be significantly reduced in the future. The development of new chips, more efficient data centers – and that do not waste water, like those that Microsoft has designed—can help with that.
AI on premises. The other alternative that can help will be the progressive adoption of “Edge AI” models that run locally. Apple Intelligence is a good example, and here the company proposes using certain functions – be careful, not all – directly on the device, which allows us to not only save energy, but also guarantee the privacy of our requests and conversations with our mobile phone or computer.