NotebookCHECK - Notebook Forum

English => News => Topic started by: Redaktion on July 26, 2024, 05:09:29

Title: Meta unveils biggest, smartest, royalty-free Llama 3.1 405B AI
Post by: Redaktion on July 26, 2024, 05:09:29
Meta has unveiled its biggest, smartest, most-neutered Llama 3.1 405B AI for royalty-free use. The 750 GB, 405 billion parameter large language model (LLM) is one of the biggest ever released and performs competitively with flagship competitors Anthropic Claude 3.5 Sonnet and OpenAI GPT-4o.

https://www.notebookcheck.net/Meta-unveils-biggest-smartest-royalty-free-Llama-3-1-405B-AI.866775.0.html
Title: Re: Meta unveils biggest, smartest, royalty-free Llama 3.1 405B AI
Post by: Dark Oopa on July 26, 2024, 10:31:57
11 GWh is totally impossible.
It's the Equivalent of the annual consumption of China + US, almost half of the world...
Title: Re: Meta unveils biggest, smartest, royalty-free Llama 3.1 405B AI
Post by: Dark Oopa on July 26, 2024, 11:54:45
Well, I said crap, totally mixed the units...

But after ccalculs, I don't understand the 11GWh
So, we had 39.3M of hours of work for 700W GPUs : which means 27.510 GWh, not 11.
The best cloud datacenter have an overhead (extra consumption like cooling etc) of 10%, 58% for the average.

Let's take 15% overhead, which means we are closer to 31.6 GWh.
Google uses around 15 TWh a year so 41 GWh/day

A US citizen uses on average 11.2MWh/year so, the training alone uses the equivalent of the yearly consumption of 2 800 americans.

And that is just the training, I'm afraid it's nothing compare to running the thing for all cloud users.