Excellent podcast interview with a very prominent person from computing word,
co-creator of RISC processor design, including modern and very popular "open hardware" RISC V".
But the most interesting part of conversation was about a mistake in a scientific paper
where energy usage used for AI training was over-estimated 120 000 times!
That was based on public data from Google, that are miss-interpreted, and without insight of how Google is actually doing AI training etc. David Patterson knows it well, since he now works in Google
and is deeply involved in energy optimization.
And now, with this wrong info published and cited many times, people are making all kinds of assumptions and even plans! Yes, data centers are using a lot of energy, but not even near to levels some people are presenting it to be.
Turing Award Special: A Conversation with David Patterson - Software Engineering Daily
Good News About the Carbon Footprint of Machine Learning Training
"Unfortunately, some ... papers misinterpreted the NAS estimate as the training cost for the model it discovered, yet emissions for this particular NAS are ~1300x larger than for training the model. These papers estimated that training the Evolved Transformer model takes two million GPU hours, costs millions of dollars, and that its carbon emissions are equivalent to five times the lifetime emissions of a car. In reality, training the Evolved Transformer model on the task examined by the UMass researchers and following the 4M best practices takes 120 TPUv2 hours, costs $40, and emits only 2.4 kg (0.00004 car lifetimes), 120,000x less. This gap is nearly as large as if one overestimated the CO2e to manufacture a car by 100x and then used that number as the CO2e for driving a car."
No comments:
Post a Comment