On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜
One of the earliest and most influential critiques of large language models. Many still look up-to-date. Unfortunately, 5 years later, the problems have never been solved.
Most LLM company still refuse to disclose their environmental impact and usage of energy resource, even including Gemini, in their latest documentation of model cards and safety reports. Astonishing data discovered:
While the average human is responsible for an estimated 5t per year, a Transformer (big) model [136] with neural architecture search and estimated that the training procedure emitted 284t of CO2.
Interesting but quite ideal viewpoints to enhance the curation and documentation by LLM company themselves. Even newly release model nowadays are not equipped with precise and clear instructions and model cards. Probably not going to work in a fast-competing AI era for accuracy and latency.
