推论
失败
能源消耗
计算机科学
乘法函数
深度学习
人工智能
消费(社会学)
指数增长
透视图(图形)
光学(聚焦)
指数函数
机器学习
数据科学
运筹学
计量经济学
经济
工程类
数学
社会学
并行计算
电气工程
光学
物理
数学分析
社会科学
作者
Radosvet Desislavov,Fernando Martínez‐Plumed,José Hernández‐Orallo
标识
DOI:10.1016/j.suscom.2023.100857
摘要
The progress of some AI paradigms such as deep learning is said to be linked to an exponential growth in the number of parameters. There are many studies corroborating these trends, but does this translate into an exponential increase in energy consumption? In order to answer this question we focus on inference costs rather than training costs, as the former account for most of the computing effort, solely because of the multiplicative factors. Also, apart from algorithmic innovations, we account for more specific and powerful hardware (leading to higher FLOPS) that is usually accompanied with important energy efficiency optimisations. We also move the focus from the first implementation of a breakthrough paper towards the consolidated version of the techniques one or two year later. Under this distinctive and comprehensive perspective, we study relevant models in the areas of computer vision and natural language processing: for a sustained increase in performance we see a much softer growth in energy consumption than previously anticipated. The only caveat is, yet again, the multiplicative factor, as future AI increases penetration and becomes more pervasive.
科研通智能强力驱动
Strongly Powered by AbleSci AI