Cresta Sun Hotel Davos
Tuesday, January 21
11:00-11:40am CET
Congress Centre, Aula
Tuesday, January 21
1:00-1:45pm CET
7270 Davos Platz, Switzerland
Wednesday, January 22
10:00-11:00am CET
Dome in Davos
Wednesday, January 22
5:00-5:30pm CET
We asked industry leaders about the state of AI and our survey results highlight a critical gap in how industry leaders perceive AI’s energy needs. Read the full release »
While 70.0% of leaders recognize the energy-intensive nature of training large language models, only 59.7% are aware of the significant power demands of inference. This gap is critical as inference workloads are set to dominate AI usage with the scaling of Agentic AI.
Only 13.0% of organizations currently monitor AI power consumption, yet 56.5% acknowledge that energy efficiency will play a crucial role in future strategic planning. The need to address rising energy demands is being driven by both cost management imperatives and operational scalability concerns.
The roll-out of Agentic AI is amplifying power concerns for enterprises. For 20.3% of companies, rising power costs are a pressing issue. Furthermore, 37.2% are experiencing increasing stakeholder pressure to improve AI energy efficiency, with a further 42.0% expecting these demands to emerge soon.
Among organizations that have widely deployed AI2, 77.4% are actively seeking to reduce power usage. Popular approaches include hardware and software optimization (40.4%), adopting energy-efficient processors (39.3%), and investing in renewable energy (34.9%). However, these measures remain insufficient compared to the rapid pace of AI adoption and scaling.