Looking at the derating graph for the standard cell libraries, we see that in the operating region, the frequency/voltage curve is roughly linear. CMOS delay is inversely proportional to supply voltage.
Logic with higher-speed capabilities is smaller which means it consumes greater leakage current which is being wasted while we are halted. Also leakage energy is proportional to supply voltage (or perhaps sublinear with exponent 0.9ish : as we raise voltage, transistors are indeed turned off more, but P=IV is still increasing).
![]() |
Voltage to a region may be varied dynamically. A higher
Operating region of the frequency/voltage curve is roughly linear.
But, logic with higher-speed capabilities is smaller which means it
Let's only raise VCC when we ramp up f: classical DVFS. |
Hence, we aim to clock at the lowest suffcient voltage and clock frequency and for the activity ratio as high as possible and minimal halt cycles.
In general, ramping voltage up linearly with clock frequency (f) results in dynamic power consumption with a cubic dependence on f. But work may be bursty, so DVFS is applied (e.g. by a laptop governor).
For all loads, this DVSF approach achieves peak performance under heavy loads and avoids cubic overhead when idle.
In a server farm processing blade we may be thermally limited, so DVFS will be throttled back by rack-level governors or Intel's RAPL.
11: (C) 2008-17, DJ Greaves, University of Cambridge, Computer Laboratory. |