If we assume that this is the energy in its aggregate form drawn from the grid, the sad news is only about 100 TWh would've really got used - the remaining having been wasted in various stages. [Refer Schneider's calculations in this regard on how energy is wasted in data centers.]
Firstly, what would it take to generate that much of electricity? Here are some options...
- Nuclear generation: Run one of the two 1,000 MegaWatt units of Kudankulam Nuclear Plant at full capacity for 1,000 hours = 42 days on a 24x7 basis
- Thermal generation: Burn 149x3 ~ 450 Kilotonnes of Coal [assuming 6.67 KWh/Kg of energy per Kg of coal and 30% as the thermodynamic efficiency of thermal power plants]
- Solar generation: Roughly 220 Sq.Km of Solar Panels [assuming 70 MilliWatts/Sq.Inch]
What can I do with it if I weren't running data centers? Well, to begin with, I can burn a tubelight or run a laptop at full power for 2.8 million years on a 24x7x365 basis. In other words, 5.6 Million houses can have at least 1 tubelight burning throughout the year for 12 hours a day - enough for 10 Million children to study if we assume each house has 2 children !