But again, that's the thing. There is only 1/Y work to do in the day; it's batch work. The work in this case is constrained on the input side, not the CPU resources side; building 1 or 2 or 1,000 nodes to do the work won't decrease how expensive it is to do the work (in fact, building more computers than you need will make it cost more!).
... so why does Google build more computers than they need? Keep in mind that at Google scale, they're always and forever building "As many computers as we can possibly afford to" under the assumption that there will always be work for those machines to do. You and I may need to consider cost of manufacturing; Google doesn't. They always have the "Build datacenter infrastructure" cranked to an 11 (more accurately, they are following an N-year plan of construction that is extremely expensive to modify).
That's the breakdown between Google's way of thinking and the way of thinking that you've presented: Google's cost of manufacture is fixed. Those computers will be built, whether or not they're going to also then run green-streamlined batch jobs. The limiting factor on the batch work is only so much work is generated in a day, and the rate the work is completed is already good enough that completing it in half the time yields no marginal value. So may as well complete it using sun instead of coal.
That may well be the way you (and most likely Google) currently look at it, but the argument presented seems to be that everything other than where you run some batch tasks is all fixed.
Assuming everything else is fixed does make the optimisation easier, though surely nearly everything is up for debate if we're actually talking about minimising the overall footprint of a compute task?
Using fossil fuels rather than allowing these fixed costs you mention to sit idle - there is a point where your carbon/CPU cycle would actually go up by not using fossil fuels some of the time (though I am assuming less carbon emitted for a unit of power than manufacture cost).
I appreciate they are ever-expanding and predicting where non-movable workloads are going to need to be run etc etc, and I'm not suggesting there are easy answers.
Is the question on the table whether, all other things being equal, this change would decrease carbon output per unit work completed, or is the question whether Google is greener after this change?
For the latter question, we have insufficient data. Google's datacenter infrastructure is huge and complicated, and if one factors in all the interdependencies and purchased carbon offsets, one needs way more data than this announcement blurb gives out to answer that question. Maybe they have an additional process to determine whether their most carbon-negative datacenters can be switched off during peak coal-use and this work unblocked them from enabling that feature? We don't know.
If I have 2 hours of work to do a day, I could sell 90% of my servers, the racks, the networking hardware, one of the air conditioning units, backup generators, and battery backups, lay off some of my IT staff, and just let the thing run 23 hours a day.
What people are trying to tell you is you can’t make having a glut of equipment work out for any reason, let alone environmental. It costs you a dozen times over to have most of you hardware not doing anything at all most of the day. You are overprovisioned.
You are so overprovisioned in fact that someone will steal your promotion by pointing this out to your boss, which is also a pretty big cost.
> What people are trying to tell you is you can’t make having a glut of equipment work out for any reason, let alone environmental.
Google has a glut of equipment for two reasons: variable loads and experimental.
The batch loads are predictable and constrained by input rate. What the coal ends up used for is spikes in demand for compute power (i.e. "Michael Jackson died and now everyone needs to watch Thriller right now") and ceiling to develop and experiment with new projects.
This optimization means that while, yes, the fossil fuel resources would be feeding those use cases, it's a net gain because those use cases are less common than the predictable batch-job work.
> You are so overprovisioned in fact that someone will steal your promotion by pointing this out to your boss, which is also a pretty big cost.
Remember the context. Nobody ever got fired at Google for adding more compute power to the behemoth, because it's all fungible and will all be used eventually. if you're trying to argue "Google is a net polluter because they continue to build machines in a fashion only bounded by land and hardware costs under the assumption more compute power is always better than less" it's an argument you could make but (a) you should factor in the carbon offsets they buy and (b) you're tipping close to arguing "All human activity is pollution; kill the species."
... so why does Google build more computers than they need? Keep in mind that at Google scale, they're always and forever building "As many computers as we can possibly afford to" under the assumption that there will always be work for those machines to do. You and I may need to consider cost of manufacturing; Google doesn't. They always have the "Build datacenter infrastructure" cranked to an 11 (more accurately, they are following an N-year plan of construction that is extremely expensive to modify).
That's the breakdown between Google's way of thinking and the way of thinking that you've presented: Google's cost of manufacture is fixed. Those computers will be built, whether or not they're going to also then run green-streamlined batch jobs. The limiting factor on the batch work is only so much work is generated in a day, and the rate the work is completed is already good enough that completing it in half the time yields no marginal value. So may as well complete it using sun instead of coal.