Not so great news for Nvidia Volta V100 as Google revealed some details of it's
TPU2.
180 TFLOP/s both for training and inferencing.
They are not direct competitors. The use case for V100 is vastly different to the TPU (they happen to overlap in Tensor based work). The TPU is pretty much the definition of an ASIC (application specific integrated circuit), in that it just does Tensor based work and thats it.
The V100 on the otherhand can be used for a variety of computations which also includes Tensor based work (it can do this stuff in parallel with FP64/FP32 work for instance). Perhaps post processing with the CUDA cores while running machine learning algorithm via Tensor cores. Not just deep learning but it can do other things which makes it different to the TPU.
And to my knowledge Google happens to be one of nVIDIAs HPC customers.