Accelerator
Cloud TPU v5p
Google TPU v5p TensorCore cloud tpu chip summary for training,
inference, and roofline-style performance analysis.
Back to Accelerator Catalog
Vendor Google
Architecture TPU v5p TensorCore
Unit Cloud TPU chip
Form factor Cloud TPU slice chip
Launch 2023-12-06
Memory 95 GB HBM2e
HBM bandwidth 2.765 TB/s
BF16 peak 459 TFLOPS
FP16 peak n/a
FP8 dense peak n/a
FP8 sparse peak n/a
FP4 dense peak n/a
FP4 sparse peak n/a
FP64 peak n/a
INT8 peak n/a
Interconnect ICI 3D torus - 1200 GB/s bidirectional per chip
Power Not published per chip
Software stack JAX, XLA, TensorFlow, PyTorch/XLA Notes
- v5p is oriented toward large-scale training and uses a 3D torus topology for full-cube and larger slices.
- Google quotes BF16 peak compute per chip rather than a broad precision table like GPU vendors.
Sources