How do I print in Google Colab which TPU version I am using and how much memory the TPUs have?
With I get the following Output
JavaScript
x
6
1
tpu = tf.distribute.cluster_resolver.TPUClusterResolver()
2
tf.config.experimental_connect_to_cluster(tpu)
3
tf.tpu.experimental.initialize_tpu_system(tpu)
4
5
tpu_strategy = tf.distribute.experimental.TPUStrategy(tpu)
6
Output
JavaScript
1
42
42
1
INFO:tensorflow:Initializing the TPU system: grpc://10.123.109.90:8470
2
INFO:tensorflow:Initializing the TPU system: grpc://10.123.109.90:8470
3
INFO:tensorflow:Clearing out eager caches
4
INFO:tensorflow:Clearing out eager caches
5
INFO:tensorflow:Finished initializing TPU system.
6
INFO:tensorflow:Finished initializing TPU system.
7
WARNING:absl:`tf.distribute.experimental.TPUStrategy` is deprecated, please use the non experimental symbol `tf.distribute.TPUStrategy` instead.
8
INFO:tensorflow:Found TPU system:
9
INFO:tensorflow:Found TPU system:
10
INFO:tensorflow:*** Num TPU Cores: 8
11
INFO:tensorflow:*** Num TPU Cores: 8
12
INFO:tensorflow:*** Num TPU Workers: 1
13
INFO:tensorflow:*** Num TPU Workers: 1
14
INFO:tensorflow:*** Num TPU Cores Per Worker: 8
15
INFO:tensorflow:*** Num TPU Cores Per Worker: 8
16
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:localhost/replica:0/task:0/device:CPU:0, CPU, 0, 0)
17
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:localhost/replica:0/task:0/device:CPU:0, CPU, 0, 0)
18
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:localhost/replica:0/task:0/device:XLA_CPU:0, XLA_CPU, 0, 0)
19
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:localhost/replica:0/task:0/device:XLA_CPU:0, XLA_CPU, 0, 0)
20
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:CPU:0, CPU, 0, 0)
21
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:CPU:0, CPU, 0, 0)
22
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:0, TPU, 0, 0)
23
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:0, TPU, 0, 0)
24
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:1, TPU, 0, 0)
25
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:1, TPU, 0, 0)
26
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:2, TPU, 0, 0)
27
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:2, TPU, 0, 0)
28
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:3, TPU, 0, 0)
29
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:3, TPU, 0, 0)
30
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:4, TPU, 0, 0)
31
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:4, TPU, 0, 0)
32
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:5, TPU, 0, 0)
33
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:5, TPU, 0, 0)
34
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:6, TPU, 0, 0)
35
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:6, TPU, 0, 0)
36
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:7, TPU, 0, 0)
37
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU:7, TPU, 0, 0)
38
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU_SYSTEM:0, TPU_SYSTEM, 0, 0)
39
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:TPU_SYSTEM:0, TPU_SYSTEM, 0, 0)
40
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:XLA_CPU:0, XLA_CPU, 0, 0)
41
INFO:tensorflow:*** Available Device: _DeviceAttributes(/job:worker/replica:0/task:0/device:XLA_CPU:0, XLA_CPU, 0, 0)
42
I want something like [IN] print('tpu.version:', tpu.version) [OUT] tpu.version: Cloud TPU v3
.
I hope someone can help me. Looking forward to an answer!
Advertisement
Answer
This answer should have what you’re looking for.
JavaScript
1
6
1
import os
2
from tensorflow.python.profiler import profiler_client
3
4
tpu_profile_service_address = os.environ['COLAB_TPU_ADDR'].replace('8470', '8466')
5
print(profiler_client.monitor(tpu_profile_service_address, 100, 2))
6
Output looks like:
JavaScript
1
4
1
Timestamp: 22:23:03
2
TPU type: TPU v2
3
Utilization of TPU Matrix Units (higher is better): 0.000%
4
TPUv2 has 8GB per-core and TPUv3 has 16GB HBM per-core (https://cloud.google.com/tpu).