If this page does not answer your question, please refer to our support channels.
What is the Edge TPU?
The Edge TPU is a small ASIC designed by Google that provides high performance ML inferencing for low-power devices. For example, it can execute state-of-the-art mobile vision models such as MobileNet V2 at 400 FPS, in a power efficient manner.
We offer multiple products that include the Edge TPU built-in.
What machine learning frameworks does the Edge TPU support?
TensorFlow Lite only.
What type of neural networks does the Edge TPU support?
The first-generation Edge TPU is capable of executing deep feed-forward neural networks (DFF) such as convolutional neural networks (CNN), making it ideal for a variety of vision-based ML applications.
For details on supported model architectures, see the model requirements.
How do I create a TensorFlow Lite model for the Edge TPU?
You need to convert your model to TensorFlow Lite and it must be quantized using either quantization-aware training (recommended) or full integer post-training quantization. Then you need to compile the model for compatibility with the Edge TPU. For more information, read TensorFlow models on the Edge TPU.
However, if you're building an image classification application, you can also use Cloud AutoML Vision to easily build a model that's compatible with the Edge TPU. This web-based tool provides a graphical UI to train a model with your own images, optimize the model, and then export it for the Edge TPU.
What's the Edge TPU's processing power?
The Edge TPU is capable of performing 4 trillion operations (tera-operations) per second (TOPS), using 0.5 watts for each TOPS (2 TOPS per watt).
What kind of performance does it actually provide?
Can the Edge TPU perform accelerated ML training?
Yes, but only to retrain the final layer. Because a TensorFlow model must be compiled for acceleration on the Edge TPU, we cannot later update the weights across all the layers. However, we do provide APIs that perform accelerated transfer-learning in two different ways:
- Backpropagation that updates weights for just the final fully-connected layer, using a cross-entropy loss function.
- Weight imprinting that uses the image embeddings from new data to imprint new activation vectors in the final layer, which allows for learning new classifications with very small small datasets.
For more information about both, read about transfer learning on-device.
What's the difference between the Dev Board and the USB Accelerator?
The Coral Dev Board is a single-board computer that includes an SOC and Edge TPU integrated on the SOM, so it's a complete system. You can also remove the SOM (or purchase it separately) and then integrate it with other hardware via three board-to-board connectors—even in this scenario, the SOM contains the complete system with SOC and Edge TPU, and all system interfaces (I2C, MIPI-CSI/DSI, SPI, etc.) are accessible via 300 pins on the board-to-board connectors.
Whereas, the Coral USB Accelerator is an accessory device that that adds the Edge TPU as a coprocessor to your existing system—you can simply connect it to any Linux-based system with a USB cable (we recommend USB 3.0 for best performance).
Can I buy just the Edge TPU chip?
No, we currently do not offer the standalone Edge TPU ASIC, but we do provide two products that make it easy to integrate the Edge TPU as a coprocessor for existing hardware systems using either USB 3.0 or a PCI-E interface. For details, see our products list.
If you have other specific requirements, please contact our sales team and we will be happy to discuss possible solutions.
Is this content helpful?