Cloud TPU pricing

    TPU pricing and quota are divided into two systems:

    • Single device TPU type pricing for individual TPU devices that are available either on-demand or as preemptible devices. You cannot combine multiple single device TPU types to collaborate on a single workload.
    • TPU Pod type pricing for clusters of TPU devices that are connected to each other over dedicated high-speed networks. These TPU types are available only if you have evaluation quota or purchase a 1 year or 3 year commitment.

    Charges for Cloud TPU accrue while your TPU node is in a READY state. You receive a bill at the end of each billing cycle that lists usage and charges for that billing cycle.

    To learn how to request quota for single device TPU types, read the Cloud TPU quota policy page.

    Single device pricing

    Single device TPU types are billed in one-second increments and are available at either an on-demand or preemptible price.

    Single device TPU types are independent TPU devices without direct network connections to other TPU devices in a Google data center. If your workloads require more TPU cores and a larger pool of memory, use a TPU Pod type instead.

    A preemptible TPU is one that Cloud TPU can terminate (preempt) at any time if Cloud TPU requires access to the resources for another task. The charges for a preemptible TPU are much lower than those for a normal TPU. You are not charged for preemptible TPUs if they are preempted in the first minute after you create them.

    You can configure your TPU nodes with the following single device TPU types:

    TPU Pod type pricing

    TPU Pod types provide access to multiple TPU devices that are all connected on a dedicated high-speed network. These TPU types provide greater compute capacity and a larger pool of TPU memory to a single TPU node. To use TPU Pod types, you must request quota using one of the following options:

    • Request access to evaluation quota so that you can test the performance of TPU Pod types. TPU nodes that you create using evaluation quota are billed in one-second increments but do not guarantee the same level of service as on-demand TPU devices or devices that you create using commitment quota. Evaluation quota persists only for a limited amount of time on your project.
    • Purchase a 1 year or 3 year commitment and create TPU nodes with up to 2048 cores. Commitments are not billed incrementally. Commitments allow for access to reserved cores for all hours of the day on an on-going month over month basis for the duration of the contract. Commitments bill you a monthly fee for the duration of your commitment term even if you do not use any TPU resources.

    You can configure your TPU nodes with the following TPU types:

    To learn about the differences between different TPU versions and configurations, read the TPU System Architecture documentation.

    Total cost for 1 year and 3 year commitments

    The following table shows the total cost for various TPU commitments during their full duration:

    To request access to TPU types with more than 8 cores, contact a sales representative.

    TensorFlow Research Cloud (TFRC) program

    If you're enrolled in the TFRC program you are granted access to Cloud TPU v2 and v3 for a limited period of time free of charge. Specifically for the TFRC program, you are not charged for Cloud TPU as long as your TPU nodes run in the us-central1-f zone.

    Virtual machine pricing

    In order to connect to a TPU, you must provision a virtual machine (VM), which is billed separately. For details on pricing for VM instances, see Compute Engine pricing.

    Pricing calculator

    To estimate the cost of using Cloud TPU with Compute Engine VM instances, see the Compute Engine pricing calculator.

    Pricing example

    The following example explains how to calculate the total cost of a training job that uses TPU resources and Compute Engine instances in the US region.

    A machine learning researcher provisions a virtual machine by creating a Compute Engine instance, and they select the n1-standard-2 machine type. They also create a TPU resource, and they accrue 10 hours of usage on both the Compute Engine instance and the TPU resource. In order to calculate the total cost of a training job, the machine learning researcher must add together:

    • the total cost of all Compute Engine instances
    • the total cost of all Cloud TPU resources
    Resource Price per machine per hour in USD Number of machines Number of hours billed Total cost of each resource Total cost of training job
    Compute Engine
    n1-standard-2 instance
    $0.095 1 10 $0.95
    Cloud TPU resource $4.50 1 10 $45.00
    $45.95

    Pricing example using a preemptible TPU

    The following example uses the same resources and time period as above, except that the researcher decides to use a preemptible TPU to save costs. The US charge for the preemptible TPU v2 is $1.35 USD per hour, as opposed to $4.50 for a normal TPU v2 .

    Resource Price per machine per hour in USD Number of machines Number of hours billed Total cost of each resource Total cost of training job
    Compute Engine
    n1-standard-2 instance
    $0.095 1 10 $0.95
    Preemptible TPU $1.35 1 10 $13.50
    $14.45

    What's next

    Was this page helpful?