Cloud Bigtable

A fully managed, scalable NoSQL database service for large analytical and operational workloads.

Try Google Cloud free
  • action/check_circle_24px Created with Sketch.

    Consistent sub-10ms latency—handle millions of requests per second

  • action/check_circle_24px Created with Sketch.

    Ideal for use cases such as personalization, ad tech, fintech, digital media, and IoT

  • action/check_circle_24px Created with Sketch.

    Seamlessly scale to match your storage needs; no downtime during reconfiguration

  • action/check_circle_24px Created with Sketch.

    Designed with a storage engine for machine learning applications leading to better predictions

  • action/check_circle_24px Created with Sketch.

    Easily connect to Google Cloud services such as BigQuery or the Apache ecosystem

Fast and performant

Use Cloud Bigtable as the storage engine that grows with you from your first gigabyte to petabyte-scale for low-latency applications as well as high-throughput data processing and analytics.

Seamless scaling and replication

Start with a single node per cluster, and seamlessly scale to hundreds of nodes dynamically supporting peak demand. Replication also adds high availability and workload isolation for live serving apps.

Simple and integrated

Fully managed service that integrates easily with big data tools like Hadoop, Dataflow, and Dataproc. Plus, support for the open source HBase API standard makes it easy for development teams to get started.

Key features

High throughput at low latency

Bigtable is ideal for storing very large amounts of data in a key-value store and supports high read and write throughput at low latency for fast access to large amounts of data. Throughput scales linearly—you can increase QPS (queries per second) by adding Bigtable nodes. Bigtable is built with proven infrastructure that powers Google products used by billions such as Search and Maps.

Cluster resizing without downtime

Scale seamlessly from thousands to millions of reads/writes per second. Bigtable throughput can be dynamically adjusted by adding or removing cluster nodes without restarting, meaning you can increase the size of a Bigtable cluster for a few hours to handle a large load, then reduce the cluster's size again—all without any downtime.

Flexible, automated replication to optimize any workload

Write data once and automatically replicate where needed with eventual consistency—giving you control for high availability and isolation of read and write workloads. No manual steps needed to ensure consistency, repair data, or synchronize writes and deletes. Benefit from a high availability SLA of 99.99% for instances with multi-cluster routing (99.9% for single-cluster instances).

Customers

Ex: A reference architecture demonstrating the ML workflow.
Dow Jones brings key historical events datasets to life with Cloud Bigtable.
Read the story

Story highlights

  • Synthesized 30+ years of news data to assess business impact

  • Uncovered hidden data relationships and insights

  • Prototype Knowledge Graph delivered with ease in 10 weeks

Partner

What's new

Sign up for Google Cloud newsletters to receive product updates, event information, special offers, and more.

Documentation

Quickstart
Quickstart using the cbt tool

Learn Cloud Bigtable basics in the quickstart that uses the Cloud Console and the cbt command-line tool.

Tutorial
Codelab: Introduction to Cloud Bigtable

Step through a Cloud Bigtable codelab that teaches you how to avoid common schema design mistakes, import data, and then query and use it.

APIs & Libraries
Cloud Bigtable client libraries

Work with Cloud Bigtable using a Google Cloud client library in your preferred programming language.

Google Cloud Basics
Table-level IAM management

Manage access control for Cloud Bigtable at the project, instance, and table level.

Tutorial
Creating a Cloud Bigtable instance

Learn how to create a Cloud Bigtable instance using the Cloud Console or command-line tools.

Google Cloud Basics
Bigtable quickstart: Command line

Learn first-hand how to use the cbt command line to connect to a Cloud Bigtable instance, perform basic admin tasks, and read and write data in a table.

Google Cloud Basics
Go global with Cloud Bigtable

Cloud Bigtable's replication capabilities give you the flexibility to make your data available across a region or worldwide.

Google Cloud Basics
Optimize schema performance with Key Visualizer

Key Visualizer lets you see key access patterns in heatmap format to optimize your Cloud Bigtable schemas for improved performance.

Google Cloud Basics
Protect and control your Google Cloud services and data

VPC Service Controls create a security perimeter around data stored in Bigtable, helping mitigate data exfiltration risks.

Use cases

Use case
Financial analysis

Build models based on historical behavior. Continually update fraud patterns and compare with real-time transactions. Store and consolidate market data, trade activity, and other data, such as social and transactional data.

Financial analysis use case diagram
Use case
IoT

Ingest and analyze large volumes of time series data from sensors in real time, matching the high speeds of IoT data to track normal and abnormal behavior. Enable customers to build dashboards and drive analytics on their data in real time.

IoT use case diagram
Use case
AdTech

Integrate large volumes of unrefined data from many sources, typically to drive consistent customer activity across channels. Collect and compare large volumes of behavior data across customers to find common patterns that can drive recommendations and sales.

Ex: A simplified, overall flow of data between an on-premises data warehouse like Teradata and BigQuery..

Pricing

Cloud Bigtable is a fast, fully managed, massively scalable NoSQL database service. For detailed pricing information, please view the pricing guide.

Integrations

Cloud Bigtable integrates with the Apache® ecosystem and other Google Cloud products to analyze, process, and store data. For more details, refer to Integrations documentation.