This guide provides an overview of how to move your on-premises Apache Hadoop system to Google Cloud Platform (GCP). It describes a migration process that not only moves your Hadoop work to GCP, but also enables you to adapt your work to take advantage of the benefits of a Hadoop system optimized for cloud computing. It also introduces some fundamental concepts you need to understand to translate your Hadoop configuration to GCP.
This is the first of several guides describing how to move from on-premises Hadoop:
- This guide, which provides context and planning advice for your migration.
- Migrating HDFS Data from On-Premises to Google Cloud Platform provides additional context for incrementally moving your data to GCP.
- Migrating data from HBase to Cloud Bigtable
- Migrating Hadoop Jobs from On-Premises to Cloud Dataproc describes the process of running your jobs on Cloud Dataproc and other GCP products.
The benefits of migrating to GCP
There are many ways in which using GCP can save you time, money, and effort compared to using an on-premises Hadoop solution. In many cases, adopting a cloud-based approach can make your overall solution simpler and easy to manage.
Built-in support for Hadoop
GCP includes Cloud Dataproc, which is a managed Hadoop and Spark environment. You can use Cloud Dataproc to run most of your existing jobs with minimal alteration, so you don't need to move away from all of the Hadoop tools you already know.
Managed hardware and configuration
When you run Hadoop on GCP, you never need to worry about physical hardware. You specify the configuration of your cluster, and Cloud Dataproc allocates resources for you. You can scale your cluster at any time.
Simplified version management
Keeping open source tools up to date and working together is one of the most complex parts of managing a Hadoop cluster. When you use Cloud Dataproc, much of that work is managed for you by Cloud Dataproc versioning.
Flexible job configuration
A typical on-premises Hadoop setup uses a single cluster that serves many purposes. When you move to GCP, you can focus on individual tasks, creating as many clusters as you need. This removes much of the complexity of maintaining a single cluster with growing dependencies and software configuration interactions.
Planning your migration
Migrating from an on-premises Hadoop solution to GCP requires a shift in approach. A typical on-premises Hadoop system consists of a monolithic cluster that supports many workloads, often across multiple business areas. As a result, the system becomes more complex over time and can require administrators to make compromises to get everything working in the monolithic cluster. When you move your Hadoop system to GCP, you can reduce the administrative complexity. However, to get that simplification and to get the most efficient processing in GCP with the minimal cost, you need to rethink how to structure your data and jobs.
Because Cloud Dataproc runs Hadoop on GCP, using a persistent Cloud Dataproc cluster to replicate your on-premises setup might seem like the easiest solution. However, there are some limitations to that approach:
- Keeping your data in a persistent HDFS cluster using Cloud Dataproc is more expensive than storing your data in Cloud Storage, which is what we recommend, as explained later. Keeping data in an HDFS cluster also limits your ability to use your data with other GCP products.
- Augmenting or replacing some of your open-source-based tools with other related GCP services can be more efficient or economical for particular use cases.
- Using a single, persistent Cloud Dataproc cluster for your jobs is more difficult to manage than shifting to targeted clusters that serve individual jobs or job areas.
The most cost-effective and flexible way to migrate your Hadoop system to GCP is to shift away from thinking in terms of large, multi-purpose, persistent clusters and instead think about small, short-lived clusters that are designed to run specific jobs. You store your data in Cloud Storage to support multiple, temporary processing clusters. This model is often called the ephemeral model, because the clusters you use for processing jobs are allocated as needed and are released as jobs finish.
The following diagram shows a hypothetical migration from an on-premises system to an ephemeral model on GCP.
The example moves four jobs that run on two on-premises clusters to Cloud Dataproc. The ephemeral clusters that are used to run the jobs in GCP are defined to maximize efficiency for individual jobs. The first two jobs use the same cluster, while the third and fourth jobs each run on their own cluster. When you migrate your own jobs, you can customize and optimize clusters for individual jobs or for groups of jobs as makes sense for your specific work. Cloud Dataproc helps you quickly define multiple clusters, bring them online, and scale them to suit your needs.
The data in the example is moved from two on-premises HDFS clusters to Cloud Storage buckets. The data in the first cluster is divided among multiple buckets, and the second cluster is moved to a single bucket. You can customize the structure of your data in Cloud Storage to suit the needs of your applications and your business.
The example migration captures the beginning and ending states of a complete migration to GCP. This implies a single step, but you'll get the best results if you don't think of moving to GCP as a one-time, complete migration. Instead, think of it as refactoring your solutions to use a new set of tools in ways that weren't possible on-premises. To make such a refactoring work, we recommend migrating incrementally.
Recommended sequence for migrating to GCP
Here are the recommended steps for migrating your workflows to GCP:
Move your data first
- Move your data into Cloud Storage buckets.
- Start small. Use backup or archived data to minimize the impact to your existing Hadoop system.
- Use a subset of data to test and experiment. Make a small-scale proof of concept for each of your jobs.
- Try new approaches to working with your data.
- Adjust to GCP and cloud-computing paradigms.
Think in terms of specialized, ephemeral clusters.
- Use the smallest clusters you can—scope them to single jobs or small groups of closely related jobs.
- Create clusters each time you need them for a job and delete them when you're done.
Use GCP tools wherever appropriate.
Moving to an ephemeral model
The biggest shift in your approach between running an on-premises Hadoop workflow and running the same workflow on GCP is the shift away from monolithic, persistent clusters to specialized, ephemeral clusters. You spin up a cluster when you need to run a job and then delete it when the job completes. The resources required by your jobs are active only when they're being used, so you only pay for what you use. This approach enables you to tailor cluster configurations for individual jobs. Because you aren't maintaining and configuring a persistent cluster, you reduce the costs of resource use and cluster administration.
This section describes how to move your existing Hadoop infrastructure to an ephemeral model.
Separate data from computation
Using Cloud Storage as the persistent storage for your workflows has the following advantages:
- It's a Hadoop Compatible File System (HCFS), so it's easy to use with your existing jobs.
- It's faster than HDFS in many cases.
- It requires less maintenance than HDFS.
- It enables you to easily use your data with the whole range of GCP products.
- It's considerably less expensive than keeping your data in HDFS on a persistent Cloud Dataproc cluster.
With your data stored persistently in Cloud Storage, you can run your jobs on ephemeral Hadoop clusters managed by Cloud Dataproc.
In some cases, it might make more sense to move data to another GCP technology, such as BigQuery or Cloud Bigtable. But most general-purpose data should persist in Cloud Storage. More detail about these alternative storage options is provided later in this guide.
Run jobs on ephemeral clusters
Cloud Dataproc makes it easy to create and delete clusters so that you can move away from using one monolithic cluster to using many ephemeral clusters. This approach has several advantages:
- You can use different cluster configurations for individual jobs, eliminating the administrative burden of managing tools across jobs.
- You can scale clusters to suit individual jobs or groups of jobs.
- You only pay for resources when your jobs are using them.
- You don't need to maintain clusters over time, because they are freshly configured every time you use them.
- You don't need to maintain separate infrastructure for development, testing, and production. You can use the same definitions to create as many different versions of a cluster as you need when you need them.
You can use Cloud Dataproc initialization actions to define the configuration of nodes in a cluster. This makes it easy to maintain the different cluster configurations you need to closely support individual jobs and related groups of jobs. You can use the provided sample initialization actions to get started. The samples demonstrate how to make your own initialization actions.
Minimize the lifetime of ephemeral clusters
The point of ephemeral clusters is to use them only for the jobs' lifetime. When it's time to run a job, follow this process:
Create a properly configured cluster.
Run your job, sending output to Cloud Storage or another persistent location.
Delete the cluster.
Use your job output however you need to.
View logs in Stackdriver or Cloud Storage.
This process is shown in the following diagram:
Use small persistent clusters only when absolutely necessary
If you can't accomplish your work without a persistent cluster, you can create one. This option may be costly and isn't recommended if there is a way to get your job done on ephemeral clusters.
You can minimize the cost of a persistent cluster by:
- Creating the smallest cluster you can.
- Scoping your work on that cluster to the smallest possible number of jobs.
- Scaling the cluster to the minimum workable number of nodes, adding more dynamically to meet demand.
There are many advantages to migrating incrementally. You can:
- Isolate individual jobs in your existing Hadoop infrastructure from the complexity that's inherent in a mature environment.
- Examine each job in isolation to evaluate its needs and to determine the best path for migration.
- Handle unexpected problems as they arise without delaying dependent tasks.
- Create a proof of concept for each complex process without affecting your production environment.
- Move your workloads to the recommended ephemeral model thoughtfully and deliberately.
Your migration is unique to your Hadoop environment, so there is no universal plan that fits all migration scenarios. Make a plan for your migration that gives you the freedom to translate each piece to a cloud-computing paradigm.
Here's a typical incremental migration sequence:
Move some portion of your data to GCP.
Experiment with that data:
Replicate your existing jobs that use the data.
Make new prototypes that work with the data.
Repeat with additional data.
Start with the least critical data that you can. In the early stages, it's a good approach to use backup data and archives.
One kind of lower-risk job that makes a good starting test is backfilling by running burst processing on archive data. You can set up jobs that fill gaps in processing for data that existed before your current jobs were in place. Starting with burst jobs can often provide experience with scaling on GCP early in your migration plan. This can help you when you begin migrating more critical jobs.
The following diagram shows an example of a typical backfill hybrid architecture.
The example has two major components. First, scheduled jobs running on the on-premises cluster push data to Cloud Storage through an internet gateway. Second, backfill jobs run on ephemeral Cloud Dataproc clusters. In addition to backfilling, you can use ephemeral clusters in GCP for experimentation and creating proofs of concept for future work.
Planning with the completed migration in mind
So far, this guide has assumed that your goal is to move your entire Hadoop system from on-premises to GCP. A Hadoop system running entirely on GCP is easier to manage than one that operates both in the cloud and on-premises. However, a hybrid approach is often necessary to meet your business or technological needs.
Designing a hybrid solution
Here are some reasons you might need a hybrid solution:
- You are in the process of developing cloud-native systems, so existing systems that depend on them must continue to run on-premises until you are finished.
- You have business requirements to keep your data on-premises.
- You have to share data with other systems that run on-premises, and those systems can't interact with GCP due to technical or business restrictions.
A typical hybrid solution has four major parts:
An on-premises Hadoop cluster.
A connection from the on-premises cluster to GCP.
Centralized data storage in GCP.
Cloud-native components that work on data in GCP.
An issue you must address with a hybrid cloud solution is how to keep your systems in sync. That is, how will you make sure that changes you make to data in one place are reflected in the other? You can simplify synchronization by making clear distinctions between how your data is used in the different environments.
For example, you might have a hybrid solution where only your archived data is stored in GCP. You can set up scheduled jobs to move your data from the on-premises cluster to GCP when the data reaches a specified age. Then you can set up all of your jobs that work on the archived data in GCP so that you never need to synchronize changes to your on-premises clusters.
Another way to divide your system is to move all of the data and work for a specific project or working group to GCP while keeping other work on premises. Then you can focus on your jobs instead of creating complex data synchronization systems.
You might have security or logistical concerns that complicate how you connect your on-premises cluster to GCP. One solution is to use a Virtual Private Cloud connected to your on-premises network using Cloud VPN.
The following diagram shows an example hybrid cloud setup:
The example setup uses Cloud VPN to connect a GCP VPC to an on-premises cluster. The system uses Cloud Dataproc inside the VPC to manage persistent clusters that process data coming from the on-premises system. This might involve synchronizing data between the systems. Those persistent Cloud Dataproc clusters also transfer data coming from the on-premises system to the appropriate storage services in GCP. For the sake of illustration, the example uses Cloud Storage, BigQuery, and Cloud Bigtable for storage—those are the most common destinations for data processed by Hadoop workloads in GCP.
The other half of the example solution shows multiple ephemeral clusters that are created as needed in the public cloud. Those clusters can be used for many tasks, including those that collect and transform new data. The results of these jobs are stored in the same storage services used by the clusters running in the VPC.
Designing a cloud-native solution
By contrast, a cloud-native solution is straightforward. Because you run all of your jobs on GCP using data stored in Cloud Storage, you avoid the complexity of data synchronization altogether, although you must still be careful about how your different jobs use the same data.
The following diagram shows an example of a cloud-native system:
The example system has some persistent clusters and some ephemeral ones. Both kinds of clusters share cloud tools and resources, including storage and monitoring. Cloud Dataproc uses standardized machine images to define software configurations on VMs in a cluster. You can use these predefined images as the basis for the VM configuration you need. The example shows most of the persistent clusters running on version 1.1, with one cluster running on version 1.2. You can create new clusters with customized VM instances whenever you need them. This lets you isolate testing and development environments from critical production data and jobs.
Working with GCP services
This section discusses some additional considerations for migrating Hadoop to GCP.
Replacing open source tools with GCP services
GCP offers many products that you can use with your Hadoop system. Using a GCP product can often have benefits over running the equivalent open source product in GCP. Learn about GCP products and services to discover the breadth of what the platform has to offer.
Using regions and zones
You should understand the repercussions of geography and regions before you configure your data and jobs. Many GCP services require you to specify regions or zones in which to allocate resources. The latency of requests can increase when the requests are made from a different region than the one where the resources are stored. Additionally, if the service's resources and your persistent data are located in different regions, some calls to GCP services might copy all of the required data from one zone to another before processing. This can have a severe impact on performance.
Configuring authentication and permissions
Your control over permissions in GCP services is likely to be less fine-grained than you are accustomed to in your on-premises Hadoop environment. Make sure you understand how access management works in GCP before you begin your migration.
Cloud Identity and Access Management (Cloud IAM) manages access to cloud resources. It works on the basis of accounts and roles. Accounts identify a user or request (authentication), and the roles granted to an account dictate the level of access (authorization). Most GCP services provide their own set of roles to help you fine-tune permissions. As part of the planning process for your migration, you should learn about how Cloud IAM interacts with Cloud Storage and with Cloud Dataproc. Learn about the permissions models of each additional GCP service as you add it to your system, and consider how to define roles that work across the services that you use.
Monitoring jobs with Stackdriver Logging
Your GCP jobs send their logs to Logging, where the logs are easily accessible. You can get your logs in these ways:
- With a browser-based graphical interface using Logs Viewer in Google Cloud Platform Console.
- From a local terminal window using the gcloud command-line tool.
- From scripts or applications using the Cloud Client Libraries for the Stackdriver Logging API.
- By calling the Logging REST API.
Managing your edge nodes with Compute Engine
You can use Compute Engine to access an edge node in a Cloud Dataproc Hadoop cluster. As with most GCP products, you have multiple options for management: through the web-based console, from the command line, and through web APIs.
Using GCP big data services
Cloud Storage is the primary way to store unstructured data in GCP, but it isn't the only storage option. Some of your data might be better suited to storage in products designed explicitly for big data.
You can use Cloud Bigtable to store large amounts of sparse data. Cloud Bigtable is an HBase-compliant API that offers low latency and high scalability to adapt to your jobs.
For data warehousing, you can use BigQuery.