Category
Compute
1. Introduction
What this service is
Hyperdisk is Google Cloud’s next-generation block storage for Compute Engine virtual machines (and services that use Compute Engine persistent disks). It is designed to deliver high performance, tunable performance, and predictable scalability compared with traditional persistent disk offerings.
Simple explanation (one paragraph)
If you need a disk for a VM and you want to control performance (IOPS and throughput) more directly—often without having to overbuy capacity just to get speed—Hyperdisk is a good fit. You create a disk, attach it to a Compute Engine VM, format and mount it, and then your applications use it like any other block device.
Technical explanation (one paragraph)
Hyperdisk is a family of Compute Engine disk types that let you provision performance characteristics (notably IOPS and throughput) more explicitly than older disk types. Hyperdisk is used as a boot or data disk for Compute Engine instances, supports common storage operations (attach/detach, snapshots, resizing), integrates with Google Cloud IAM and audit logging, and is billed using capacity plus (for some types) provisioned performance dimensions. Exact capabilities can vary by Hyperdisk type and region—always validate your chosen type against the official documentation.
What problem it solves
Traditional block storage often ties performance to disk size (or requires very large disks to achieve certain throughput/IOPS). Hyperdisk targets problems like:
- Needing high IOPS or high throughput without allocating unnecessary storage capacity.
- Needing to tune performance for changing workloads (databases, analytics pipelines, ML training) with fewer redesigns.
- Needing enterprise-grade operations (snapshots, encryption, IAM, monitoring) in a managed, cloud-native way.
Service name check: Hyperdisk is the current Google Cloud product name for this disk family (not a renamed legacy term). Confirm the latest disk-type lineup and regional availability in the official docs: https://cloud.google.com/compute/docs/disks/hyperdisks
2. What is Hyperdisk?
Official purpose
Hyperdisk is Google Cloud’s high-performance persistent block storage offering for Compute Engine, providing disk types engineered for demanding workloads and performance tuning.
Core capabilities
Common, currently documented capabilities include (verify exact support per disk type/region):
- Block storage for Compute Engine instances (boot or data volumes).
- Provisioned performance characteristics (IOPS and/or throughput) depending on disk type.
- Online operations typical of persistent disks (attach/detach, resize) subject to instance and disk constraints.
- Snapshots and images integration via Compute Engine disk snapshot features (capability depends on disk type and workflow; verify).
- Encryption by default with Google-managed keys; support for Customer-Managed Encryption Keys (CMEK) for disks in many cases (verify for your Hyperdisk type and region).
Major components
- Hyperdisk volume (disk): The block device resource you create in a project and zone/region, then attach to one or more VMs (subject to attachment mode and limitations).
- Compute Engine VM: The consumer of the disk via SCSI/NVMe presentation (exact interface details depend on machine series and configuration—verify).
- Snapshots / Images (Compute Engine): Data protection and cloning workflows.
- IAM & Audit Logs: Access control and tracking for disk operations.
- Cloud Monitoring & Cloud Logging: Observability for VM and disk-related metrics/logs.
Hyperdisk “storage pools” exist as a related concept in Compute Engine (used for capacity/performance management). Whether you use pools depends on your workload and operational model. Validate current documentation for Hyperdisk storage pools if you plan to use them.
Service type
- Managed block storage (Compute Engine persistent disk family).
- API-driven resource managed at the project level.
- Typically zonal (many disk resources are zonal) and in some cases regional replication is available for certain disk types and configurations—verify in official docs for Hyperdisk specifically.
Scope (zonal/regional/project)
- Project-scoped resources that are created in a zone (zonal disks) or region (regional disks), depending on the disk type and configuration.
- The VM that attaches the disk must be compatible with the disk’s location and attachment rules.
How it fits into the Google Cloud ecosystem
Hyperdisk is part of the Google Cloud Compute storage stack:
- Compute Engine VM storage options include:
- Persistent disks (including Hyperdisk)
- Local SSD
- Network file storage (Filestore)
- Object storage (Cloud Storage)
- Hyperdisk is typically chosen when you want managed block storage with predictable performance and cloud-native operations.
Official entry points:
– Hyperdisk docs: https://cloud.google.com/compute/docs/disks/hyperdisks
– Disks overview: https://cloud.google.com/compute/docs/disks
3. Why use Hyperdisk?
Business reasons
- Right-size cost vs performance: For performance-sensitive workloads, Hyperdisk can reduce the tendency to overprovision disk capacity just to meet performance targets.
- Faster time-to-production: Standardized disk types and provisioning patterns can simplify infrastructure templates.
- Supports growth: Easier scaling paths when performance needs increase over time.
Technical reasons
- Performance tuning: Hyperdisk types are designed for high IOPS and/or throughput use cases with explicit performance provisioning (varies by type).
- Predictability: Better alignment between declared performance requirements and achieved performance (subject to VM, filesystem, application pattern, and quotas).
- Integration: Works with Compute Engine snapshots, images, IAM, monitoring, and common Linux tooling.
Operational reasons
- Automation-friendly: Manage via gcloud, API, Terraform, and instance templates.
- Standard lifecycle operations: Create, attach, resize, snapshot, and delete in controlled workflows.
- Observability: Combine Compute Engine metrics with application metrics to track I/O saturation and plan changes.
Security / compliance reasons
- Encryption at rest by default (Google-managed keys).
- CMEK support (where available) to align with key custody and regulatory policies.
- Auditability: Admin operations are visible in Cloud Audit Logs.
Scalability / performance reasons
- Supports higher performance ceilings than baseline disk types (depending on your VM, disk type, and region).
- Designed for workloads that need sustained I/O, not just burst.
When teams should choose Hyperdisk
Choose Hyperdisk when you need one or more of the following:
- A database or stateful service that needs higher IOPS and/or higher throughput than general-purpose disks.
- A workload where performance must be tuned independently from capacity (for example, 200 GB disk but very high IOPS).
- A standard platform disk type for multiple teams with consistent performance behavior.
When teams should not choose Hyperdisk
Consider alternatives when:
- Your workload is not I/O bound (CPU/memory bound). Paying for performance you don’t use is wasteful.
- You need shared file semantics (choose Filestore or another file service).
- You need object storage or data lake patterns (choose Cloud Storage).
- Your VM is ephemeral and you can use Local SSD with replication at the app layer for maximum performance (with different durability tradeoffs).
- Your region/zone or machine family doesn’t support the Hyperdisk type you want (verify in official docs).
4. Where is Hyperdisk used?
Industries
- Financial services (transaction systems, risk analytics)
- SaaS and enterprise software
- Gaming backends (state stores, match services)
- Media processing pipelines
- Healthcare and life sciences (controlled environments)
- Retail and e-commerce (catalog, session, checkout state)
- Manufacturing/IoT (time series, near-real-time analytics)
Team types
- Platform teams standardizing VM storage
- SRE/operations teams handling stateful reliability and performance
- DevOps engineers building infrastructure-as-code modules
- Database administrators migrating or scaling DBs on Compute Engine
- ML/DS engineering teams working with high-throughput training data pipelines
Workloads
- OLTP databases (e.g., PostgreSQL, MySQL) on Compute Engine
- NoSQL stores (self-managed)
- Search clusters (e.g., Elasticsearch/OpenSearch self-managed)
- Message queue persistence (self-managed)
- Analytics staging, ETL scratch, intermediate compute datasets
- High-performance application caches with persistence requirements
Architectures
- Single VM with attached disks
- HA pairs using regional replication (if supported for your disk type—verify)
- Instance groups for stateless tiers plus dedicated stateful nodes
- Multi-zone architectures with application-level replication
Real-world deployment contexts
- Production: persistent state, higher SLA expectations, snapshots, CMEK, monitoring, controlled IAM
- Dev/test: performance experiments, cost-controlled smaller disks, automated cleanup
5. Top Use Cases and Scenarios
Below are realistic Hyperdisk use cases. Exact fit depends on the Hyperdisk type (Balanced/Extreme/Throughput/ML, etc.) and regional availability—confirm in docs.
1) Performance-tuned database volume
- Problem: A database needs high IOPS but doesn’t need a multi-terabyte disk.
- Why Hyperdisk fits: Provision IOPS/throughput to match DB workload without inflating capacity (type-dependent).
- Scenario: A 300 GB PostgreSQL VM needs predictable read/write IOPS for peak traffic; Hyperdisk is tuned to the target.
2) High-throughput ETL scratch disk
- Problem: Data pipelines need sustained sequential throughput for temporary staging.
- Why Hyperdisk fits: Hyperdisk throughput-oriented types target sustained bandwidth.
- Scenario: Nightly ETL jobs process multi-GB files; throughput is tuned for the job window.
3) Search index storage
- Problem: Search nodes need balanced random reads and writes with consistent latency.
- Why Hyperdisk fits: Balanced performance profiles and scalable throughput.
- Scenario: A self-managed search cluster keeps shard data on Hyperdisk to stabilize tail latency.
4) Stateful microservice on Compute Engine
- Problem: A stateful service requires durability and straightforward VM attachment semantics.
- Why Hyperdisk fits: Managed persistent block storage with snapshot workflows.
- Scenario: A licensing service stores a small embedded DB on a Hyperdisk volume.
5) Boot disks for performance-sensitive VMs
- Problem: Fast boot and package load times matter for frequent rollout cycles.
- Why Hyperdisk fits: Hyperdisk can be used as a boot disk (verify per type and constraints).
- Scenario: A CI fleet uses faster boot disks for quicker VM readiness.
6) ML training data disk (Compute Engine-based training)
- Problem: Training jobs need predictable data read throughput to keep GPUs busy.
- Why Hyperdisk fits: Higher throughput and performance tuning (type-dependent; Hyperdisk ML exists—verify current usage patterns).
- Scenario: GPU VMs mount Hyperdisk volumes containing sharded training datasets.
7) Low-latency transactional service (self-managed)
- Problem: The service needs consistent disk latency under load spikes.
- Why Hyperdisk fits: Tunable performance and high ceilings help reduce I/O-induced latency.
- Scenario: A payment processing component runs on Compute Engine with tuned disk IOPS.
8) Migration target from older Persistent Disk types
- Problem: Existing PD types are hitting performance limits or require oversized volumes for performance.
- Why Hyperdisk fits: Move to a disk type with explicit performance provisioning.
- Scenario: A MySQL VM on pd-ssd is migrated to Hyperdisk to meet peak I/O without doubling disk size.
9) Multi-environment platform standard disk
- Problem: Platform teams want one “default disk type” for most stateful services.
- Why Hyperdisk fits: Consistent model for performance tuning; can simplify internal documentation.
- Scenario: A platform module provisions Hyperdisk Balanced by default with environment-specific performance presets.
10) Temporary high-performance volume for backfill/reindex
- Problem: Short-term operations require high disk performance for hours/days.
- Why Hyperdisk fits: Provision performance for the duration of the operation, then scale down (if supported for your type—verify).
- Scenario: A reindex job provisions higher throughput during the run, then reduces it after completion.
11) Log ingestion buffer (self-managed)
- Problem: High-volume log ingestion spikes can overwhelm storage.
- Why Hyperdisk fits: Throughput tuning supports sustained writes.
- Scenario: A self-managed ingestion pipeline uses a disk buffer before shipping to a data lake.
12) Build cache for large monorepo builds
- Problem: Builds are I/O heavy and benefit from fast local persistence.
- Why Hyperdisk fits: Better performance than baseline disks with managed durability.
- Scenario: A remote build runner stores cache artifacts on a high-IOPS Hyperdisk volume.
6. Core Features
Hyperdisk is a family; features and limits vary by disk type and region. Always verify: https://cloud.google.com/compute/docs/disks/hyperdisks
Feature 1: Hyperdisk disk types for different performance profiles
- What it does: Provides multiple Hyperdisk variants (for example, Balanced, Extreme, Throughput, ML—names and availability can change).
- Why it matters: Lets you pick a disk aligned to your workload pattern: mixed I/O, high IOPS, high throughput, or ML-oriented access patterns.
- Practical benefit: Better cost/performance fit than “one size fits all.”
- Caveats: Not all types are available in all regions/zones; verify quotas and compatibility.
Feature 2: Provisioned performance (IOPS and throughput) (type-dependent)
- What it does: Lets you explicitly provision IOPS and/or throughput, often independently from disk size.
- Why it matters: Avoids overprovisioning capacity just to achieve performance.
- Practical benefit: Easier performance planning: choose disk size for data, performance for workload.
- Caveats: Provisioned performance typically affects billing. Minimums/maximums apply and are enforced by API; confirm current ranges in docs.
Feature 3: Integration with Compute Engine VMs (attach/detach)
- What it does: Hyperdisk volumes attach to VMs similarly to other persistent disks.
- Why it matters: Operationally familiar: standard Linux/Windows disk workflows.
- Practical benefit: Use standard tooling (LVM, mdadm, filesystems) and common automation.
- Caveats: Attachment limits per VM apply. Some advanced attachment modes (e.g., multi-writer) are not universal—verify support.
Feature 4: Resizing (capacity changes)
- What it does: Increase disk size without recreating the disk (common persistent disk behavior).
- Why it matters: Supports growth without downtime (filesystem expansion still required).
- Practical benefit: Incremental expansion as data grows.
- Caveats: Shrinking disks is generally not supported in-place; plan for migrations for downsizing.
Feature 5: Snapshots for backup, cloning, and DR
- What it does: Create point-in-time snapshots of disk data using Compute Engine snapshot features.
- Why it matters: Backups, environment cloning, and recovery are standard requirements.
- Practical benefit: Automate backups and restore workflows.
- Caveats: Snapshot cost and performance characteristics vary; cross-region copy and retention policies affect cost. Verify best practices for app-consistent snapshots.
Feature 6: Encryption at rest and CMEK (where supported)
- What it does: Encrypts disk data at rest; optional customer-managed keys via Cloud KMS for many disk scenarios.
- Why it matters: Security and compliance controls.
- Practical benefit: Key rotation, separation of duties, and centralized key governance.
- Caveats: CMEK introduces operational dependencies on KMS availability and permissions. Confirm Hyperdisk CMEK support for your chosen type.
Feature 7: Observability via Cloud Monitoring metrics (indirect)
- What it does: VM/disk I/O metrics help you understand throughput, IOPS, latency, and saturation patterns.
- Why it matters: Without metrics, teams overprovision or misdiagnose performance issues.
- Practical benefit: Data-driven tuning of provisioned performance.
- Caveats: Some disk-level metrics are surfaced at instance level. Validate which metrics apply to Hyperdisk in your environment.
Feature 8: Automation via API/CLI/IaC
- What it does: Manage disks with gcloud, REST, Terraform, and deployment pipelines.
- Why it matters: Repeatability and governance.
- Practical benefit: Standard modules, policy-as-code, consistent tags/labels.
- Caveats: Keep tooling versions current; new disk types may require recent provider/plugin versions.
Feature 9: Compatibility with Google Kubernetes Engine (via CSI) (verify)
- What it does: Use Hyperdisk-backed persistent volumes for GKE workloads (where supported).
- Why it matters: Many stateful workloads run on Kubernetes.
- Practical benefit: Standard PV/PVC workflows with tuned block storage.
- Caveats: Confirm GKE versions, CSI driver capabilities, and supported Hyperdisk types in official GKE docs.
7. Architecture and How It Works
High-level architecture
Hyperdisk is a Compute Engine disk resource. You create the disk in a project and zone/region, attach it to a VM, and the guest OS reads/writes blocks through the VM’s I/O stack. Control-plane actions (create, attach, snapshot) are handled by Google Cloud APIs; data-plane I/O flows between the VM and the storage backend over Google’s internal network.
Control flow vs data flow
- Control plane (management):
- You call the Compute Engine API (via Console, gcloud, Terraform).
- IAM authorizes the request.
- The disk resource is created/updated/attached/snapshotted.
- Operations are recorded in Cloud Audit Logs.
- Data plane (reads/writes):
- The VM issues I/O to the attached block device.
- Performance depends on the Hyperdisk type, provisioned settings, VM limits, and workload pattern.
- You observe results via application metrics and Cloud Monitoring.
Integrations with related services
- Compute Engine: Primary consumer (VMs).
- Cloud IAM: Permissions for disk operations.
- Cloud KMS: CMEK for disks (where supported).
- Cloud Monitoring: I/O-related metrics (often instance-level).
- Cloud Logging / Audit Logs: Who did what (disk create/delete/attach).
- Backup and DR tooling: Often built on snapshots (native or third-party marketplace tools).
- GKE (verify support): Persistent disks provisioned via CSI driver.
Dependency services
- Compute Engine API
- (Optional) Cloud KMS (if using CMEK)
- Cloud Logging / Monitoring (for ops)
Security / authentication model
- API requests are authenticated via Google Cloud identity (user accounts, service accounts).
- Authorized via IAM roles (least privilege recommended).
- For CMEK: service accounts need permissions to use KMS keys.
Networking model
- Hyperdisk is accessed over Google’s internal infrastructure; you don’t place disks in a VPC subnet.
- VPC rules don’t “firewall” disk access directly; access is controlled by who can attach disks to instances and who can log in to the VM.
Monitoring / logging / governance
- Use Cloud Audit Logs for administrative operations.
- Use Cloud Monitoring dashboards for instance disk I/O and saturation.
- Use labels on disk resources for cost allocation (team, env, app, data-classification).
Simple architecture diagram (Mermaid)
flowchart LR
user[Engineer / CI] -->|gcloud / Console| api[Compute Engine API]
api --> disk[Hyperdisk Volume]
api --> vm[Compute Engine VM]
vm -->|attach| disk
vm --> app[Application + Filesystem]
Production-style architecture diagram (Mermaid)
flowchart TB
subgraph Project["Google Cloud Project"]
subgraph ZoneA["Zone A"]
vm1[Compute Engine VM (primary)]
hd1[(Hyperdisk volume)]
vm1 -->|read/write| hd1
end
subgraph ZoneB["Zone B"]
vm2[Compute Engine VM (standby/secondary)]
hd2[(Replica / second disk\n(if using regional replication - verify))]
vm2 -->|read/write (failover)| hd2
end
snap[Snapshots (scheduled)]
kms[Cloud KMS (CMEK optional)]
mon[Cloud Monitoring]
audit[Cloud Audit Logs]
end
vm1 -->|metrics| mon
vm2 -->|metrics| mon
vm1 -->|admin events| audit
vm2 -->|admin events| audit
hd1 -->|snapshot| snap
hd2 -->|snapshot| snap
kms -.->|encrypt/decrypt keys| hd1
kms -.->|encrypt/decrypt keys| hd2
Note: The “replica/second disk” depiction depends on whether regional Hyperdisk is supported for your chosen disk type and region. Confirm current replication options in official docs.
8. Prerequisites
Account / project requirements
- A Google Cloud account with a billing-enabled project.
- Compute Engine API enabled:
- https://console.cloud.google.com/apis/library/compute.googleapis.com
Permissions / IAM roles
Minimum roles depend on what you do in the lab:
- To create/manage disks and VMs (broad):
roles/compute.admin(powerful; not least privilege)- More scoped options (typical building blocks):
roles/compute.instanceAdmin.v1roles/compute.storageAdmin(for disks/snapshots)roles/iam.serviceAccountUser(if attaching service accounts to VMs)
IAM reference: – https://cloud.google.com/compute/docs/access/iam
Billing requirements
- Hyperdisk is a paid resource. There is no universal “free tier” for persistent disks. Any free credits/promotions are account-specific—verify in your billing account.
CLI / tools
- Google Cloud CLI installed and authenticated:
- https://cloud.google.com/sdk/docs/install
- Optional:
- Terraform (if doing IaC)
fiofor benchmarking inside the VM
Region availability
- Hyperdisk types are not necessarily available in every region/zone.
- Before committing to an architecture, verify availability in:
- Hyperdisk docs: https://cloud.google.com/compute/docs/disks/hyperdisks
- The Cloud Console disk creation UI (it will show allowed types)
- Quotas pages for Compute Engine in your region
Quotas / limits
- Persistent disk quotas (GB, disk count, IOPS/throughput, snapshots) vary by region and project.
- VM attachment limits apply.
- If you hit errors like “quota exceeded” or “resource not available,” check:
- IAM permissions
- Regional/zone availability
- Compute Engine quotas
Prerequisite services
- Compute Engine
- Cloud Logging and Cloud Monitoring are enabled by default in most projects (verify organizational policies).
9. Pricing / Cost
Hyperdisk pricing is usage-based and varies by region and disk type. Do not rely on blog posts for exact numbers—use official sources.
Official pricing sources
- Compute disk pricing (official): https://cloud.google.com/compute/disks-pricing
- Google Cloud Pricing Calculator: https://cloud.google.com/products/calculator
Pricing dimensions (typical model—verify for your disk type)
Depending on the Hyperdisk type, you are commonly billed for:
- Provisioned capacity (GB-month)
- Provisioned IOPS (IOPS-month) (type-dependent)
- Provisioned throughput (MB/s-month or similar) (type-dependent)
- Snapshots: – Snapshot storage (GB-month) – Potential charges for snapshot operations or retrieval patterns (verify)
- Images (if you create custom images from disks)
- Operations/indirect costs: – VM costs (CPU/RAM) often dominate total cost – Backups/retention (snapshot growth) – Monitoring/logging ingestion (usually small, but can grow)
Free tier
- Persistent disks typically do not have a general free tier large enough for production use. Any always-free allowances are limited and may not apply to Hyperdisk. Verify in current Google Cloud free tier documentation.
Key cost drivers
- Provisioned performance: If you provision high IOPS/throughput “just in case,” you pay for it.
- Over-retention of snapshots: Long retention and frequent snapshots can accumulate.
- Environment sprawl: Multiple dev/test environments with large disks.
- High availability replication: If you use replicated/regional disks (where supported), you pay for replicated capacity/performance.
Hidden or indirect costs to watch
- Snapshot sprawl from automated pipelines without lifecycle policies.
- Performance misconfiguration: paying for performance above what the VM can actually use (VM-side bottleneck).
- Data egress / cross-region copies: Snapshot copy to another region or exporting images can incur network/storage charges (verify your workflow).
Network / data transfer implications
- Disk I/O itself is internal to Google’s infrastructure and is not usually billed as “network egress” like internet traffic.
- Cross-region data movement (e.g., snapshot copy across regions, DR replication patterns) can introduce additional costs—verify the specific SKU.
Cost optimization strategies
- Start with measured requirements: Use metrics to estimate required IOPS/throughput.
- Match VM and disk: Ensure the VM machine type supports the disk performance you provision.
- Use labels on disks and snapshots for cost allocation.
- Automate snapshot lifecycle: retention policies and deletion of orphaned snapshots.
- Scale performance only when needed: If your disk type supports performance changes after creation, adjust for peak windows and scale down after (verify supported operations and billing granularity).
Example low-cost starter estimate (conceptual)
A minimal lab generally includes: – 1 small VM (e2 or similar) – 1 Hyperdisk with small capacity – Optional: a snapshot
Because pricing varies, the right approach is: 1. Choose a region. 2. Use the pricing page and calculator to estimate: – Disk GB-month – Provisioned IOPS/throughput (if applicable) 3. Keep the lab short (minutes to a few hours) and delete resources.
Example production cost considerations (conceptual)
For production, estimate: – Disk capacity (including growth) – Provisioned IOPS/throughput for peak and steady-state – Snapshot cadence and retention – Replication/DR strategy (regional disks and/or cross-region backups) – Operational overhead: monitoring, logs, and backup tooling
10. Step-by-Step Hands-On Tutorial
This lab creates a Compute Engine VM, provisions a Hyperdisk volume, attaches it, formats and mounts it, then runs a quick benchmark and verifies metrics. It is designed to be beginner-friendly and low-risk, but Hyperdisk is a billed resource—clean up at the end.
Objective
- Create a Hyperdisk volume in Google Cloud
- Attach it to a Compute Engine VM
- Format and mount it on Linux
- Run a simple I/O test
- Validate the disk is working
- Clean up all resources
Lab Overview
You will:
1. Configure gcloud and choose a zone
2. Create a VM
3. Create a Hyperdisk volume (Hyperdisk Balanced in this example)
4. Attach the disk to the VM
5. Format and mount the disk
6. Run fio to confirm I/O works
7. Validate via lsblk, df -h, and basic performance output
8. Clean up
Important: Hyperdisk types and flags can vary over time. If any command fails due to unsupported flags or unavailable disk types in your zone, use the Console UI to confirm supported options, or check the latest docs: – Hyperdisk docs: https://cloud.google.com/compute/docs/disks/hyperdisks
– gcloud reference: https://cloud.google.com/sdk/gcloud/reference/compute/disks/create
Step 1: Set project, region, and zone
1) Authenticate and select a project:
gcloud auth login
gcloud config set project YOUR_PROJECT_ID
2) Choose a zone that supports Hyperdisk in your project. Example:
gcloud config set compute/region us-central1
gcloud config set compute/zone us-central1-a
Expected outcome: gcloud commands default to your selected project/zone.
Verification:
gcloud config list
Step 2: Enable the Compute Engine API
gcloud services enable compute.googleapis.com
Expected outcome: API enabled successfully.
Common error: Permission denied.
Fix: Ensure you have serviceusage.services.enable permission (e.g., Project Owner or Service Usage Admin).
Step 3: Create a small Linux VM
Create a low-cost VM suitable for testing:
gcloud compute instances create hd-lab-vm \
--machine-type=e2-standard-2 \
--image-family=debian-12 \
--image-project=debian-cloud \
--boot-disk-size=20GB
Expected outcome: VM hd-lab-vm is created.
Verification:
gcloud compute instances describe hd-lab-vm --format="get(status,zone)"
Step 4: Create a Hyperdisk volume
Create a Hyperdisk disk in the same zone.
Notes: – Disk type names are specific. Common examples include
hyperdisk-balanced,hyperdisk-extreme,hyperdisk-throughput,hyperdisk-ml(availability varies). – Some Hyperdisk types allow/require provisioning IOPS and throughput explicitly. – If the command fails, verify the exact supported type and flags in your zone using the Cloud Console create-disk flow or the current docs.
Example (Hyperdisk Balanced):
gcloud compute disks create hd-lab-disk \
--type=hyperdisk-balanced \
--size=100GB
Optional: If you intend to provision performance explicitly (type-dependent), the flags may look like:
# Verify supported flags and units in official docs before using.
# Example only; adjust values to allowed ranges for your type/region.
gcloud compute disks create hd-lab-disk-pp \
--type=hyperdisk-balanced \
--size=100GB \
--provisioned-iops=3000 \
--provisioned-throughput=150
Expected outcome: Disk is created.
Verification:
gcloud compute disks describe hd-lab-disk --format="yaml(name,type,sizeGb,zone)"
If you see an error like “Unknown disk type”:
– The zone may not support that Hyperdisk type
– The type string may differ
– Your org policy may restrict disk types
Use the Console to confirm supported types for that zone, or check: https://cloud.google.com/compute/docs/disks/hyperdisks
Step 5: Attach the Hyperdisk to the VM
gcloud compute instances attach-disk hd-lab-vm \
--disk=hd-lab-disk
Expected outcome: Disk attached successfully.
Verification:
gcloud compute instances describe hd-lab-vm \
--format="value(disks[].source)"
Step 6: SSH into the VM and identify the new disk
gcloud compute ssh hd-lab-vm
Inside the VM:
lsblk
Expected outcome: You see an unformatted disk (often sdb or sdc) with ~100G size.
If you’re unsure which device is the new disk:
sudo dmesg | tail -n 50
Step 7: Format the disk and mount it
Replace /dev/sdb with your actual device name.
1) Create an ext4 filesystem:
sudo mkfs.ext4 -F /dev/sdb
2) Create a mount point:
sudo mkdir -p /mnt/hyperdisk
3) Mount it:
sudo mount /dev/sdb /mnt/hyperdisk
4) Verify:
df -h /mnt/hyperdisk
Expected outcome: /mnt/hyperdisk shows mounted space close to 100G.
(Optional but recommended) Persist mount in /etc/fstab
Get the UUID:
sudo blkid /dev/sdb
Add an fstab entry (use your UUID output):
sudo nano /etc/fstab
Example line:
UUID=YOUR_UUID_HERE /mnt/hyperdisk ext4 defaults,nofail 0 2
Test:
sudo umount /mnt/hyperdisk
sudo mount -a
df -h /mnt/hyperdisk
Step 8: Run a quick I/O test with fio
Install fio:
sudo apt-get update
sudo apt-get install -y fio
Run a small test (keep it short to limit cost and time):
sudo fio --name=hdtest \
--filename=/mnt/hyperdisk/testfile \
--size=2G \
--rw=randread \
--bs=4k \
--iodepth=16 \
--numjobs=1 \
--time_based \
--runtime=30 \
--direct=1
Expected outcome: fio outputs IOPS and latency numbers. Exact results vary by VM, disk type, and provisioned performance.
Clean up the test file:
sudo rm -f /mnt/hyperdisk/testfile
Exit SSH:
exit
Validation
Validate disk attachment and mount (outside and inside VM)
Outside VM:
gcloud compute instances describe hd-lab-vm --format="table(disks.deviceName,disks.source)"
Inside VM (SSH):
lsblk
df -h | grep hyperdisk
Validate you are using Hyperdisk
From your workstation:
gcloud compute disks describe hd-lab-disk --format="get(type)"
You should see a Hyperdisk type string (for example, .../diskTypes/hyperdisk-balanced).
Troubleshooting
Issue: “Unknown disk type hyperdisk-balanced” – Cause: The zone doesn’t support that type, or the type name differs. – Fix: – Check Hyperdisk docs: https://cloud.google.com/compute/docs/disks/hyperdisks – Use the Console “Create disk” flow in that zone to see valid disk types. – Try a different zone/region that supports Hyperdisk.
Issue: Attach-disk fails with permission denied
– Cause: Missing IAM permissions.
– Fix: Ensure roles like roles/compute.storageAdmin and roles/compute.instanceAdmin.v1 (or equivalent) are granted.
Issue: Disk not visible in VM
– Cause: Attach operation not complete, or OS didn’t rescan.
– Fix:
– Re-check attachment: gcloud compute instances describe ...
– Inside VM: sudo dmesg | tail
– As a last resort, reboot the VM (not ideal for production, OK for lab).
Issue: fio performance is lower than expected
– Common causes:
– VM limits (machine type bottleneck)
– Single job/iodepth not saturating the disk
– Filesystem overhead
– Provisioned IOPS/throughput too low (type-dependent)
– Fix:
– Increase numjobs and iodepth for testing
– Verify VM type supports the expected I/O rates
– Verify provisioned performance settings (if applicable)
Cleanup
Delete resources to stop billing.
1) Delete the VM:
gcloud compute instances delete hd-lab-vm --quiet
2) Delete the Hyperdisk disk:
gcloud compute disks delete hd-lab-disk --quiet
If you created an additional disk (like hd-lab-disk-pp), delete it too:
gcloud compute disks delete hd-lab-disk-pp --quiet
3) (Optional) Verify everything is gone:
gcloud compute instances list
gcloud compute disks list
11. Best Practices
Architecture best practices
- Choose the Hyperdisk type to match I/O pattern
- Random IOPS-heavy (databases) vs sequential throughput-heavy (ETL).
- Design for failure
- Use snapshots, tested restores, and app-level replication where needed.
- If using regional replication features, validate RPO/RTO and actual failover steps (verify for your type).
- Separate boot and data disks
- Keep OS on a smaller boot disk; keep application data on dedicated data disks for easier tuning and migration.
IAM / security best practices
- Least privilege: Grant only the permissions required to manage disks and attach them to instances.
- Separate duties: Ops team manages disk resources; app team manages instance-level access where appropriate.
- Use CMEK if required: Apply Cloud KMS keys for regulated workloads (verify Hyperdisk CMEK support and plan for KMS permissions and availability).
Cost best practices
- Measure before provisioning performance: Use monitoring to determine needed IOPS/throughput.
- Avoid idle overprovisioning: Especially in dev/test—use smaller disks and delete after use.
- Snapshot hygiene: Implement lifecycle policies; delete unused snapshots; avoid retaining duplicates forever.
- Label everything:
env,team,app,cost-center,data-classification.
Performance best practices
- Match VM capabilities: VM type and interface can limit achievable disk performance.
- Use appropriate filesystem settings: For databases, follow vendor filesystem and mount options guidance (noatime, etc.), but test.
- Benchmark correctly: Use realistic block sizes, concurrency, and read/write mix.
- Watch queue depth: Many workloads need concurrency to reach target IOPS.
Reliability best practices
- Use app-consistent snapshot strategies:
- Quiesce or flush DBs when needed.
- Consider filesystem freeze for consistent backups (Linux
fsfreeze)—test and validate. - Regular restore tests: A backup not tested is a guess.
Operations best practices
- Use Infrastructure as Code for repeatability.
- Standard naming:
pd|hd-<app>-<env>-<zone>-<purpose>- Runbooks:
- Attach/detach procedures
- Snapshot restore
- Disk expansion steps
- Track quotas: Plan capacity and performance headroom; request quota increases early.
Governance best practices
- Org policies: Control external IPs, restrict disk CMEK usage patterns, enforce labels (where supported).
- Resource hierarchy: Separate prod vs non-prod projects for clean billing and access boundaries.
12. Security Considerations
Identity and access model
- Hyperdisk is controlled through Compute Engine IAM permissions.
- Key permissions include:
- Create/delete disks
- Attach/detach disks to instances
- Create/delete snapshots
- Use service accounts for automation and CI/CD; avoid long-lived user keys.
Encryption
- Encryption at rest is standard for Google Cloud storage services, including persistent disks.
- CMEK (Cloud KMS):
- Use when you need customer-controlled key management.
- Ensure the right identities have
cloudkms.cryptoKeyEncrypterDecrypter. - Plan for operational impact: if KMS access is blocked, disk operations can fail.
CMEK docs (persistent disks): – https://cloud.google.com/compute/docs/disks/customer-managed-encryption
Network exposure
- Disks aren’t accessed via VPC firewall rules; they are attached to VMs.
- The main network risk is VM compromise, leading to disk data exposure.
- Mitigations:
- OS hardening, patching, minimal packages
- Restrict SSH/RDP access, use IAP where appropriate
- Use Shielded VM features where relevant (for VM security posture)
Secrets handling
- Don’t store secrets in plaintext on disks.
- Use Secret Manager or KMS-encrypted secrets with application-level controls.
Audit / logging
- Cloud Audit Logs can show disk operations like create/delete/attach/snapshot.
- Route audit logs to centralized logging projects if required.
- Monitor for suspicious operations (unexpected snapshot creation, disk deletion attempts).
Audit logs overview: – https://cloud.google.com/logging/docs/audit
Compliance considerations
- Data residency: choose regions appropriately.
- CMEK and access logs are often required for regulated industries.
- Establish retention policies for snapshots and logs.
Common security mistakes
- Overly broad roles like
roles/ownerorroles/compute.adminused everywhere. - No separation between prod and dev/test projects.
- Unencrypted secrets on disks.
- No restore tests; no detection for unexpected snapshot exports.
Secure deployment recommendations
- Use least privilege IAM and separate projects/environments.
- Use CMEK where required, with documented key rotation.
- Use hardened base images and OS Login / IAP for access controls.
- Use labels to tag data classification and enforce policies with org tooling.
13. Limitations and Gotchas
These are common patterns; exact limits depend on Hyperdisk type, VM type, region, and evolving product constraints. Always verify current limits in the official docs.
Known limitations / constraints (verify)
- Regional availability: Some Hyperdisk types are only in certain regions/zones.
- VM attachment limits: Maximum number of disks per VM; maximum aggregate performance per VM.
- Performance limits: Provisioned IOPS/throughput have minimum and maximum values; VM can cap actual achieved performance.
- Boot disk support: Not all disk types may be allowed/ideal as boot disks; validate before standardizing.
- Multi-attach / multi-writer: If you require shared block access, verify if the specific Hyperdisk type supports it and under what constraints.
- Snapshots behavior: Snapshot speed and restore performance depend on many variables; plan and test.
Quotas
- Disk capacity quotas (GB)
- Disk count quotas
- Snapshot quotas
- Performance-related quotas (IOPS/throughput), if applicable
- API rate limits
Regional constraints
- Zonal vs regional disks: not all disk types have both forms.
- Cross-zone/region migrations require planned workflows.
Pricing surprises
- Provisioning high IOPS/throughput can cost more than expected.
- Snapshots retained for long periods can accumulate significant costs.
- DR copies across regions can add storage and transfer costs.
Compatibility issues
- Older guest OS images may require updated drivers or kernel versions for optimal performance.
- Some machine families may not deliver expected throughput; verify machine-specific limits.
Operational gotchas
- After increasing disk size, you still must grow the filesystem (and sometimes partition) inside the VM.
- If using CMEK, broken KMS permissions can block disk operations unexpectedly.
- Snapshot-based backups must consider application consistency (especially databases).
Migration challenges
- Moving between disk types may require:
- Snapshot/restore
- Image-based workflows
- rsync-level copy with downtime windows
- Benchmark before and after migration; tune performance settings.
14. Comparison with Alternatives
Hyperdisk sits within Google Cloud’s block storage lineup. Your main decision is usually between Hyperdisk and other persistent disk types, Local SSD, or a file/object service.
Comparison table
| Option | Best For | Strengths | Weaknesses | When to Choose |
|---|---|---|---|---|
| Hyperdisk (Compute Engine) | Tunable high-performance block storage for VMs | Provisioned performance model (type-dependent), strong integration with Compute Engine, snapshots, encryption | Availability/limits vary; misconfig can be costly; VM caps apply | You need predictable performance and tuning flexibility |
| Persistent Disk (pd-balanced / pd-ssd / pd-standard) | General VM storage | Widely available, simple sizing, familiar | Performance often tied to size or has lower ceilings | Most general workloads, cost-sensitive dev/test |
| Persistent Disk Extreme (pd-extreme) | Very high IOPS for specific workloads | High IOPS for select workloads | May require large minimums/constraints; availability limits | Legacy extreme-performance needs; compare with Hyperdisk Extreme |
| Local SSD | Ultra-low latency, highest IOPS on a single VM | Very high performance | Ephemeral; data loss on stop/terminate; requires replication | Caches, scratch, distributed storage systems |
| Filestore | Shared file storage (NFS) | Managed NFS, simple POSIX shared access | Not block storage; different performance/cost model | Shared home dirs, content repos, lift-and-shift file apps |
| Cloud Storage | Object storage, data lakes | Cheap at scale, durable, global tooling | Not POSIX; higher latency; different access semantics | Data lakes, backups, content distribution |
| AWS EBS (gp3/io2) | Similar VM block storage on AWS | Mature ecosystem; gp3 also decouples performance | Different cloud; migration effort | Multi-cloud parity or AWS-native workloads |
| Azure Managed Disks (Premium SSD v2/Ultra Disk) | Similar VM block storage on Azure | Explicit performance tiers, high performance options | Different cloud; migration effort | Azure-native workloads needing high-performance disks |
| Self-managed storage (Ceph, etc.) | Full control, custom features | Flexibility; multi-cloud | Operational burden; reliability complexity | Specialized needs and strong storage ops expertise |
15. Real-World Example
Enterprise example: regulated payments platform on Compute Engine
- Problem: A payments platform runs self-managed PostgreSQL on Compute Engine and needs predictable peak IOPS during business hours, with strict audit and encryption requirements.
- Proposed architecture:
- Compute Engine VMs (primary/standby across zones as per HA design)
- Hyperdisk volumes for data
- Scheduled snapshots with tested restore process
- CMEK via Cloud KMS (where supported)
- Centralized audit logging and monitoring dashboards
- Why Hyperdisk was chosen:
- Ability to tune performance without scaling capacity excessively (type-dependent).
- Integration with IAM, audit logs, and CMEK to satisfy compliance requirements.
- Expected outcomes:
- More predictable DB latency at peak.
- Clearer capacity/performance planning and cost allocation.
- Improved audit posture for disk operations.
Startup / small-team example: SaaS search + analytics pipeline
- Problem: A small SaaS team runs a search cluster and ETL jobs on Compute Engine and sees inconsistent performance when indexing and running analytics jobs at the same time.
- Proposed architecture:
- Separate VMs for indexing and ETL stages
- Hyperdisk Throughput (or suitable Hyperdisk type) for ETL staging volume
- Hyperdisk Balanced for search data volumes
- Snapshot policies for quick recovery
- Why Hyperdisk was chosen:
- Tune throughput for ETL and balanced I/O for search without oversizing disks.
- Simple operational model for a small team (managed disk, standard VM workflows).
- Expected outcomes:
- Fewer indexing slowdowns.
- Faster ETL completion in batch windows.
- More predictable cloud spend tied to explicit performance needs.
16. FAQ
1) Is Hyperdisk a separate product from Compute Engine Persistent Disk?
Hyperdisk is part of the Compute Engine disk ecosystem (persistent block storage). It’s best understood as a newer disk family/type within Compute Engine storage options.
2) What’s the main reason to use Hyperdisk?
To get better performance tuning (IOPS/throughput) and performance scaling characteristics compared to traditional disks—while keeping managed disk operations.
3) Which Hyperdisk type should I pick?
Pick based on your workload pattern: – Mixed I/O: typically “Balanced” – Very high IOPS: typically “Extreme” – Sustained bandwidth: typically “Throughput” – ML-oriented: “ML” type (if applicable) Always verify the current lineup and guidance: https://cloud.google.com/compute/docs/disks/hyperdisks
4) Can I use Hyperdisk as a boot disk?
Often yes for some types, but boot disk support and recommendations can vary. Verify in official docs for your specific Hyperdisk type and VM configuration.
5) Can I change Hyperdisk performance settings after creation?
Some Hyperdisk types are designed around adjustability. However, exact operations and constraints can change—verify in official docs for “update disk” workflows.
6) Is Hyperdisk zonal or regional?
Compute Engine disks can be zonal or regional depending on disk type. Hyperdisk availability for regional replication depends on the type and region—verify in the Hyperdisk documentation.
7) How do snapshots work with Hyperdisk?
Hyperdisk uses Compute Engine snapshot mechanisms. Use snapshots for backups and cloning; plan retention carefully to manage cost. Snapshot docs: https://cloud.google.com/compute/docs/disks/create-snapshots
8) Does Hyperdisk support CMEK?
Persistent disks often support CMEK via Cloud KMS, but support can vary by disk type/region. Verify CMEK support for your disk type: https://cloud.google.com/compute/docs/disks/customer-managed-encryption
9) What’s the difference between Hyperdisk and Local SSD?
Local SSD is ephemeral and extremely fast; Hyperdisk is persistent managed storage with snapshots and durability.
10) Why is my Hyperdisk slower than expected?
Common reasons: – VM instance limits cap throughput/IOPS – Workload doesn’t generate enough parallel I/O (low queue depth) – Filesystem or application configuration bottlenecks – Provisioned performance too low (type-dependent) Benchmark with realistic concurrency and validate VM limits.
11) How do I estimate Hyperdisk cost?
Use official pricing pages and the pricing calculator:
– https://cloud.google.com/compute/disks-pricing
– https://cloud.google.com/products/calculator
Model capacity plus any provisioned performance dimensions and snapshot retention.
12) Is Hyperdisk supported for GKE Persistent Volumes?
GKE uses the Compute Engine PD CSI driver; Hyperdisk support depends on GKE version/driver and Hyperdisk type. Verify in official GKE docs and test in a sandbox.
13) What filesystem should I use on Hyperdisk?
For Linux, ext4 and XFS are common. Choose based on workload requirements and operational familiarity. For databases, follow vendor guidance and test.
14) Can multiple VMs share the same Hyperdisk?
Some disk types and modes support multi-attach or multi-writer patterns, but not universally. Verify if your Hyperdisk type supports your intended access mode.
15) What’s the safest backup strategy?
- Regular snapshots with retention policy
- Periodic restore tests
- Application-consistent procedures for databases (quiesce/flush) For stronger DR, consider cross-region copies and app replication patterns (verify costs and RPO/RTO).
17. Top Online Resources to Learn Hyperdisk
| Resource Type | Name | Why It Is Useful |
|---|---|---|
| Official documentation | Hyperdisk documentation | Canonical feature set, disk types, limits, supported regions: https://cloud.google.com/compute/docs/disks/hyperdisks |
| Official documentation | Compute Engine disks overview | Broader context of disk types and operations: https://cloud.google.com/compute/docs/disks |
| Official pricing page | Compute disk pricing | Official SKUs and pricing dimensions: https://cloud.google.com/compute/disks-pricing |
| Pricing tool | Google Cloud Pricing Calculator | Region-specific estimates and scenario modeling: https://cloud.google.com/products/calculator |
| CLI reference | gcloud compute disks create |
Exact flags for disk type and provisioning: https://cloud.google.com/sdk/gcloud/reference/compute/disks/create |
| Official guide | Snapshots documentation | Backup/restore workflows: https://cloud.google.com/compute/docs/disks/create-snapshots |
| Official guide | CMEK for disks | Key management and encryption controls: https://cloud.google.com/compute/docs/disks/customer-managed-encryption |
| Official docs | IAM for Compute Engine | Roles and permissions design: https://cloud.google.com/compute/docs/access/iam |
| Observability docs | Cloud Monitoring | Build dashboards and alerts for VM I/O: https://cloud.google.com/monitoring/docs |
| Video (official) | Google Cloud Tech YouTube channel | Often includes storage/Compute Engine deep dives (verify relevant Hyperdisk content): https://www.youtube.com/@googlecloudtech |
18. Training and Certification Providers
| Institute | Suitable Audience | Likely Learning Focus | Mode | Website URL |
|---|---|---|---|---|
| DevOpsSchool.com | DevOps engineers, SREs, platform teams | DevOps tooling, cloud operations, CI/CD, infrastructure practices (verify course catalog for Hyperdisk/Google Cloud coverage) | check website | https://www.devopsschool.com/ |
| ScmGalaxy.com | Beginners to intermediate DevOps learners | SCM, DevOps fundamentals, automation practices | check website | https://www.scmgalaxy.com/ |
| CLoudOpsNow.in | Cloud ops practitioners | Cloud operations, reliability, cost basics | check website | https://www.cloudopsnow.in/ |
| SreSchool.com | SREs and operations engineers | SRE principles, monitoring, incident response | check website | https://www.sreschool.com/ |
| AiOpsSchool.com | Ops teams adopting AIOps | AIOps concepts, automation, operational analytics | check website | https://www.aiopsschool.com/ |
19. Top Trainers
| Platform/Site | Likely Specialization | Suitable Audience | Website URL |
|---|---|---|---|
| RajeshKumar.xyz | Cloud/DevOps training content (verify current offerings) | Beginners to intermediate engineers | https://www.rajeshkumar.xyz/ |
| devopstrainer.in | DevOps training programs (verify cloud-specific modules) | DevOps engineers, students | https://www.devopstrainer.in/ |
| devopsfreelancer.com | Freelance DevOps enablement (verify services) | Teams needing short-term guidance | https://www.devopsfreelancer.com/ |
| devopssupport.in | DevOps support and training resources (verify scope) | Ops teams and practitioners | https://www.devopssupport.in/ |
20. Top Consulting Companies
| Company Name | Likely Service Area | Where They May Help | Consulting Use Case Examples | Website URL |
|---|---|---|---|---|
| cotocus.com | Cloud/DevOps consulting (verify offerings) | Cloud migration, platform engineering, operations | Disk/storage design reviews; performance troubleshooting; IaC modules | https://cotocus.com/ |
| DevOpsSchool.com | DevOps consulting and training services | DevOps transformation, tooling, process | Implement IaC + CI/CD; operational runbooks; monitoring setup | https://www.devopsschool.com/ |
| DEVOPSCONSULTING.IN | DevOps consulting (verify offerings) | Cloud ops, automation, reliability practices | Cost optimization for VM + disk; governance/labeling strategy; backup runbooks | https://www.devopsconsulting.in/ |
21. Career and Learning Roadmap
What to learn before Hyperdisk
- Compute Engine fundamentals:
- VM lifecycle, zones/regions, instance templates
- Linux storage basics:
- Partitions, filesystems (ext4/XFS), mount points, fstab
- IAM fundamentals:
- Projects, roles, service accounts
- Observability basics:
- Metrics vs logs, dashboards, alerting
What to learn after Hyperdisk
- Advanced storage operations:
- Snapshot strategy design, restore drills, DR planning
- Performance engineering:
- fio methodology, queue depth, syscall patterns
- Infrastructure as Code:
- Terraform modules for disks, instances, labels, CMEK
- Kubernetes storage (if applicable):
- CSI concepts, PV/PVC, StatefulSets (verify Hyperdisk support for your GKE version)
Job roles that use it
- Cloud Engineer / Platform Engineer
- DevOps Engineer
- Site Reliability Engineer (SRE)
- Solutions Architect
- Systems/Infrastructure Engineer
- Database Administrator (DBA) on cloud infrastructure
Certification path (Google Cloud)
Google Cloud certifications change over time; common relevant ones include: – Associate Cloud Engineer – Professional Cloud Architect – Professional DevOps Engineer
Verify the current certification list: – https://cloud.google.com/learn/certification
Project ideas for practice
- Build a “stateful VM module”:
- VM + Hyperdisk + snapshot schedule + labels + CMEK option
- Create a performance test harness:
- Deploy VM, run fio profiles, collect metrics, store results
- Implement backup/restore drill:
- Snapshot nightly, restore to a new VM weekly, validate data integrity
- Cost governance:
- Label enforcement + budget alerts + automated cleanup for dev disks
22. Glossary
- Hyperdisk: Google Cloud Compute Engine next-generation persistent block storage disk family.
- Block storage: Storage presented as a raw block device to an OS (formatted with a filesystem or used directly by databases).
- Compute Engine: Google Cloud’s Infrastructure-as-a-Service VM offering.
- IOPS: Input/Output Operations Per Second; measures how many read/write operations happen per second.
- Throughput: Data transfer rate (e.g., MB/s) for sequential or aggregate I/O.
- Latency: Time per I/O operation; important for transactional workloads.
- Zonal resource: Exists in a single zone (e.g.,
us-central1-a). - Regional disk: A replicated disk across multiple zones in a region (availability depends on disk type—verify for Hyperdisk).
- Snapshot: Point-in-time backup of a disk.
- CMEK: Customer-Managed Encryption Keys, managed in Cloud KMS.
- Cloud KMS: Key Management Service for creating and controlling encryption keys.
- IAM: Identity and Access Management; controls who can do what in a project.
- Cloud Audit Logs: Logs for administrative actions taken in Google Cloud.
- fio: A flexible I/O workload generator for benchmarking storage.
23. Summary
Hyperdisk is Google Cloud’s Compute Engine persistent block storage family designed for tunable, high-performance workloads. It matters when you need more control over disk performance characteristics (like IOPS and throughput) and want to avoid overprovisioning storage capacity just to hit performance targets.
In Google Cloud architectures, Hyperdisk fits as the primary disk option for stateful VMs and performance-sensitive systems, integrating with IAM, snapshots, monitoring, and (where supported) CMEK via Cloud KMS. Cost management is largely about controlling provisioned performance, capacity growth, and snapshot retention, and ensuring the VM can actually use the performance you pay for.
Use Hyperdisk when you have measurable I/O requirements and want predictable scaling; choose simpler disk types or other storage services (Filestore/Cloud Storage/Local SSD) when the workload semantics or cost profile demand it.
Next step: read the official Hyperdisk documentation, then replicate the lab using your target region and disk type, and validate performance with real workload patterns: https://cloud.google.com/compute/docs/disks/hyperdisks