Appendix: GCP deployment guidelines
Recommended instance types
As a general guideline, we recommend:
-
Processor Type: ARM-based CPU
-
Sizing: 2:1 disk-to-RAM ratio with spill-to-disk enabled.
When operating on GCP in production, we recommend the following machine types that support local SSD attachment:
| Series | Examples |
|---|---|
| N2 high-memory series | n2-highmem-16 or n2-highmem-32 with local NVMe SSDs |
| N2D high-memory series | n2d-highmem-16 or n2d-highmem-32 with local NVMe SSDs |
To maintain the recommended 2:1 disk-to-RAM ratio for your machine type, see
Number of local SSDs to determine the number of local
SSDs
(disk_support_config.local_ssd_count)
to use.
See also Locally attached NVMe storage.
Number of local SSDs
Each local NVMe SSD in GCP provides 375GB of storage. Use the appropriate number
of local SSDs
(disk_support_config.local_ssd_count)
to ensure your total disk space is at least twice the amount of RAM in your
machine type for optimal Materialize performance.
n2d-highmem-32 allows only the following number of local
SSDs: 4,8,16, or 24. To determine the valid number of Local SSDs to attach for your machine type, see the GCP
documentation.
For example, the following table provides a minimum local SSD count to ensure the 2:1 disk-to-RAM ratio. Your actual count will depend on the your machine type.
| Machine Type | RAM | Required Disk | Minimum Local SSD Count | Total SSD Storage |
|---|---|---|---|---|
n2-highmem-8 |
64GB |
128GB |
1 | 375GB |
n2-highmem-16 |
128GB |
256GB |
1 | 375GB |
n2-highmem-32 |
256GB |
512GB |
2 | 750GB |
n2-highmem-64 |
512GB |
1024GB |
3 | 1125GB |
n2-highmem-80 |
640GB |
1280GB |
4 | 1500GB |
Locally-attached NVMe storage
For optimal performance, Materialize requires fast, locally-attached NVMe storage. Having a locally-attached storage allows Materialize to spill to disk when operating on datasets larger than main memory as well as allows for a more graceful degradation rather than OOMing. Network-attached storage (like EBS volumes) can significantly degrade performance and is not supported.
Swap support
Starting in v0.6.1 of Materialize on Google Cloud PRovider (GCP) Terraform, disk support (using swap on NVMe instance storage) may be enabled for Materialize. With this change, the Terraform:
- Creates a node group for Materialize.
- Configures NVMe instance store volumes as swap using a daemonset.
- Enables swap at the Kubelet.
For swap support, the following configuration options are available:
See Upgrade Notes.
CPU affinity
It is strongly recommended to enable the Kubernetes static CPU management policy.
This ensures that each worker thread of Materialize is given exclusively access to a vCPU. Our benchmarks have shown this
to substantially improve the performance of compute-bound workloads.
TLS
When running with TLS in production, run with certificates from an official Certificate Authority (CA) rather than self-signed certificates.
Storage bucket versioning
Starting in v0.3.1 of Materialize on GCP Terraform, storage bucket versioning is
disabled (i.e.,
storage_bucket_versioning
is set to false by default) to facilitate cleanup of resources during testing.
When running in production, versioning should be turned on with a sufficient TTL
(storage_bucket_version_ttl)
to meet any data-recovery requirements.