Installation on GCP
The diagram below depicts the target topology for GCP.

Requirements for GCP:
Two GCS storage buckets for gazette and druid data.
A GKE cluster with a minimum of two node pools: base pool and druid pool.
The base node pool should be labeled with
arize=true
andarize-base=true
.The druid node pool should be labeled with
arize=true
anddruid-historical=true
.A GCP service account attached to a role with the following permissions:
bigquery.jobs.create
storage.objects.create
storage.objects.delete
storage.objects.get
storage.objects.list
artifactregistry.repositories.downloadArtifacts
aiplatform.endpoints.predict
If using Workload Identity (default):
The service account must grant permissions to these GKE namespace/service-account pairs with role
Workload Identity User
:arize/arize
arize-operator/arize-operator
arize-spark/spark
If not using Workload Identity:
A JSON key from the GCP service account is required
Storage classes
premium-rwo
andstandard-rwo
are preferred and used by default.A GCR or docker registry is optional as Arize pulls images from Arize's central image registry by default.
Namespaces
arize
,arize-operator
, andarize-spark
can be pre-existing or created later by the helm chart.If using workload identity, the GCP service account must have role bindings to
<namespace>/<k8-service-account>
pairs.If not using workload identity, a JSON key from the service account is required.
Contact Arize for the <sizing option> field. This field controls the size of the deployment and must align
with the size of the cluster. Common values used would be small1b
or medium2b
.
values.yaml:
Choose an approach based on the deployment, for helm:
Last updated
Was this helpful?