Pivotal Greenplum for Kubernetes 1.1.0 Release Notes
Pivotal Greenplum for Kubernetes 1.1.0 introduces new and changed features as compared to previous releases of the product.
Refer to the Pivotal Greenplum Database documentation for detailed information on Pivotal Greenplum Database.
Pivotal Greenplum for Kubernetes includes the Pivotal Greenplum Database version identified below:
|Greenplum for Kubernetes Version||Pivotal Greenplum Database Version|
Pivotal Greenplum for Kubernetes 1.1.0 is supported on the following platforms:
- Pivotal Container Service (PKS) 1.3.1 (contains Kubernetes 1.12.4)
- Google Kubernetes Engine (GKE) Kubernetes 1.12.4
Additional Kubernetes environments, such as Minikube, can be used for testing or demonstration purposes.
This release introduces the following new and changed features:
All Kubernetes resources related to the Greenplum Operator are now labeled using the
appkey. Previous releases used the
namekey. For example, in this release you could list all Greenplum Operator resources using the command:
kubectl get all -l app=greenplum-operator
New status and event messages were added to provide additional information about the lifecycle of a Greenplum for Kubernetes cluster during create, update, and delete operations. For example, the
kubectl deletecommand now sets the deletion status and events as shown in this example:
Status: Instance Image: greenplum-for-kubernetes:latest Operator Version: greenplum-operator:latest Phase: DeletingGreenplumCluster Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal CreatingGreenplumCluster 3m greenplumOperator Creating Greenplum cluster my-greenplum in default Normal CreatedGreenplumCluster 1m greenplumOperator Successfully created Greenplum cluster my-greenplum in default Normal DeletingGreenplumCluster 6s greenplumOperator Deleting Greenplum cluster my-greenplum in default
During cluster deletion, an error message is now logged if
gpstopfails, but the operator continues to delete Kubernetes resources.
Greenplum for Kubernetes cluster configurations are now validated via a Kubernetes admission webhook when you apply changes with applications such as
kubectl. Note that the following properties are validated and cannot be modified for an existing cluster instance:
If you redeploy an instance using updated
storageClassNameproperties, the request is now rejected if the corresponding persistent volume claims (PVCs) already exist. To redeploy using new values for these properties, you must first delete existing PVCs.
The upgrade process was improved to reject attempts to upgrade older, unsupported Greenplum clusters.
To upgrade your existing Greenplum for Kubernetes operator and clusters, follow the instructions in Upgrading Greenplum for Kubernetes.
- Some common Greenplum Database backup and restore utilities are not supported. This includes
gprestoreutilities, and the related plug-ins for s3 and Data Domain Boost.
- Greenplum Database connector functionality is not supported. This includes Greenplum Streaming Server (gpss) and its API, Greenplum-Informatica Connector, Greenplum-Spark Connector, and the Greenplum-Kafka Integration.
- Pivotal extension framework (PXF) and its associated plug-ins are not supported.
- Greenplum for Kubernetes does not currently support cluster monitoring with Greenplum Command Center, the
gpperfmondatabase, or related utilities. Use system-level monitoring tools such as Prometheus and Grafana until Greenplum Command Center support is available.
- Greenplum for Kubernetes does not support the built-in SNMP features that are available in Pivotal Greenplum Database.
- Greenplum for Kubernetes clusters do not support installing Greenplum extensions that use the
gppkgutility). Future releases will include these extensions as part of the distribution, as with MADlib.
- The Greenplum Operator does not yet support changing all attribute values of a deployed Greenplum cluster. See Greenplum Operator Manifest File for details.