Deploying PXF with Greenplum (Beta)

This section describes procedures for deploying a Greenplum for Kubernetes cluster that includes the Pivotal extension framework (PXF).

Note: Pivotal PXF in Greenplum for Kubernetes is a Beta feature.

About PXF on Greenplum for Kubernetes

When you deploy PXF to Greenplum for Kubernetes, the Greenplum Operator creates one or more dedicated pods, or replicas, to host the PXF server instances. This differs from Pivotal Greenplum deployed to other platforms, where a PXF server instance is deployed to each Greenplum segment host. With Greenplum for Kubernetes, you can choose to deploy as many PXF server replicas as needed to provide redundancy should a PXF pod fail and to distribute load.

When you install a new Greenplum cluster using the template PXF manifest file, workspace/samples/my-gp-with-pxf-instance.yaml, PXF is installed and initialized with a default (empty) PXF configuration directory. After deploying the cluster, you can customize the configuration by creating PXF server configurations for multiple data sources, using the instructions in the Pivotal Greenplum Documentation such as:

For subsequent or repeatable deployments, you can export your customized PXF configuration directory ($PXF_CONF) to an S3 bucket, and then describe the bucket-path in the Greenplum for Kubernetes deployment manifest. The specified S3 location is downloaded and used as the PXF_CONF directory when deploying the cluster.

Deploying PXF with the Default Configuration

Follow these steps to deploy PXF with the default, initialized configuration. You will need to modify the template PXF configuration files to access the external data sources required for your system.

  1. Use the procedure described in Deploying a New Greenplum Cluster to deploy the cluster, but use the samples/my-gp-with-pxf-instance.yaml as the basis for your deployment. Copy the file into your /workspace directory. For example:

    $ cd ./greenplum-for-kubernetes-*/workspace
    $ cp ./samples/my-gp-with-pxf-instance.yaml .
  2. Edit the file as necessary for your deployment. my-gp-with-pxf-instance.yaml includes properties to configure PXF in the basic Greenplum cluster:

    apiVersion: ""
    kind: "GreenplumCluster"
      name: my-greenplum
        hostBasedAuthentication: |
          # host   all   gpadmin   trust
          # host   all   gpuser   md5
        memory: "800Mi"
        cpu: "0.5"
        storageClassName: standard
        storage: 1G
        antiAffinity: "yes"
        workerSelector: {}
        primarySegmentCount: 1
        memory: "800Mi"
        cpu: "0.5"
        storageClassName: standard
        storage: 2G
        antiAffinity: "yes"
        workerSelector: {}
        serviceName: "my-greenplum-pxf"    
    apiVersion: ""
    kind: "GreenplumPXFService"
      name: my-greenplum-pxf
      replicas: 2
      cpu: "0.5"
      memory: "1Gi"
      workerSelector: {}
    #  pxfConf:
    #    s3Source:
    #      secret: "my-greenplum-pxf-configs"
    #      endpoint: ""
    #      bucket: "YOUR_S3_BUCKET_NAME"
    #      folder: "YOUR_S3_BUCKET_FOLDER-Optional"
    # Note: If using pxfConf.s3Source, in addition to applying the above yaml be sure to create a secret using a command similar to:
    # kubectl create secret generic my-greenplum-pxf-configs --from-literal=‘access_key_id=XXX’ --from-literal=‘secret_access_key=XXX’

    The entry:

        serviceName: "my-greenplum-pxf"    

    Indicates that the cluster will use the PXF service configuration named my-greenplum-pxf that follows at the end of the yaml file. The sample configuration creates two PXF replica pods for redundancy with minimal settings for CPU and memory. You can customize these values as needed, as well as the workerSelector value if you want to constrain the replica pods to labeled nodes in your cluster. See Greenplum PXF Service Properties for information about each available property.

    The commented properties can be used when you have created a PXF configuration directory that you want to reuse when deploying the cluster. See Exporting a PXF Configuration to S3 for more information.

  3. Use kubectl apply command with your modified PXF manifest file to send the deployment request to the Greenplum Operator. For example:

    $ kubectl apply -f ./my-gp-with-pxf-instance.yaml  created created

    If you are deploying another instance of a Greenplum cluster, specify the Kubernetes namespace where you want to deploy the new cluster. For example, if you previously deployed a cluster in the namespace gpinstance-1, you could deploy a second Greenplum cluster in the gpinstance-2 namespace using the command:

    $ kubectl apply -f ./my-gp-with-pxf-instance.yaml -n gpinstance-2

    The Greenplum Operator deploys the necessary Greenplum and PXF resources according to your specification, and also initializes the Greenplum cluster.

  4. Execute the following command to monitor the deployment of the cluster. While the cluster is initializing the status will be Pending:

    $ watch kubectl get all
    NAME                                      READY     STATUS    RESTARTS   AGE
    pod/greenplum-operator-79cddcf586-ctftb   1/1       Running   0          2m40s
    pod/master-0                              1/1       Running   0          23s
    pod/master-1                              1/1       Running   0          22s
    pod/my-greenplum-pxf-676fd6fdd7-825gq     0/1       Running   0          28s
    pod/my-greenplum-pxf-676fd6fdd7-mjt6w     0/1       Running   0          28s
    pod/segment-a-0                           1/1       Running   0          22s
    pod/segment-b-0                           1/1       Running   0          22s
    NAME                                                            TYPE           CLUSTER-IP       EXTERNAL-IP   PORT(S)         AGE
    service/agent                                                   ClusterIP      None             <none>        22/TCP          23s
    service/greenplum                                               LoadBalancer   <pending>     5432:32294/TCP  23s
    service/greenplum-validating-webhook-service-79cddcf586-ctftb   ClusterIP     <none>        443/TCP         2m38s
    service/kubernetes                                              ClusterIP        <none>        443/TCP         19m
    service/my-greenplum-pxf                                        ClusterIP   <none>        5888/TCP        29s
    NAME                                 READY     UP-TO-DATE   AVAILABLE   AGE
    deployment.apps/greenplum-operator   1/1       1            1           2m40s
    deployment.apps/my-greenplum-pxf     0/2       2            0           28s
    NAME                                            DESIRED   CURRENT   READY     AGE
    replicaset.apps/greenplum-operator-79cddcf586   1         1         1         2m40s
    replicaset.apps/my-greenplum-pxf-676fd6fdd7     2         2         0         28s
    NAME                         READY     AGE
    statefulset.apps/master      2/2       23s
    statefulset.apps/segment-a   1/1       23s
    statefulset.apps/segment-b   1/1       23s
    NAME                                                 STATUS    AGE   Pending   29s
    NAME                                                        AGE   29s

    Note that the Greenplum PXF service, deployment, and replicas are created in addition to the Greenplum cluster.

  5. Describe your Greenplum cluster to verify that it was created successfully. The Phase should eventually transition to Running:

    $ kubectl describe greenplumClusters/my-greenplum
    Name:         my-greenplum
    Namespace:    default
    Labels:       <none>
                    {"apiVersion":"","kind":"GreenplumCluster", "metadata":{"annotations":{},"name":"my-greenplum", "namespace":"default"...
    API Version:
    Kind:         GreenplumCluster
      Creation Timestamp:  2019-04-01T15:19:17Z
      Generation:          1
      Resource Version:    1469567
      Self Link:           /apis/
      UID:                 83e0bdfd-5491-11e9-a268-c28bb5ff3d1c
      Master And Standby:
        Anti Affinity:              yes
        Cpu:                        0.5
        Host Based Authentication:  # host   all   gpadmin   trust
    # host   all   gpuser   md5
        Memory:              800Mi
        Storage:             1G
        Storage Class Name:  standard
        Worker Selector:
        Anti Affinity:          yes
        Cpu:                    0.5
        Memory:                 800Mi
        Primary Segment Count:  1
        Storage:                2G
        Storage Class Name:     standard
        Worker Selector:
      Instance Image:    greenplum-for-kubernetes:latest
      Operator Version:  greenplum-operator:latest
      Phase:             Running
      Type    Reason                    Age   From               Message
      ----    ------                    ----  ----               -------
      Normal  CreatingGreenplumCluster  2m    greenplumOperator  Creating Greenplum cluster my-greenplum in default
      Normal  CreatedGreenplumCluster   8s    greenplumOperator  Successfully created Greenplum cluster my-greenplum in default

    If you are deploying a brand new cluster, the Greenplum Operator automatically initializes the Greenplum cluster. The Phase should eventually transition from Pending to Running and the Events should match the output above.

    Note: If you redeployed a previously-deployed Greenplum cluster, the phase will begin at Pending. The cluster uses its existing Persistent Volume Claims if they are available. In this case, the master and segment data directories will already exist in their former state. The master-0 pod automatically starts the Greenplum Cluster, after which the phase transitions to Running.

  6. At this point, you can work with the deployed Greenplum cluster by executing Greenplum utilities from within Kubernetes, or by using a locally-installed tool, such as psql, to access the Greenplum instance running in Kubernetes. To begin PXF configuration, examine the PXF_CONF directory on master:

    $ kubectl exec -it master-0 bash -- -c "ls -R /etc/pxf"
    conf  keytabs  lib  logs  servers  templates
    /etc/pxf/conf:  pxf-profiles.xml
    adl-site.xml   hbase-site.xml  jdbc-site.xml    s3-site.xml
    core-site.xml  hdfs-site.xml   mapred-site.xml  wasbs-site.xml
    gs-site.xml    hive-site.xml   minio-site.xml   yarn-site.xml

    The Pivotal PXF service has just been initialized, and the PXF_CONF directory (/etc/pxf) contains the subdirectories and template configuration files required to begin configuring PXF for your external data sources. Continue with the instructions in Verifying PXF if you want to verify basic PXF functionality in the new cluster.

    Or, follow the relevant instructions in the Pivotal Greenplum Documentation to configure PXF for connectivity with one or more data sources:

- [Configuring PXF Hadoop Connectors (Optional)](
- [Configuring Connectors to Azure, Google Cloud Storage, Minio, and S3 Object Stores (Optional)](
- [Configuring the JDBC Connector (Optional)](

Verifying PXF

Follow these steps to quickly verify PXF operation in your new cluster, using MinIO as an example data source.

  1. Use helm to install a standalone MinIO server to your cluster. For example:

    $ helm install stable/minio
    NAME:   voting-coral                                                                                                                                                                                                LAST DEPLOYED: Wed Oct 16 07:44:38 2019
    NAMESPACE: default
    ==> v1/ConfigMap
    NAME                DATA  AGE
    voting-coral-minio  1     0s
    ==> v1/Deployment
    voting-coral-minio  0/1    1           0          0s
    ==> v1/PersistentVolumeClaim
    NAME                STATUS  VOLUME                                    CAPACITY  ACCESS MODES  STORAGECLASS      AGE
    voting-coral-minio  Bound   pvc-a7d03ab0-4fda-4328-a1ed-2f603888ed13  10Gi      RWO           standard      0s
    ==> v1/Pod(related)
    NAME                                 READY  STATUS             RESTARTS  AGE
    voting-coral-minio-7fd9b4c78b-ddp97  0/1    ContainerCreating  0         0s
    ==> v1/Secret
    NAME                TYPE    DATA  AGE
    voting-coral-minio  Opaque  2     0s
    ==> v1/Service
    NAME                TYPE       CLUSTER-IP    EXTERNAL-IP  PORT(S)   AGE
    voting-coral-minio  ClusterIP  <none>       9000/TCP  0s
    ==> v1/ServiceAccount
    NAME                SECRETS  AGE
    voting-coral-minio  1        0s
    To access Minio from localhost, run the below commands:                                                   
    1. export POD_NAME=$(kubectl get pods --namespace default -l "release=voting-coral" -o jsonpath="{.item[0]}")                                                                                         
    2. kubectl port-forward $POD_NAME 9000 --namespace default                                                
    You can now access Minio server on http://localhost:9000. Follow the below steps to connect to Minio server with mc client:                                                                                    
    3. mc ls voting-coral-minio-local                                                                         
    Alternately, you can use your browser or the Minio SDK to access the server -
  2. Execute the commands shown at the end of the MinIO deployment output to make the MinIO server accessible from the local host. Using the above output as an example:

    $ export POD_NAME=$(kubectl get pods --namespace default -l "release=voting-coral" -o jsonpath="{.item[0]}")
    $ kubectl port-forward $POD_NAME 9000 --namespace default
    Forwarding from -> 9000
    Forwarding from [::1]:9000 -> 9000

    Also make note of the MinIO service name used within the cluster (“voting-coral-minio” in the above example). You will use this name when defining the MinIO endpoint in the PXF configuration.

  3. You can create the sample data in MinIO using either the web interface, accessible from http://localhost:9000, or using the mc client. The example steps that follow use mc commands:

    1. Configure the mc client to use the MinIO server you just deployed:

      $ mc config host add minio http://localhost:9000 AKIAIOSFODNN7EXAMPLE wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY
      Added `minio` successfully.

      The accessKey and secretKey in the above command are used in the default helm chart deployment.

    2. Make a new bucket named pxf to store the sample data:

      $ mc mb minio/pxf
      Bucket created successfully `minio/pxf`.
    3. Create a delimited plain text data file named pxf_s3_simple.txt to provide the sample data:

      $ echo 'Prague,Jan,101,4875.33
      Beijing,Jul,411,11600.67' > ./pxf_s3_simple.txt
    4. Copy the sample data file to the MinIO bucket you created:

      $  mc cp ./pxf_s3_simple.txt minio/pxf
      ./pxf_s3_simple.txt:                      192 B / 192 B [===============] 100.00% 6.46 KiB/s 0s
  4. To configure PXF to access MinIO, begin by copying the template PXF MinIO configuration file to your local host:

    $ kubectl cp master-0:/etc/pxf/templates/minio-site.xml ./minio-site.xml
  5. Open the copied template file in a text editor, and edit the file entries to access the MinIO server that you deployed with helm. The file contents should be similar to:

    <?xml version="1.0" encoding="UTF-8"?>

    Be sure to change the first property value, fs.s3a.endpoint, to the URL of the MinIO service that was deployed in your cluster. The access.key and secret.key values in the above output are used in the default helm chart deployment. All other property values are the defaults provided in the template file.

  6. Save the file and exit your text editor.

  7. Use kubectl get all and record the names of the PXF server pods that were created for your cluster. For example:

    $ watch kubectl get all
    NAME                                      READY     STATUS    RESTARTS   AGE
    pod/greenplum-operator-79cddcf586-ctftb   1/1       Running   0          2m40s
    pod/master-0                              1/1       Running   0          23s
    pod/master-1                              1/1       Running   0          22s
    pod/my-greenplum-pxf-676fd6fdd7-825gq     0/1       Running   0          28s
    pod/my-greenplum-pxf-676fd6fdd7-mjt6w     0/1       Running   0          28s
    pod/segment-a-0                           1/1       Running   0          22s
    pod/segment-b-0                           1/1       Running   0          22s

    The remaining steps use the example names my-greenplum-pxf-676fd6fdd7-825gq and my-greenplum-pxf-676fd6fdd7-mjt6w from the above output. Substitute the correct names for your cluster.

  8. Copy your modified minio-site.xml file for use as the default PXF server configuration on all PXF pods deployed in your server. For example:

    $ kubectl cp ./minio-site.xml my-greenplum-pxf-676fd6fdd7-825gq:/etc/pxf/servers/default/minio-site.xml
    $ kubectl cp ./minio-site.xml my-greenplum-pxf-676fd6fdd7-mjt6w:/etc/pxf/servers/default/minio-site.xml
  9. At this point, you have deployed a MinIO server with sample data, and configured your Greenplum cluster with a default PXF server configuration to access the MinIO server. Perform the remaining steps on the Greenplum master pod to create and query an external table that references the sample MinIO data:

    1. Open a bash shell on the master-0 pod:

      $ kubectl exec -it master-0 bash
    2. Start the psql subsystem:

      $ psql -d postgres
      psql (8.3.23)
      Type "help" for help.
    3. Create the PXF extension in the database:

      postgres=# create extension pxf;
    4. Use the PXF s3:text profile to create a Greenplum Database external table that references the pxf_s3_simple.txt file that you just created and added to MinIO. This command omits the typical &SERVER=<server_name> option in the PXF location URL, because the procedure created only the default server configuration:

      postgres=# CREATE EXTERNAL TABLE pxf_s3_textsimple(location text, month text, num_orders int, total_sales float8) 
                 LOCATION ('pxf://pxf/pxf_s3_simple.txt?PROFILE=s3:text') 
                 FORMAT 'TEXT' (delimiter=E',');
    5. Query the external table to access the sample data stored on MinIO:

      postgres=# SELECT * FROM pxf_s3_textsimple;
       location  | month | num_orders | total_sales
       Prague    | Jan   |        101 |     4875.33
       Rome      | Mar   |         87 |     1557.39
       Bangalore | May   |        317 |     8936.99
       Beijing   | Jul   |        411 |    11600.67
      (4 rows)

    If you receive any errors when querying the external table, verify the contents of the /etc/pxf/servers/default/minio-site.xml file on each PXF server in the cluster. Also use the mc client to verify the contents and location of the sample data file on MinIO.

    Further PXF troubleshooting information is available in the Greenplum Database documentation at Troubleshooting PXF.

Exporting a PXF Configuration to S3

Follow these steps to export an existing PXF Configuration to an S3 bucket, so that you can use the configuration in later deployments of Pivotal Greenplum for Kubernetes. This procedure requires that you have a working PXF service configuration (a customized $PXF_CONF directory) available in a Greenplum for Kubernetes cluster. The sample commands use the default PXF server configuration that is created in Verifying PXF.

  1. Create a temporary directory, and copy in the complete PXF_CONF directory from a PXF server pod in your Greenplum for Kubernetes cluster. For example:

    $ mkdir ./pxf-temp
    $ kubectl cp master-0:/etc/pxf ./pxf-temp/
  2. Copy the PXF_CONF contents from your temporary directory to a S3 bucket and folder. For example:

    $ mc cp ./pxf-temp minio/pxf --recursive
  3. Create a secret that can be used to authenticate access to the S3 bucket and folder that contains the PXF configuration directory. For example:

    $ kubectl create secret generic my-greenplum-pxf-configs --from-literal='access_key_id=AKIAIOSFODNN7EXAMPLE' --from-literal='secret_access_key=wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY'
    secret/my-greenplum-pxf-configs created

    The above command creates a secret named my-greenplum-pxf-configs using the S3 access and secret keys that you provide. Replace the access and secret key values with the actual values for your system. If necessary, use your S3 implementation documentation to generate a secret access key.

  4. Edit the Greenplum for Kubernetees manifest file you use to deploy Greenplum and PXF. Uncomment or add the required PXF configuration properties. For example:

          secret: "my-greenplum-pxf-configs"
          endpoint: "http://voting-coral-minio:9000"
          bucket: "PXF"
          folder: "pxf-temp"

    Replace my-greenplum-pxf-configs with the actual secret that you created in the previous step (if you used a different name). Similarly, replace the remaining property values with the endpoint, bucket name, and folder in which you have placed the full contents of the PXF_CONF directory.

    When you deploy a new cluster with the above property values, the Greenplum operator supplies the secret and S3 location that you specified to all PXF pods in your cluster so that they can copy the PXF_CONF directory.

  5. Use the procedure described in Deploying a New Greenplum Cluster to deploy a new cluster, but use your modified manifest file containing the S3 PXF configuration.

  6. After the new cluster deployment is complete, validate your PXF configuration and test PXF connectivity. For example, examine the PXF_CONF directory on a deployed PXF server pod to ensure that your customized PXF configuration was installed:

    $ kubectl exec -it my-greenplum-pxf-676fd6fdd7-825gq bash -- -c "ls -R /etc/pxf"