Red Hat Marketplace #

Red Hat Marketplace

Starburst Enterprise platform (SEP) is available on Red Hat OpenShift.

OpenShift from Red Hat Marketplace (RHM) is a container platform using Kubernetes Helm charts that automate the provisioning, management, and scaling of applications to any cloud platform or even on-prem. Starburst Enterprise platform (SEP) is available on RHM as of OpenShift version 4.

Prerequisites #

Before you get started, here are some things you need:

  • Access to an OpenShift cluster using IAM credentials, and with sufficient Elastic IPs
  • Previously installed and configured Kubernetes, including access to kubectl
  • A Helm chart repository for your organization
  • An editor suitable for editing YAML files
  • Your SEP license file

For more information, please refer to our Kubernetes with Helm documentation.

Quick start #

After you have signed up through RHM, download the latest OpenShift Container Platform (OCP) client for your platform from the OpenShift mirror site, and copy the oc executable into your path, usually /usr/local/bin. Once this is done, you are ready to install the operator in OCP4.

Using your administrator login for Red Hat OCP, log in to the OCP web console and click Operators > OperatorHub in the left-hand menu.

Once there, select “Presto” from the Project: drop-down menu, and navigate through the projects to All Items > Big Data > Starburst until you see Starburst Enterprise. Click on that tile, then click the Install button.

When the Create Operator Subscription page appears, select the Starburst project as the specific namespace on the cluster, leave all other options as default, and click Subscribe.

When the operation is complete, you are subscribed to the SEP operator, and it is installed and accessible to you in OCP4.

Getting up and running #

Installation #

Once you have your operator subscription in place, it’s time to install. There are several steps to getting SEP installed and deployed:

  • Installing the SEP cluster
  • Installing the Hive Metastore Service (HMS)
  • Installing Apache Ranger (optional)

You must install the HMS to connect and query any objects storage with the Hive connector. This is typically a core use case for SEP, and then a required step. The HMS is used by SEP to manage the metadata of any objects storage.

Once these installations are complete, you need to assign containers to specific nodes or pods in your cluster.

Read more about installation and node assignments in our Kubernetes with Helm installation guide.

Configuration #

When your installations are complete and the nodes are assigned containers, you must configure your cluster. Just like with installation, there are several steps to configuring Starburst Enterprise:

Each of these steps uses a specific Helm chart. Click on the links for detailed instructions on configuring each of the SEP Helm charts.

Make sure to configure your desired data sources as catalogs when you configure the SEP chart.

Next steps #

Your cluster is now operational! You can now connect to it with your client tools and start querying your data sources.

We’ve created an operations guide to get you started with common first steps in cluster operations.

It includes some great advice about starting with a small, initial configuration that is built upon in our cluster sizing and performance video training.

Troubleshooting #

SEP is powerful, enterprise-grade software with many moving parts. As such, if you find you need help troubleshooting, here are some helpful resources:

FAQs #

Q: Once it’s deployed, how do I access my cluster? #

A: You can use the CLI on a terminal or the Web UI to access your cluster. For example:

  • Presto CLI command: ./presto --server --catalog hive

  • Web UI URL:

  • Many other client applications can be connected, and used to run queries, created dashboards and more.

Q: I need to make administrative changes that require a shell prompt. How to I get a command line shell prompt in a container within my cluster? #

A: On OCP, you’ll get a shell prompt for a pod. To get a shell prompt for a pod, you’ll need the name of the pod you want to work from. To do so, log in to your cluster as per your RHM documentation. For example:

oc login -u kubeadmin -p XXXXX-XXXXX-XXXXX-XXXX

Get the list of running pods:

❯ oc get pod -o wide
NAME                                                 READY   STATUS    RESTARTS   AGE   IP            NODE                                         NOMINATED NODE   READINESS GATES
hive-metastore-example-presto-XXXXXXXXX-lhj7l        1/1     Running   0          27m   10.131.2.XX    <none>           <none>
presto-coordinator-example-presto-XXXXXXXXX-4bzrv   1/1     Running   0          27m   10.129.2.XX     <none>           <none>
presto-operator-7c4ff6dd8f-2xxrr                     1/1     Running   0          41m   10.131.2.XX    <none>           <none>
presto-worker-example-presto-XXXXXXXXX-522j8        1/1     Running   0          27m   10.131.2.XX    <none>           <none>
presto-worker-example-presto-XXXXXXXXX-kwxhr        1/1     Running   0          27m   10.130.2.XX   <none>           <none>
presto-worker-example-presto-XXXXXXXXX-phlqq        1/1     Running   0          27m   10.129.2.XX     <none>           <none>

The pod name is the first value in a record. Use the pod name to open a shell:

❯ oc rsh presto-coordinator-example-presto-XXXXXXXXX-4bzrv

A shell prompt will appear. For example, on OCP 4.4:


Q: Is there a way to get a shell prompt through the OCP web console? #

A: Yes. Log in to your OCP web console and navigate to Workloads > Pods. Select the pod you want a terminal for, and click the Terminal tab.

Q: I’ve added a new data source. How do I update the configuration to recognize it? #

A: Using the making configuration changes section to edit your YAML configuration, find additionalCatalogs, and add an entry for your new data source. For example, to add a PostgreSQL data source called mydatabase:

    mydatabase: |

Once your changes are complete, click Save and then Reload to deploy your changes. Note that this restarts the coordinator and all workers on the cluster, and might take a little while.