This is a Kubernetes Operator to manage Apache Hive Metastore
The metastore-operator is an important part of the Nineinfra Data Platform, a curated combination of open-source projects including Spark, Flink, HDFS, Kafka, ClickHouse, Kyuubi, and Superset, all working together seamlessly to provide users with a stable and user-friendly big data processing platform. Nineinfra is a full-stack data platform built on Kubernetes, capable of running on public cloud, private cloud, or on-premises environments.
You’ll need a Kubernetes cluster to run against. You can use KIND to get a local cluster for testing, or run against a remote cluster.
Note: Your controller will automatically use the current context in your kubeconfig file (i.e. whatever cluster kubectl cluster-info
shows).
Install metastore operator by helm:
helm repo add nineinfra-charts https://nineinfra.github.io/nineinfra-charts/
kubectl create namespace metastore-operator
helm install metastore-operator nineinfra-charts/metastore-operator -n metastore-operator
- Obtain the necessary configuration information, including HDFS cluster or Minio cluster, and Spark cluster configuration information. example:
kubectl get svc -n dwh |grep hdfs
hdfs ClusterIP 10.100.208.68 <none> 9820/TCP,9870/TCP,9871/TCP
kubectl get svc -n minio-t1 |grep minio
minio LoadBalancer 10.106.61.173 <pending> 80:31063/TCP 11d
minio-t1-console LoadBalancer 10.96.125.48 <pending> 9090:30848/TCP 11d
minio-t1-hl ClusterIP None <none> 9000/TCP 11d
And the service of hdfs is hdfs with the suffix .dwh.svc
- Edit the cr yaml, there is a sample file like config/samples/metastore_v1alpha1_metastorecluster.yaml
apiVersion: metastore.nineinfra.tech/v1alpha1
kind: MetastoreCluster
metadata:
labels:
app.kubernetes.io/name: metastorecluster
app.kubernetes.io/instance: metastorecluster-sample
app.kubernetes.io/part-of: metastore-operator
app.kubernetes.io/managed-by: kustomize
app.kubernetes.io/created-by: metastore-operator
name: metastorecluster-sample
spec:
metastoreVersion: 3.1.3
metastoreImage:
repository: nineinfra/metastore
tag: v3.1.3
metastoreResource:
replicas: 1
metastoreConf:
"hive.metastore.warehouse.dir": "/usr/hive/warehouse"
clusterRefs:
- name: "database"
type: "database"
database:
connectionUrl: "jdbc:postgresql://postgresql:5432/hive"
dbType: postgres
userName: hive
password: hive
- name: "minio"
type: "minio"
minio:
endpoint: "http://192.168.123.24:31063"
accessKey: "984GcQyUWobTVl3B"
secretKey: "wE5ffRYxSacalsYT5UAVgo1AMlK2uune"
sslEnabled: "false"
pathStyleAccess: "true"
- Deploy a metastore cluster
kubectl apply -f metastore_v1alpha1_metastorecluster_postgres.yaml -n metastore-operator
- Get status of the metastore cluster,you can access the metastore cluster with the service in the output
kubectl get metastorecluster metastorecluster-sample -n metastore-operator -o yaml
status:
creationTime: "2023-10-15T15:01:48Z"
exposedInfos:
- exposedType: thrift-http
name: thrift-http-0
serviceName: metastorecluster-sample-metastore
servicePort:
name: thrift-http
port: 9083
protocol: TCP
targetPort: 9083
updateTime: "2023-10-15T15:01:48Z"
The hive.metastore.uris value is from the metastorecluster's status above
apiVersion: kyuubi.nineinfra.tech/v1alpha1
kind: KyuubiCluster
metadata:
labels:
app.kubernetes.io/name: kyuubicluster
app.kubernetes.io/instance: kyuubicluster-sample
app.kubernetes.io/part-of: kyuubi-operator
app.kubernetes.io/managed-by: kustomize
app.kubernetes.io/created-by: kyuubi-operator
name: kyuubicluster-sample
spec:
kyuubiVersion: 1.8.1
kyuubiImage:
repository: 172.18.123.24:30003/library/kyuubi
tag: v1.8.1-minio
kyuubiResource:
replicas: 1
kyuubiConf:
kyuubi.kubernetes.namespace: dwh
kyuubi.frontend.connection.url.use.hostname: 'false'
kyuubi.frontend.thrift.binary.bind.port: '10009'
kyuubi.frontend.thrift.http.bind.port: '10010'
kyuubi.frontend.rest.bind.port: '10099'
kyuubi.frontend.mysql.bind.port: '3309'
kyuubi.frontend.protocols: REST,THRIFT_BINARY
kyuubi.metrics.enabled: 'false'
clusterRefs:
- name: spark
type: spark
spark:
sparkMaster: k8s
sparkImage:
repository: 172.18.123.24:30003/library/spark
tag: v3.2.4-minio
sparkNamespace: dwh
sparkDefaults:
spark.hadoop.fs.s3a.access.key: 984GcQyUWobTVl3B
spark.hadoop.fs.s3a.secret.key: wE5ffRYxSacalsYT5UAVgo1AMlK2uune
spark.hadoop.fs.s3a.path.style.access: "true"
spark.hadoop.fs.s3a.connection.ssl.enabled: "false"
spark.hadoop.fs.s3a.endpoint: http://172.18.123.24:31063
- name: hdfs
type: hdfs
hdfs:
coreSite:
hdfsSite:
dfs.client.block.write.retries: '3'
- name: metastore
type: metastore
metastore:
hiveSite:
hive.metastore.uris: thrift://metastorecluster-sample-metastore.dwh.svc:9083
hive.metastore.warehouse.dir: s3a://usr/hive/warehouse
- Login the kyuubi pod
kubectl get pod -n kyuubi-operator
NAME READY STATUS RESTARTS AGE
kyuubi-kyuubi-user-spark-sql-anonymous-default-2f33bda8-a9b0-4584-a726-1d653c809d15-2f33bda8-a9b0-4584-a726-1d653c809d15-exec-1 0/1 Completed 0 92m
kyuubi-kyuubi-user-spark-sql-anonymous-default-2f33bda8-a9b0-4584-a726-1d653c809d15-2f33bda8-a9b0-4584-a726-1d653c809d15-exec-2 0/1 Completed 0 92m
kyuubi-operator-deployment-57b54cbc6-fc8bz 1/1 Running 0 85m
kyuubicluster-sample-kyuubi-0 1/1 Running 0 83m
kubectl exec -it kyuubicluster-sample-kyuubi-0 -n kyuubi-operator -- bash
- Run beeline command
kyuubi@kyuubicluster-sample-kyuubi-0:/opt/kyuubi$ cd bin
kyuubi@kyuubicluster-sample-kyuubi-0:/opt/kyuubi/bin$ ./beeline
Warn: Not find kyuubi environment file /opt/kyuubi/conf/kyuubi-env.sh, using default ones...
Beeline version 1.8.0-SNAPSHOT by Apache Kyuubi
beeline>
- Connect the kyuubi cluster by the thrift-binary protocol through the service kyuubicluster-sample-kyuubi
beeline> !connect jdbc:hive2://kyuubicluster-sample-kyuubi:10009
- Execute some example sql statements
0: jdbc:hive2://kyuubicluster-sample-kyuubi:1> show databases;
+------------+
| namespace |
+------------+
| default |
| test |
+------------+
0: jdbc:hive2://kyuubicluster-sample-kyuubi:1> use test;
0: jdbc:hive2://kyuubicluster-sample-kyuubi:1> create table test3 (name string,id int);
0: jdbc:hive2://kyuubicluster-sample-kyuubi:1> insert into test3 values("kyuubi",1);
0: jdbc:hive2://kyuubicluster-sample-kyuubi:1> select * from test3;
+---------+-----+
| name | id |
+---------+-----+
| kyuubi | 1 |
+---------+-----+
// TODO(user): Add detailed information on how you would like others to contribute to this project
This project aims to follow the Kubernetes Operator pattern.
It uses Controllers, which provide a reconcile function responsible for synchronizing resources until the desired state is reached on the cluster.
- Install the CRDs into the cluster:
make install
- Run your controller (this will run in the foreground, so switch to a new terminal if you want to leave it running):
make run
NOTE: You can also run this in one step by running: make install run
If you are editing the API definitions, generate the manifests such as CRs or CRDs using:
make manifests
NOTE: Run make --help
for more information on all potential make
targets
More information can be found via the Kubebuilder Documentation
Copyright 2023 nineinfra.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.