You are looking at the documentation of a prior release. To read the documentation of the latest release, please
visit here.
Backup and Restore Percona XtraDB Cluster using Stash
Stash 0.9.0+ supports backup and restoration of Percona XtraDB cluster databases. This guide will show you how you can backup and restore your Percona XtraDB cluster with Stash.
Before You Begin
- At first, you need to have a Kubernetes cluster, and the
kubectl
command-line tool must be configured to communicate with your cluster. If you do not already have a cluster, you can create one by using Minikube. - Install Stash in your cluster following the steps here.
- Install KubeDB in your cluster following the steps here. This step is optional. You can deploy your database using any method you want. We are using KubeDB because KubeDB simplifies many of the difficult or tedious management tasks to run a production-grade database on private and public clouds.
- If you are not familiar with how Stash takes backup and restores Percona XtraDB, please check the following guide here.
You have to be familiar with the following custom resources:
To keep things isolated, we are going to use a separate namespace called demo
throughout this tutorial. Create demo
namespace if you haven’t created yet.
$ kubectl create ns demo
namespace/demo created
Note: YAML files used in this tutorial are stored here.
Backup Percona XtraDB Cluster
This section will demonstrate how to backup a Percona XtraDB cluster. Here, we are going to deploy a Percona XtraDB cluster using KubeDB. Then, we are going to back up this database into a GCS bucket. Finally, we are going to restore the backed up data into another Percona XtraDB cluster.
Deploy Sample Percona XtraDB Cluster
Let’s deploy a sample Percona XtraDB cluster and insert some data into it.
Create Percona XtraDB CRD
Below is the YAML of a sample PerconaXtraDB
CRD that we are going to create for this tutorial:
apiVersion: kubedb.com/v1alpha2
kind: PerconaXtraDB
metadata:
name: sample-xtradb-cluster
namespace: demo
spec:
version: "5.7-cluster"
replicas: 3
storageType: Durable
storage:
storageClassName: "standard"
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 1Gi
terminationPolicy: WipeOut
Create the above PerconaXtraDB
CRD,
$ kubectl apply -f https://github.com/stashed/docs/raw/v2024.4.8/docs/addons/percona-xtradb/cluster/examples/sample-xtradb-cluster.yaml
perconaxtradb.kubedb.com/sample-xtradb-cluster created
KubeDB will deploy a Percona XtraDB cluster according to the above specification. It will also create the necessary Secrets and Services to access the database.
Let’s check if the database is ready to use,
$ kubectl get px -n demo sample-xtradb-cluster
NAME VERSION STATUS AGE
sample-xtradb-cluster 5.7-cluster Ready 7m46s
The database is Ready
. Verify that KubeDB has created a Secret and a Service for this database using the following commands,
$ kubectl get secret -n demo -l=app.kubernetes.io/instance=sample-xtradb-cluster
NAME TYPE DATA AGE
sample-xtradb-cluster-auth Opaque 2 9m2s
$ kubectl get service -n demo -l=app.kubernetes.io/instance=sample-xtradb-cluster
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
sample-xtradb-cluster ClusterIP 10.103.37.141 <none> 3306/TCP 11m
sample-xtradb-cluster-gvr ClusterIP None <none> 3306/TCP 11m
Here, we have to use service sample-xtradb-cluster
and secret sample-xtradb-cluster-auth
to connect with the database. KubeDB creates an AppBinding CRD that holds the necessary information to connect with the database.
Verify AppBinding
Verify that the AppBinding
has been created successfully using the following command,
$ kubectl get appbindings -n demo
NAME AGE
sample-xtradb-cluster 14m
Let’s check the YAML of the above AppBinding,
$ kubectl get appbindings -n demo sample-xtradb-cluster -o yaml
apiVersion: appcatalog.appscode.com/v1alpha1
kind: AppBinding
metadata:
creationTimestamp: "2019-10-30T11:41:20Z"
generation: 1
labels:
app.kubernetes.io/component: database
app.kubernetes.io/managed-by: kubedb.com
app.kubernetes.io/name: perconaxtradbs.kubedb.com
app.kubernetes.io/instance: sample-xtradb-cluster
name: sample-xtradb-cluster
namespace: demo
ownerReferences:
- apiVersion: kubedb.com/v1alpha2
blockOwnerDeletion: false
kind: PerconaXtraDB
name: sample-xtradb-cluster
uid: 79d90fc4-f5e8-4a8c-83d7-3eae7c12f01a
resourceVersion: "12319"
selfLink: /apis/appcatalog.appscode.com/v1alpha1/namespaces/demo/appbindings/sample-xtradb-cluster
uid: 977cb8fd-b5e5-4830-a50f-58de9eb5d82c
spec:
clientConfig:
service:
name: sample-xtradb-cluster
path: /
port: 3306
scheme: mysql
url: tcp(sample-xtradb-cluster:3306)/
parameters:
address: gcomm://sample-xtradb-cluster-0.sample-xtradb-cluster-gvr.demo,sample-xtradb-cluster-1.sample-xtradb-cluster-gvr.demo,sample-xtradb-cluster-2.sample-xtradb-cluster-gvr.demo
apiVersion: config.kubedb.com/v1alpha2
group: sample-xtradb-cluster
kind: GaleraArbitratorConfiguration
sstMethod: xtrabackup-v2
secret:
name: sample-xtradb-cluster-auth
type: kubedb.com/perconaxtradb
version: "5.7-cluster"
Stash uses the AppBinding CRD to connect with the target database. It requires the following two fields to be set in the AppBinding’s .spec
section.
.spec.clientConfig.service.name
specifies the name of the Service that connects to the database..spec.secret
specifies the name of the Secret that holds the necessary credentials to access the database..spec.type
specifies the type of the app that this AppBinding is pointing to. The format KubeDB generated AppBinding follows to set the value of.spec.type
is<app_group>/<app_resource_type>
.
Creating AppBinding Manually
If you deploy the Percona XtraDB cluster without KubeDB, you have to create the AppBinding CRD manually in the same namespace as the service and secret of the database.
The following YAML shows a minimal AppBinding specification that you have to create if you deploy the Percona XtraDB cluster without KubeDB.
apiVersion: appcatalog.appscode.com/v1alpha1
kind: AppBinding
metadata:
name: your-custom-appbinding-name
namespace: your-database-namespace
spec:
clientConfig:
service:
name: your-database-service-name
port: 3306
scheme: mysql
secret:
name: your-database-auth-secret-name
# type field is optional. you can keep it empty.
# if you keep it empty then the value of TARGET_APP_RESOURCE variable
# will be set to "appbinding" during auto-backup.
type: kubedb.com/perconaxtradb
You have to replace the <...>
quoted part with proper values in the above YAML.
Insert Sample Data
Now, we are going to exec into the database pod and create some sample data. At first, find out the database pods using the following command,
$ kubectl get pods -n demo --selector="app.kubernetes.io/instance=sample-xtradb-cluster"
NAME READY STATUS RESTARTS AGE
sample-xtradb-cluster-0 1/1 Running 0 39m
sample-xtradb-cluster-1 1/1 Running 0 38m
sample-xtradb-cluster-2 1/1 Running 0 37m
And copy the username and password of the root
user to access into mysql
shell.
$ kubectl get secret -n demo sample-xtradb-cluster-auth -o jsonpath='{.data.username}'| base64 -d
root⏎
$ kubectl get secret -n demo sample-xtradb-cluster-auth -o jsonpath='{.data.password}'| base64 -d
CZYWy7MDXiedL2EG⏎
Now, let’s exec into the Pod to enter into mysql
shell and create a database and a table,
$ kubectl exec -it -n demo sample-xtradb-cluster-0 -- mysql --user=root --password=CZYWy7MDXiedL2EG
mysql: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 275
Server version: 5.7.25-28-57 Percona XtraDB Cluster (GPL), Release rel28, Revision a2ef85f, WSREP version 31.35, wsrep_31.35
Copyright (c) 2009-2019 Percona LLC and/or its affiliates
Copyright (c) 2000, 2019, Oracle and/or its affiliates. All rights reserved.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
mysql> CREATE DATABASE playground;
Query OK, 1 row affected (0.01 sec)
mysql> SHOW DATABASES;
+--------------------+
| Database |
+--------------------+
| information_schema |
| mysql |
| performance_schema |
| playground |
| sys |
+--------------------+
5 rows in set (0.00 sec)
mysql> CREATE TABLE playground.equipment ( id INT NOT NULL AUTO_INCREMENT, type VARCHAR(50), quant INT, color VARCHAR(25), PRIMARY KEY(id));
Query OK, 0 rows affected (0.01 sec)
mysql> SHOW TABLES IN playground;
+----------------------+
| Tables_in_playground |
+----------------------+
| equipment |
+----------------------+
1 row in set (0.01 sec)
mysql> INSERT INTO playground.equipment (type, quant, color) VALUES ("slide", 2, "blue");
Query OK, 1 row affected (0.01 sec)
mysql> SELECT * FROM playground.equipment;
+----+-------+-------+-------+
| id | type | quant | color |
+----+-------+-------+-------+
| 1 | slide | 2 | blue |
+----+-------+-------+-------+
1 row in set (0.00 sec)
mysql> exit
Bye
Now, we are ready to back up the database.
Prepare Backend
We are going to store our backed up data into a GCS bucket. At first, we need to create a secret with GCS credentials then we need to create a Repository
CRD. If you want to use a different backend, please read the respective backend configuration doc from here.
Create Storage Secret
Let’s create a secret called gcs-secret
with access credentials to our desired GCS bucket,
$ echo -n 'changeit' > RESTIC_PASSWORD
$ echo -n '<your-project-id>' > GOOGLE_PROJECT_ID
$ cat downloaded-sa-json.key > GOOGLE_SERVICE_ACCOUNT_JSON_KEY
$ kubectl create secret generic -n demo gcs-secret \
--from-file=./RESTIC_PASSWORD \
--from-file=./GOOGLE_PROJECT_ID \
--from-file=./GOOGLE_SERVICE_ACCOUNT_JSON_KEY
secret/gcs-secret created
Create Repository
Now, crete a Repository
using this secret. Below is the YAML of Repository CRD we are going to create,
apiVersion: stash.appscode.com/v1alpha1
kind: Repository
metadata:
name: gcs-repo-xtradb-cluster
namespace: demo
spec:
backend:
gcs:
bucket: appscode-qa
prefix: /demo/xtradb/sample-xtradb-cluster
storageSecretName: gcs-secret
Let’s create the Repository
we have shown above,
$ kubectl create -f https://github.com/stashed/docs/raw/v2024.4.8/docs/addons/percona-xtradb/cluster/examples/repository.yaml
repository.stash.appscode.com/gcs-repo-xtradb-cluster created
Now, we are ready to back up our database to our desired backend.
Backup
We have to create a BackupConfiguration
targeting respective AppBinding CRD of our desired database. Then Stash will create a CronJob to periodically backup the database.
Create BackupConfiguration
Below is the YAML for BackupConfiguration
CRD to backup the sample-xtradb-cluster
database we have deployed earlier,
apiVersion: stash.appscode.com/v1beta1
kind: BackupConfiguration
metadata:
name: sample-xtradb-cluster-backup
namespace: demo
spec:
schedule: "*/5 * * * *"
task:
name: perconaxtradb-backup-5.7
repository:
name: gcs-repo-xtradb-cluster
target:
ref:
apiVersion: appcatalog.appscode.com/v1alpha1
kind: AppBinding
name: sample-xtradb-cluster
retentionPolicy:
name: keep-last-5
keepLast: 5
prune: true
Here,
.spec.schedule
specifies that we want to back up the database at 5 minutes interval..spec.task.name
specifies the name of the Task CRD that specifies the necessary Functions and their execution order to backup a Percona XtraDB cluster..spec.target.ref
refers to the AppBinding CRD that was created for thesample-xtradb-cluster
database.
Let’s create the BackupConfiguration
CRD we have shown above,
$ kubectl create -f https://github.com/stashed/docs/raw/v2024.4.8/docs/addons/percona-xtradb/cluster/examples/backupconfiguration.yaml
backupconfiguration.stash.appscode.com/sample-xtradb-cluster-backup created
Verify Backup Setup Successful
If everything goes well, the phase of the BackupConfiguration
should be Ready
. The Ready
phase indicates that the backup setup is successful. Let’s verify the Phase
of the BackupConfiguration,
$ kubectl get backupconfiguration -n demo
NAME TASK SCHEDULE PAUSED PHASE AGE
sample-xtradb-cluster-backup perconaxtradb-backup-5.7 */5 * * * * Ready 11s
Verify CronJob
Stash will create a CronJob with the schedule specified in .spec.schedule
field of BackupConfiguration
CRD.
Verify that the CronJob has been created using the following command,
$ kubectl get cronjob -n demo
NAME SCHEDULE SUSPEND ACTIVE LAST SCHEDULE AGE
sample-xtradb-cluster-backup */5 * * * * False 0 49s 2m22s
Wait for BackupSession
The sample-xtradb-cluster-backup
CronJob will trigger a backup on each scheduled slot by creating a BackupSession
CRD.
Wait for a schedule to appear. Run the following command to watch BackupSession
CRD,
$ kubectl get backupsession -n demo -l=stash.appscode.com/invoker-name=sample-xtradb-cluster-backup --watch
NAME INVOKER-TYPE INVOKER-NAME PHASE AGE
sample-xtradb-cluster-backup-1572439801 BackupConfiguration sample-xtradb-cluster-backup Succeeded 4m27s
Here, the phase Succeeded
means that the backupsession has been succeeded.
Note: Backup CronJob creates
BackupSession
CRD the labelstash.appscode.com/invoker-name=<BackupConfiguration_crd_name>
. We can use this label to watch only theBackupSession
of our desiredBackupConfiguration
.
Verify Backup
Now, we are going to verify whether the backed up data is in the backend. Once a backup is completed, Stash will update the respective Repository
CRD to reflect the backup completion. Check that the repository gcs-repo-xtradb-cluster
has been updated by the following command,
$ kubectl get repository -n demo gcs-repo-xtradb-cluster
NAME INTEGRITY SIZE SNAPSHOT-COUNT LAST-SUCCESSFUL-BACKUP AGE
gcs-repo-xtradb-cluster true 304.165 MiB 3 97s 13m
Now, if we navigate to the GCS bucket, we will see the backed up data has been stored in demo/xtradb/sample-xtradb-cluster
directory as specified by .spec.backend.gcs.prefix
field of Repository CRD.
Note: Stash keeps all the backed up data encrypted. So, data in the backend will not make any sense until they are decrypted.
Restore Percona XtraDB Cluster
In this section, we are going to restore the database from the backup we have taken in the previous section. We are going to deploy a new database and initialize it from the backup.
Stop Taking Backup of the Old Database
At first, let’s stop taking any further backup of the old database so that no backup is taken during the restore process. We are going to pause the BackupConfiguration
CRD that we had created to backup the sample-xtradb-cluster
database. Then, Stash will stop taking any further backup for this database.
Let’s pause the sample-xtradb-cluster-backup
BackupConfiguration,
$ kubectl patch backupconfiguration -n demo sample-xtradb-cluster-backup --type="merge" --patch='{"spec": {"paused": true}}'
backupconfiguration.stash.appscode.com/sample-xtradb-cluster-backup patched
Now, wait for a moment. Stash will pause the BackupConfiguration. Verify that the operator has paused the BackupConfiguration object,
$ kubectl get backupconfiguration -n demo sample-xtradb-cluster-backup
NAME TASK SCHEDULE PAUSED PHASE AGE
sample-xtradb-cluster-backup perconaxtradb-backup-5.7 */5 * * * * true Ready 50m
Notice the PAUSED
column. Value true
for this field means that the BackupConfiguration has been paused.
Deploy Restored Database
Now, we have to deploy the restored database similarly as we have deployed the original sample-xtradb-cluster
database. However, this time there will be the following differences:
- We have to use the same secret that was used in the original database. We are going to specify it using
.spec.databaseSecret
field. - We are going to specify
.spec.init.waitForInitialRestore: true
which tells KubeDB to wait for the initial restore to complete before marking this database as ready to use.
Below is the YAML for PerconaXtraDB
CRD we are going deploy to initialize from backup,
apiVersion: kubedb.com/v1alpha2
kind: PerconaXtraDB
metadata:
name: restored-xtradb-cluster
namespace: demo
spec:
version: "5.7-cluster"
replicas: 3
authSecret:
name: sample-xtradb-cluster-auth
storageType: Durable
storage:
storageClassName: "standard"
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 1Gi
init:
waitForInitialRestore: true
terminationPolicy: WipeOut
Let’s create the above database,
$ kubectl apply -f https://github.com/stashed/docs/raw/v2024.4.8/docs/addons/percona-xtradb/cluster/examples/restored-xtradb-cluster.yaml
perconaxtradb.kubedb.com/restored-xtradb-cluster created
If you check the database status, you will see it is stuck in Provisioning
state.
$ kubectl get px -n demo restored-xtradb-cluster
NAME VERSION STATUS AGE
restored-xtradb-cluster 5.7-cluster Provisioning 4m10s
Create RestoreSession
Now, we need to create a RestoreSession
CRD pointing to the newly created restored database.
In case of Percona XtraDB cluster, the RestoreSession object contains some different configurations unlike other databases supported by KubeDB. To restore Percona XtraDB cluster, Stash operator will create the required number of PVCs and mount the data in the data directory /var/lib/mysql
with proper ownership and permission. After completing the PVC creation, KubeDB then creates AppBinding, Secret, Services, etc. objects.
Below is the contents of YAML file of the RestoreSession CRD that we are going to create to restore the backed up data into the newly created database provisioned by PerconaXtrDB CRD named restored-xtradb-cluster
.
apiVersion: stash.appscode.com/v1beta1
kind: RestoreSession
metadata:
name: restored-xtradb-cluster-restore
namespace: demo
spec:
task:
name: perconaxtradb-restore-5.7
repository:
name: gcs-repo-xtradb-cluster
target:
replicas: 3
ref:
apiVersion: appcatalog.appscode.com/v1alpha1
kind: AppBinding
name: restored-xtradb-cluster
volumeMounts:
- name: data-restored-xtradb-cluster
mountPath: /var/lib/mysql
volumeClaimTemplates:
- metadata:
name: data-restored-xtradb-cluster-${POD_ORDINAL}
spec:
accessModes: [ "ReadWriteOnce" ]
storageClassName: "standard"
resources:
requests:
storage: 1Gi
rules:
- targetHosts: [] # empty host match all hosts
sourceHost: "host-0" # restore same data on all pvc
snapshots: ["latest"]
Here,
.spec.task.name
specifies the name of the Task CRD that specifies the necessary Functions and their execution order to restore a Percona XtraDB cluster..spec.repository.name
specifies the Repository CRD that holds the backend information where our backed up data has been stored..spec.target.replicas
specifies the number of PVCs where snapshot data will be restored..spec.target.ref
refers to the AppBinding object for therestored-xtradb-cluster
PerconaXtraDB object. Though the KubeDB operator will create this AppBinding object later, we need to tell Stash operator about this AppBinding object ref. Because the AppBinding object name is identical with the corresponding PerconaXtraDB object name and the names of the PVCs directly depend on this name..spec.target.volumeClaimTemplates
specifies the template used for the PVCs. The important thing here is the.metadata.name
. In KubeDB side, the PVC name is formed by following the ruledata-<xtradb_crd_object_name>-<statefulset_pod_ordinal>
. Since we have created our restore database namedrestored-xtradb-cluster
and later KubeDB will create a StatefulSet for this database, the PVC names will bedata-restored-xtradb-cluster-0
,data-restored-xtradb-cluster-1
,data-restored-xtradb-cluster-2
, etc. up to the number of replicas. Here Stash operator will create these PVCs by following the same convention as KubeDB. We just need to provide the.metadata.name
asdata-<xtradb_crd_object_name>-${POD_ORDINAL}
. You must insert${POD_ORDINAL}
at the end of the name. Stash will create the required PVCs by replacing this with the corresponding pod index. That means if the value of.spec.target.replicas
is 3, then Stash will create 3 PVCs nameddata-restored-xtradb-cluster-0
,data-restored-xtradb-cluster-1
, anddata-restored-xtradb-cluster-2
..spec.target.volumeMounts
specifies the mount path for the volume. ThemountPath
must be/var/lib/mysql
as expected by Percona XtraDB server. And the volume name is form as"data-<xtradb_crd_object_name>"
. Since for restoring purpose, we have created a PerconaXtraDB object namedrestored-xtradb-cluster
, the volume name will be"data-restored-xtradb-cluster"
..spec.rules
specifies that we are restoring data from thelatest
backup snapshot of the database. Empty ([]
)targetHosts
means snapshot data will be restored in all specified number of PVCs. And another obvious thing is we want to restore the same data fromhost-0
to all PVCs. During the backup procedure, we took backup data ashost-0
from the Percona XtraDB cluster. So, here the source host ishost-0
.
Let’s create the RestoreSession CRD object we have shown above,
$ kubectl apply -f https://github.com/stashed/docs/raw/v2024.4.8/docs/addons/percona-xtradb/cluster/examples/restoresession.yaml
restoresession.stash.appscode.com/restored-xtradb-cluster-restore created
Once you have created the RestoreSession object, Stash will create a restore Job. We can watch the phase of the RestoreSession object to check whether the restore process has succeeded or not.
Run the following command to watch the phase of the RestoreSession object,
$ kubectl get restoresession -n demo restored-xtradb-cluster-restore --watch
NAME REPOSITORY PHASE AGE
restored-xtradb-cluster-restore gcs-repo-xtradb-cluster Running 3m33s
restored-xtradb-cluster-restore gcs-repo-xtradb-cluster Running 3m51s
restored-xtradb-cluster-restore gcs-repo-xtradb-cluster Running 3m58s
restored-xtradb-cluster-restore gcs-repo-xtradb-cluster Succeeded 3m58s
Here, we can see from the output of the above command that the restore process succeeded.
Verify Restored Data
In this section, we are going to verify whether the desired data has restored successfully. We are going to connect to the database server and check whether the database and the table we created earlier in the original database have restored.
At first, check if the database has gone into Running
state,
$ kubectl get px -n demo restored-xtradb-cluster --watch
NAME VERSION STATUS AGE
restored-xtradb-cluster 5.7-cluster Provisioning 3m36s
restored-xtradb-cluster 5.7-cluster Provisioning 4m4s
restored-xtradb-cluster 5.7-cluster Ready 4m4s
Now, find out the database Pod,
$ kubectl get pods -n demo --selector="app.kubernetes.io/instance=restored-xtradb-cluster" --watch
NAME READY STATUS RESTARTS AGE
restored-xtradb-cluster-0 1/1 Running 0 115s
restored-xtradb-cluster-1 1/1 Running 0 77s
restored-xtradb-cluster-2 1/1 Running 0 41s
And then copy the user name and password of the root
user to access into mysql
shell.
Notice: We used the same Secret for the
restored-xtradb-cluster
object. So, we will use the same commands as before.
$ kubectl get secret -n demo sample-xtradb-cluster-auth -o jsonpath='{.data.username}'| base64 -d
root⏎
$ kubectl get secret -n demo sample-xtradb-cluster-auth -o jsonpath='{.data.password}'| base64 -d
CZYWy7MDXiedL2EG⏎
Now, let’s exec into the Pod to enter into mysql
shell and check the database and the table we created before,
$ kubectl exec -it -n demo restored-xtradb-cluster-0 -- mysql --user=root --password=CZYWy7MDXiedL2EG
mysql: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 275
Server version: 5.7.25-28-57 Percona XtraDB Cluster (GPL), Release rel28, Revision a2ef85f, WSREP version 31.35, wsrep_31.35
Copyright (c) 2009-2019 Percona LLC and/or its affiliates
Copyright (c) 2000, 2019, Oracle and/or its affiliates. All rights reserved.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
mysql> SHOW DATABASES;
+--------------------+
| Database |
+--------------------+
| information_schema |
| mysql |
| performance_schema |
| playground |
| sys |
+--------------------+
5 rows in set (0.00 sec)
mysql> SHOW TABLES IN playground;
+----------------------+
| Tables_in_playground |
+----------------------+
| equipment |
+----------------------+
1 row in set (0.00 sec)
mysql> SELECT * FROM playground.equipment;
+----+-------+-------+-------+
| id | type | quant | color |
+----+-------+-------+-------+
| 1 | slide | 2 | blue |
+----+-------+-------+-------+
1 row in set (0.00 sec)
mysql> exit
Bye
So, from the above output, we can see that the playground
database and the equipment
table we created before in the original database are restored successfully.
Cleanup
To cleanup the Kubernetes resources created by this tutorial, run:
kubectl delete restoresession -n demo restored-xtradb-cluster-restore
kubectl delete px -n demo restored-xtradb-cluster
kubectl delete repository -n demo gcs-repo-xtradb-cluster
kubectl delete backupconfiguration -n demo sample-xtradb-cluster-backup
kubectl delete px -n demo sample-xtradb-cluster