Alpha Percona Xtradb Cluster release Only for limited production use
pxc-release is a BOSH release of MySQL Galera that can be used as a backing store for Cloudfoundry. The Galera Cluster Provider is Percona Xtradb Cluster. This release is intended as a drop-in replacement for cf-mysql-release.
If you were previously using the cf-mysql
release, we recommend using the pxc-mysql
job. Even if you were deploying the cf-mysql
release with only one node, this is probably the best choice for you. It will be relatively easy to migrate to, and the properties will mostly be familiar.
If you are using pxc for creating deployments with alternative replication strategies like leader-follower, then the mysql
job is for you.
The pxc-mysql
BOSH job runs mysql using Galera replication, across 1 or several nodes.
The typical clustered topology is 2 proxy nodes and 3 mysql nodes running the pxc-mysql
BOSH job. The proxies can be separate vms or co-located with the pxc-mysql nodes.
You can also run this topology with a single mysql node running the pxc-mysql
BOSH job and a single proxy job. In this case you would have a galera cluster of size 1, which does not provide high-availability.
The number of mysql nodes should always be odd, with a minimum count of three, to avoid split-brain. When a failed node comes back online, it will automatically rejoin the cluster and sync data from one of the healthy nodes.
Two proxy instances are recommended. The second proxy is intended to be used in a failover capacity. You can also choose to place a load balancer in front of both proxies, or use BOSH DNS to send traffic to both proxies.
In the event the first proxy fails, the second proxy will still be able to route requests to the mysql nodes.
The proxies both will route traffic to the lowest-indexed healthy galera node, according to the galera index (not bosh index).
Traffic to the MySQL cluster is routed through one or more proxy nodes. The current proxy implementation is Switchboard. This proxy acts as an intermediary between the client and the MySQL server, providing failover between MySQL nodes. The number of nodes is configured by the proxy job instance count in the deployment manifest.
NOTE: If the number of proxy nodes is set to zero, apps will be bound to the IP address of the first MySQL node in the cluster. If that IP address should change for any reason (e.g. loss of a VM) or a proxy was subsequently added, one would need to re-bind all apps to the IP address of the new node.
For more details see the proxy documentation.
The mysql
BOSH job runs standard mysql 5.7 without Galera.
Use the cf-deployment manifests with the experimental/use-pxc.yml
ops file.
<a name='deploying-clustered>
To deploy a clustered deployment, use the pxc-deployment.yml manifest and apply the use-clustered opsfile:
bosh -d <deployment> deploy --ops-file operations/use-clustered.yml pxc-deployment.yml
To deploy a standalone deployment, use the pxc-deployment.yml manifest:
bosh -d <deployment> deploy pxc-deployment.yml
Requirements:
cf-mysql-release v36.12.0 or greater
Use the cf-deployment manifests with the experimental/migrate-cf-mysql-to-pxc.yml
ops file. It is advisable to take a backup first.
⚠️ migrate-cf-mysql-to-pxc.yml
will scale down a cluster to a single node. This is required for migration. Be sure to re-set to the appropriate number of instances when switching touse-pxc.yml
subsequently.
The ops file will trigger the same migration procedure described in Using PXC release with other deployments
- If your cf-deployment uses CredHub, be sure to also include the secure-service-credentials-with-pxc-release.yml ops file.
After migrating, use the Deploying CF with pxc-release docs for your next deploy.
-
Make backups according to your usual backup procedure.
-
Get the latest pxc bosh release from bosh.io
-
Add the release to your manifest
-
⚠️ Scale down to 1 node and ensure the persistent disk has enough free space to double the size of the mysql data. -
Make the following changes to your bosh manifest:
-
Add the
pxc-mysql
job frompxc-release
to the instance group that has themysql
job fromcf-mysql-release
-
Configure the
pxc-mysql
job with the same credentials and property values as themysql
job -
To run the migration:
- Set the
cf_mysql_enabled: false
property on themysql
job - Set the
pxc_enabled: true
property onpxc-mysql
job - Switch the proxies to use the proxy job from
pxc-release
instead ofcf-mysql-release
- Deploy using BOSH
- Set the
-
To prepare for the migration, but not run it immediately:
- Set the
cf_mysql_enabled: true
property on themysql
job - Set the
pxc_enabled: false
property onpxc-mysql
job - Deploy using BOSH
- The MySQL will run as normal with only the
cf-mysql-release
running - In order to trigger the migration, redeploy with
cf_mysql_enabled: false
andpxc_enabled: true
- Set the
-
⚠️ Do not enable both releases or disable both releases. Only enable one at a time.
-
-
The migration is triggered by deploying with
cf_mysql_enabled: false
andpxc_enabled: true
. Thepre-start
script for thepxc-mysql
job inpxc-release
starts both the Mariadb MySQL from thecf-mysql-release
and the Percona MySQL frompxc-release
. The migration dumps the MariaDB MySQL and loads that data into the Percona MySQL. This is done using pipes, so the dump is not written to disk, in order to reduce the use of disk space. The MariaDB MySQL is then stopped, leaving only the Percona MySQL running.⚠️ MySQL DB will experience downtime during the migration
-
After the migration, you can optionally clean up your deployment:
- The migration will make a copy of the MySQL data on the persistent disk. To reduce disk usage, you can delete the old copy of the data in
/var/vcap/store/mysql
after you feel comfortable in the success of your migration. Do NOT delete the new copy of the data in/var/vcap/store/pxc-mysql
. - Deploy only the
pxc-release
and not thecf-mysql-release
in future deployments per Deploying new deployments, to free up disk space used by thecf-mysql-release
.
- The migration will make a copy of the MySQL data on the persistent disk. To reduce disk usage, you can delete the old copy of the data in
-
Scale back up to the recommended 3 nodes, if desired.
The Cloud Foundry team uses GitHub and accepts contributions via pull request.
Follow these steps to make a contribution to any of our open source repositories:
-
Ensure that you have completed our CLA Agreement for individuals or corporations.
-
Set your name and email (these should match the information on your submitted CLA)
git config --global user.name "Firstname Lastname" git config --global user.email "your_email@example.com"
- Fork the repository
- Check out
master
of pxc-release - Create a feature branch (
git checkout -b <my_new_branch>
) - Make changes on your branch
- Deploy your changes using pxc as the database for cf-deployment to your dev environment and run CF Acceptance Tests (CATS)
- Push to your fork (
git push origin <my_new_branch>
) and submit a pull request
We favor pull requests with very small, single commits with a single purpose.
Your pull request is much more likely to be accepted if it is small and focused with a clear message that conveys the intent of your change.