2016-11-23 22:19:13 +00:00
|
|
|
---
|
2018-02-18 19:29:37 +00:00
|
|
|
reviewers:
|
2016-11-23 22:19:13 +00:00
|
|
|
- enisoc
|
|
|
|
- erictune
|
|
|
|
- foxish
|
|
|
|
- janetkuo
|
|
|
|
- kow3ns
|
|
|
|
- smarterclayton
|
2017-06-08 20:15:43 +00:00
|
|
|
title: Run a Replicated Stateful Application
|
2018-05-05 16:00:51 +00:00
|
|
|
content_template: templates/tutorial
|
2018-05-20 04:52:49 +00:00
|
|
|
weight: 30
|
2016-11-23 22:19:13 +00:00
|
|
|
---
|
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% capture overview %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
This page shows how to run a replicated stateful application using a
|
2017-10-04 04:07:11 +00:00
|
|
|
[StatefulSet](/docs/concepts/workloads/controllers/statefulset/) controller.
|
2016-11-23 22:19:13 +00:00
|
|
|
The example is a MySQL single-master topology with multiple slaves running
|
|
|
|
asynchronous replication.
|
|
|
|
|
|
|
|
Note that **this is not a production configuration**.
|
|
|
|
In particular, MySQL settings remain on insecure defaults to keep the focus
|
|
|
|
on general patterns for running stateful applications in Kubernetes.
|
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% /capture %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% capture prerequisites %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
* {{< include "task-tutorial-prereqs.md" >}} {{< version-check >}}
|
|
|
|
* {{< include "default-storage-class-prereqs.md" >}}
|
2016-11-23 22:19:13 +00:00
|
|
|
* This tutorial assumes you are familiar with
|
2017-04-19 17:56:47 +00:00
|
|
|
[PersistentVolumes](/docs/concepts/storage/persistent-volumes/)
|
2017-10-04 04:07:11 +00:00
|
|
|
and [StatefulSets](/docs/concepts/workloads/controllers/statefulset/),
|
2017-04-19 17:56:47 +00:00
|
|
|
as well as other core concepts like [Pods](/docs/concepts/workloads/pods/pod/),
|
|
|
|
[Services](/docs/concepts/services-networking/service/), and
|
2018-01-05 02:05:27 +00:00
|
|
|
[ConfigMaps](/docs/tasks/configure-pod-container/configure-pod-configmap/).
|
2016-12-02 21:27:43 +00:00
|
|
|
* Some familiarity with MySQL helps, but this tutorial aims to present
|
2016-11-23 22:19:13 +00:00
|
|
|
general patterns that should be useful for other systems.
|
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% /capture %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% capture objectives %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-11-29 18:04:17 +00:00
|
|
|
* Deploy a replicated MySQL topology with a StatefulSet controller.
|
2016-11-23 22:19:13 +00:00
|
|
|
* Send MySQL client traffic.
|
|
|
|
* Observe resistance to downtime.
|
2016-11-29 18:04:17 +00:00
|
|
|
* Scale the StatefulSet up and down.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% /capture %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% capture lessoncontent %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2017-06-08 20:15:43 +00:00
|
|
|
## Deploy MySQL
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-11-29 18:04:17 +00:00
|
|
|
The example MySQL deployment consists of a ConfigMap, two Services,
|
|
|
|
and a StatefulSet.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
### ConfigMap
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
Create the ConfigMap from the following YAML configuration file:
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
```shell
|
2017-05-18 03:14:09 +00:00
|
|
|
kubectl create -f https://k8s.io/docs/tasks/run-application/mysql-configmap.yaml
|
2016-11-23 22:19:13 +00:00
|
|
|
```
|
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{< code file="mysql-configmap.yaml" >}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-11-29 18:04:17 +00:00
|
|
|
This ConfigMap provides `my.cnf` overrides that let you independently control
|
2016-12-02 21:27:43 +00:00
|
|
|
configuration on the MySQL master and slaves.
|
2016-11-23 22:19:13 +00:00
|
|
|
In this case, you want the master to be able to serve replication logs to slaves
|
|
|
|
and you want slaves to reject any writes that don't come via replication.
|
|
|
|
|
|
|
|
There's nothing special about the ConfigMap itself that causes different
|
|
|
|
portions to apply to different Pods.
|
2016-12-02 21:27:43 +00:00
|
|
|
Each Pod decides which portion to look at as it's initializing,
|
2016-11-29 18:04:17 +00:00
|
|
|
based on information provided by the StatefulSet controller.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
### Services
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
Create the Services from the following YAML configuration file:
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
```shell
|
2017-05-18 03:14:09 +00:00
|
|
|
kubectl create -f https://k8s.io/docs/tasks/run-application/mysql-services.yaml
|
2016-11-23 22:19:13 +00:00
|
|
|
```
|
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{< code file="mysql-services.yaml" >}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-11-29 18:04:17 +00:00
|
|
|
The Headless Service provides a home for the DNS entries that the StatefulSet
|
2016-12-02 21:27:43 +00:00
|
|
|
controller creates for each Pod that's part of the set.
|
|
|
|
Because the Headless Service is named `mysql`, the Pods are accessible by
|
2016-11-23 22:19:13 +00:00
|
|
|
resolving `<pod-name>.mysql` from within any other Pod in the same Kubernetes
|
|
|
|
cluster and namespace.
|
|
|
|
|
|
|
|
The Client Service, called `mysql-read`, is a normal Service with its own
|
2016-12-02 21:27:43 +00:00
|
|
|
cluster IP that distributes connections across all MySQL Pods that report
|
|
|
|
being Ready. The set of potential endpoints includes the MySQL master and all
|
|
|
|
slaves.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
Note that only read queries can use the load-balanced Client Service.
|
2016-12-02 21:27:43 +00:00
|
|
|
Because there is only one MySQL master, clients should connect directly to the
|
|
|
|
MySQL master Pod (through its DNS entry within the Headless Service) to execute
|
|
|
|
writes.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
### StatefulSet
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
Finally, create the StatefulSet from the following YAML configuration file:
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
```shell
|
2017-05-18 03:14:09 +00:00
|
|
|
kubectl create -f https://k8s.io/docs/tasks/run-application/mysql-statefulset.yaml
|
2016-11-23 22:19:13 +00:00
|
|
|
```
|
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{< code file="mysql-statefulset.yaml" >}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
You can watch the startup progress by running:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl get pods -l app=mysql --watch
|
|
|
|
```
|
|
|
|
|
|
|
|
After a while, you should see all 3 Pods become Running:
|
|
|
|
|
|
|
|
```
|
|
|
|
NAME READY STATUS RESTARTS AGE
|
|
|
|
mysql-0 2/2 Running 0 2m
|
|
|
|
mysql-1 2/2 Running 0 1m
|
|
|
|
mysql-2 2/2 Running 0 1m
|
|
|
|
```
|
|
|
|
|
|
|
|
Press **Ctrl+C** to cancel the watch.
|
2016-11-29 18:04:17 +00:00
|
|
|
If you don't see any progress, make sure you have a dynamic PersistentVolume
|
2016-11-23 22:19:13 +00:00
|
|
|
provisioner enabled as mentioned in the [prerequisites](#before-you-begin).
|
|
|
|
|
|
|
|
This manifest uses a variety of techniques for managing stateful Pods as part of
|
2016-11-29 18:04:17 +00:00
|
|
|
a StatefulSet. The next section highlights some of these techniques to explain
|
|
|
|
what happens as the StatefulSet creates Pods.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
## Understanding stateful Pod initialization
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-11-29 18:04:17 +00:00
|
|
|
The StatefulSet controller starts Pods one at a time, in order by their
|
2016-11-23 22:19:13 +00:00
|
|
|
ordinal index.
|
|
|
|
It waits until each Pod reports being Ready before starting the next one.
|
|
|
|
|
|
|
|
In addition, the controller assigns each Pod a unique, stable name of the form
|
|
|
|
`<statefulset-name>-<ordinal-index>`.
|
|
|
|
In this case, that results in Pods named `mysql-0`, `mysql-1`, and `mysql-2`.
|
|
|
|
|
2016-11-29 18:04:17 +00:00
|
|
|
The Pod template in the above StatefulSet manifest takes advantage of these
|
2016-11-23 22:19:13 +00:00
|
|
|
properties to perform orderly startup of MySQL replication.
|
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
### Generating configuration
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
Before starting any of the containers in the Pod spec, the Pod first runs any
|
2017-09-25 06:45:50 +00:00
|
|
|
[Init Containers](/docs/concepts/workloads/pods/init-containers/)
|
2016-11-23 22:19:13 +00:00
|
|
|
in the order defined.
|
|
|
|
|
|
|
|
The first Init Container, named `init-mysql`, generates special MySQL config
|
|
|
|
files based on the ordinal index.
|
|
|
|
|
|
|
|
The script determines its own ordinal index by extracting it from the end of
|
|
|
|
the Pod name, which is returned by the `hostname` command.
|
|
|
|
Then it saves the ordinal (with a numeric offset to avoid reserved values)
|
|
|
|
into a file called `server-id.cnf` in the MySQL `conf.d` directory.
|
2016-11-29 18:04:17 +00:00
|
|
|
This translates the unique, stable identity provided by the StatefulSet
|
2016-11-23 22:19:13 +00:00
|
|
|
controller into the domain of MySQL server IDs, which require the same
|
|
|
|
properties.
|
|
|
|
|
|
|
|
The script in the `init-mysql` container also applies either `master.cnf` or
|
2016-11-29 18:04:17 +00:00
|
|
|
`slave.cnf` from the ConfigMap by copying the contents into `conf.d`.
|
2016-12-02 21:27:43 +00:00
|
|
|
Because the example topology consists of a single MySQL master and any number of
|
2016-11-29 18:25:09 +00:00
|
|
|
slaves, the script simply assigns ordinal `0` to be the master, and everyone
|
|
|
|
else to be slaves.
|
2016-11-29 18:40:15 +00:00
|
|
|
Combined with the StatefulSet controller's
|
2017-09-25 06:45:50 +00:00
|
|
|
[deployment order guarantee](/docs/concepts/workloads/controllers/statefulset/#deployment-and-scaling-guarantees/),
|
2016-12-02 21:27:43 +00:00
|
|
|
this ensures the MySQL master is Ready before creating slaves, so they can begin
|
2016-11-29 18:40:15 +00:00
|
|
|
replicating.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
### Cloning existing data
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
In general, when a new Pod joins the set as a slave, it must assume the MySQL
|
|
|
|
master might already have data on it. It also must assume that the replication
|
|
|
|
logs might not go all the way back to the beginning of time.
|
2016-12-23 07:26:12 +00:00
|
|
|
These conservative assumptions are the key to allow a running StatefulSet
|
2016-11-23 22:19:13 +00:00
|
|
|
to scale up and down over time, rather than being fixed at its initial size.
|
|
|
|
|
|
|
|
The second Init Container, named `clone-mysql`, performs a clone operation on
|
2016-11-29 18:04:17 +00:00
|
|
|
a slave Pod the first time it starts up on an empty PersistentVolume.
|
2016-11-23 22:19:13 +00:00
|
|
|
That means it copies all existing data from another running Pod,
|
|
|
|
so its local state is consistent enough to begin replicating from the master.
|
|
|
|
|
|
|
|
MySQL itself does not provide a mechanism to do this, so the example uses a
|
|
|
|
popular open-source tool called Percona XtraBackup.
|
2016-11-29 18:25:09 +00:00
|
|
|
During the clone, the source MySQL server might suffer reduced performance.
|
2016-12-02 21:27:43 +00:00
|
|
|
To minimize impact on the MySQL master, the script instructs each Pod to clone
|
|
|
|
from the Pod whose ordinal index is one lower.
|
|
|
|
This works because the StatefulSet controller always ensures Pod `N` is
|
2016-11-23 22:19:13 +00:00
|
|
|
Ready before starting Pod `N+1`.
|
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
### Starting replication
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
After the Init Containers complete successfully, the regular containers run.
|
|
|
|
The MySQL Pods consist of a `mysql` container that runs the actual `mysqld`
|
|
|
|
server, and an `xtrabackup` container that acts as a
|
2018-06-15 17:19:29 +00:00
|
|
|
[sidecar](https://kubernetes.io/blog/2015/06/the-distributed-system-toolkit-patterns).
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
The `xtrabackup` sidecar looks at the cloned data files and determines if
|
|
|
|
it's necessary to initialize MySQL replication on the slave.
|
|
|
|
If so, it waits for `mysqld` to be ready and then executes the
|
|
|
|
`CHANGE MASTER TO` and `START SLAVE` commands with replication parameters
|
|
|
|
extracted from the XtraBackup clone files.
|
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
Once a slave begins replication, it remembers its MySQL master and
|
|
|
|
reconnects automatically if the server restarts or the connection dies.
|
2016-11-29 18:25:09 +00:00
|
|
|
Also, because slaves look for the master at its stable DNS name
|
2016-12-02 21:27:43 +00:00
|
|
|
(`mysql-0.mysql`), they automatically find the master even if it gets a new
|
2016-11-29 18:25:09 +00:00
|
|
|
Pod IP due to being rescheduled.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
Lastly, after starting replication, the `xtrabackup` container listens for
|
|
|
|
connections from other Pods requesting a data clone.
|
2016-11-29 18:04:17 +00:00
|
|
|
This server remains up indefinitely in case the StatefulSet scales up, or in
|
|
|
|
case the next Pod loses its PersistentVolumeClaim and needs to redo the clone.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
## Sending client traffic
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
You can send test queries to the MySQL master (hostname `mysql-0.mysql`)
|
2016-11-23 22:19:13 +00:00
|
|
|
by running a temporary container with the `mysql:5.7` image and running the
|
|
|
|
`mysql` client binary.
|
|
|
|
|
|
|
|
```shell
|
Release 1.8 (#5659)
* GC now supports non-core resources
* Add two examples about how to analysis audits of kube-apiserver (#4264)
* Deprecate system:nodes binding
* [1.8] StatefulSet `initialized` annotation is now ignored.
* inits the kubeadm upgrade docs
addresses kubernetes/kubernetes.github.io/issues/4689
* adds kubeadm upgrade cmd to ToC
addresses kubernetes/kubernetes.github.io/issues/4689
* add workload placement docs
* ScaleIO - document udpate for 1.8
* Add documentation on storageClass.mountOptions and PV.mountOptions (#5254)
* Add documentation on storageClass.mountOptions and PV.mountOptions
* convert notes into callouts
* Add docs for CustomResource validation
add info about supported fields
* advanced audit beta features (#5300)
* Update job workload doc with backoff failure policy (#5319)
Add to the Jobs documentation how to use the new backoffLimit field that
limit the number of Pod failure before considering the Job as failed.
* Documented additional AWS Service annotations (#4864)
* Add device plugin doc under concepts/cluster-administration. (#5261)
* Add device plugin doc under concepts/cluster-administration.
* Update device-plugins.md
* Update device-plugins.md
Add meta description. Fix typo. Change bare metal deployment to manual deployment.
* Update device-plugins.md
Fix typo again.
* Update page.version. (#5341)
* Add documentation on storageClass.reclaimPolicy (#5171)
* [Advanced audit] use new herf for audit-api (#5349)
This tag contains all the changes in v1beta1 version. Update it now.
* Added documentation around creating the InitializerConfiguration for the persistent volume label controller in the cloud-controller-manager (#5255)
* Documentation for kubectl plugins (#5294)
* Documentation for kubectl plugins
* Update kubectl-plugins.md
* Update kubectl-plugins.md
* Updated CPU manager docs to match implementation. (#5332)
* Noted limitation of alpha static cpumanager.
* Updated CPU manager docs to match implementation.
- Removed references to CPU pressure node condition and evictions.
- Added note about new --cpu-manager-reconcile-period flag.
- Added note about node allocatable requirements for static policy.
- Noted limitation of alpha static cpumanager.
* Move cpu-manager task link to rsc mgmt section.
* init containers annotation removed in 1.8 (#5390)
* Add documentation for TaintNodesByCondition (#5352)
* Add documentation for TaintNodesByCondition
* Update nodes.md
* Update taint-and-toleration.md
* Update daemonset.md
* Update nodes.md
* Update taint-and-toleration.md
* Update daemonset.md
* Fix deployments (#5421)
* Document extended resources and OIR deprecation. (#5399)
* Document extended resources and OIR deprecation.
* Updated extended resources doc per reviews.
* reverts extra spacing in _data/tasks.yml
* addresses `kubeadm upgrade` review comments
Feedback from @chenopis, @luxas, and @steveperry-53 addressed with this commit
* HugePages documentation (#5419)
* Update cpu-management-policies.md (#5407)
Fixed the bad link.
Modified "cpu" to "CPU".
Added more 'yaml' as supplement.
* Update RBAC docs for v1 (#5445)
* Add user docs for pod priority and preemption (#5328)
* Add user docs for pod priority and preemption
* Update pod-priority-preemption.md
* More updates
* Update docs/admin/kubeadm.md for 1.8 (#5440)
- Made a couple of minor wording changes (not strictly 1.8 related).
- Did some reformatting (not strictly 1.8 related).
- Updated references to the default token TTL (was infinite, now 24 hours).
- Documented the new `--discovery-token-ca-cert-hash` and `--discovery-token-unsafe-skip-ca-verification` flags for `kubeadm join`.
- Added references to the new `--discovery-token-ca-cert-hash` flag in all the default examples.
- Added a new _Security model_ section that describes the security tradeoffs of the various discovery modes.
- Documented the new `--groups` flag for `kubeadm token create`.
- Added a note of caution under _Automating kubeadm_ that references the _Security model_ section.
- Updated the component version table to drop 1.6 and add 1.8.
- Update `_data/reference.yml` to try to get the sidebar fixed up and more consistent with `kubefed`.
* Update StatefulSet Basics for 1.8 release (#5398)
* addresses `kubeadm upgrade` review comments
2nd iteration review comments by @luxas
* adds kubelet upgrade section to kubeadm upgrade
* Fix a bulleted list on docs/admin/kubeadm.md. (#5458)
I updated this doc yesterday and I was absolutely sure I fixed this, but I just saw that this commit got lost somehow.
This was introduced recently in https://github.com/kubernetes/kubernetes.github.io/pull/5440.
* Clarify the API to check for device plugins
* Moving Flexvolume to separate out-of-tree section
* addresses `kubeadm upgrade` review comments
CC: @luxas
* fixes kubeadm upgrade index
* Update Stackdriver Logging documentation (#5495)
* Re-update WordPress and MySQL PV doc to use apps/v1beta2 APIs (#5526)
* Update statefulset concepts doc to use apps/v1beta2 APIs (#5420)
* add document on kubectl's behavior regarding initializers (#5505)
* Update docs/admin/kubeadm.md to cover self-hosting in 1.8. (#5497)
This is a new beta feature in 1.8.
* Update kubectl patch doc to use apps/v1beta2 APIs (#5422)
* [1.8] Update "Run Applications" tasks to apps/v1beta2. (#5525)
* Update replicated stateful application task for 1.8.
* Update single instance stateful app task for 1.8.
* Update stateless app task for 1.8.
* Update kubectl patch task for 1.8.
* fix the link of persistent storage (#5515)
* update the admission-controllers.md index.md what-is-kubernetes.md link
* fix the link of persistent storage
* Add quota support for local ephemeral storage (#5493)
* Add quota support for local ephemeral storage
update the doc to this alpha feature
* Update resource-quotas.md
* Updated Deployments concepts doc (#5491)
* Updated Deployments concepts doc
* Addressed comments
* Addressed more comments
* Modify allocatable storage to ephemeral-storage (#5490)
Update the doc to use ephemeral-storage instead of storage
* Revamped concepts doc for ReplicaSet (#5463)
* Revamped concepts doc for ReplicaSet
* Minor changes to call out specific versions for selector defaulting and
immutability
* Addressed doc review comments
* Remove petset documentations (#5395)
* Update docs to use batch/v1beta1 cronjobs (#5475)
* add federation job doc (#5485)
* add federation job doc
* Update job.md
Edits for clarity and consistency
* Update job.md
Fixed a typo
* update DaemonSet concept for 1.8 release (#5397)
* update DaemonSet concept for 1.8 release
* Update daemonset.md
Fix typo. than -> then
* Update bootstrap tokens doc for 1.8. (#5479)
* Update bootstrap tokens doc for 1.8.
This has some changes I missed when I was updating the main kubeadm documention:
- Bootstrap tokens are now beta, not alpha (https://github.com/kubernetes/features/issues/130)
- The apiserver flag to enable the authenticator changedin 1.8 (https://github.com/kubernetes/kubernetes/pull/51198)
- Added `auth-extra-groups` documentaion (https://github.com/kubernetes/kubernetes/pull/50933)
- Updated the _Token Management with `kubeadm`_ section to link to the main kubeadm docs, since it was just duplicated information.
* Update bootstrap-tokens.md
* Updated the Cassandra tutorial to use apps/v1beta2 (#5548)
* add docs for AllowPrivilegeEscalation (#5448)
Signed-off-by: Jess Frazelle <acidburn@microsoft.com>
* Add local ephemeral storage alpha feature in managing compute resource (#5522)
* Add local ephemeral storage alpha feature in managing compute resource
Since 1.8, we add the local ephemeral storage alpha feature as one
resource type to manage. Add this feature into the doc.
* Update manage-compute-resources-container.md
* Update manage-compute-resources-container.md
* Update manage-compute-resources-container.md
* Update manage-compute-resources-container.md
* Update manage-compute-resources-container.md
* Update manage-compute-resources-container.md
* Added documentation for Metrics Server (#5560)
* authorization: improve authorization debugging docs (#5549)
* Document mount propagation (#5544)
* Update /docs/setup/independent/create-cluster-kubeadm.md for 1.8. (#5524)
This introduction needed a couple of small tweaks to cover the `--discovery-token-ca-cert-hash` flag added in https://github.com/kubernetes/kubernetes/pull/49520 and some version bumps.
* Add task doc for alpha dynamic kubelet configuration (#5523)
* Fix input/output of selfsubjectaccess review (#5593)
* Add docs for implementing resize (#5528)
* Add docs for implementing resize
* Update admission-controllers.md
* Added link to PVC section
* minor typo fixes
* Update NetworkPolicy concept guide with egress and CIDR changes (#5529)
* update zookeeper tutorial for 1.8 release
* add doc for hostpath type (#5503)
* Federated Hpa feature doc (#5487)
* Federated Hpa feature doc
* Federated Hpa feature doc review fixes
* Update hpa.md
* Update hpa.md
* update cloud controller manager docs for v1.8
* Update cronjob with defaults information (#5556)
* Kubernetes 1.8 reference docs (#5632)
* Kubernetes 1.8 reference docs
* Kubectl reference docs for 1.8
* Update side bar with 1.8 kubectl and api ref docs links
* remove petset.md
* update on state of HostAlias in 1.8 with hostNetwork Pod support (#5644)
* Fix cron job deletion section (#5655)
* update imported docs (#5656)
* Add documentation for certificate rotation. (#5639)
* Link to using kubeadm page
* fix the command output
fix the command output
* fix typo in api/resources reference: "Worloads"
* Add documentation for certificate rotation.
* Create TOC entry for cloud controller manager. (#5662)
* Updates for new versions of API types
* Followup 5655: fix link to garbage collection (#5666)
* Temporarily redirect resources-reference to api-reference. (#5668)
* Update config for 1.8 release. (#5661)
* Update config for 1.8 release.
* Address reviewer comments.
* Switch references in HPA docs from alpha to beta (#5671)
The HPA docs still referenced the alpha version. This switches them to
talk about v2beta1, which is the appropriate version for Kubernetes 1.8
* Deprecate openstack heat (#5670)
* Fix typo in pod preset conflict example
Move container port definition to the correct line.
* Highlight openstack-heat provider deprecation
The openstack-heat provider for kube-up is being deprecated and will be
removed in a future release.
* Temporarily fix broken links by redirecting. (#5672)
* Fix broken links. (#5675)
* Fix render of code block (#5674)
* Fix broken links. (#5677)
* Add a small note about auto-bootstrapped CSR ClusterRoles (#5660)
* Update kubeadm install doc for v1.8 (#5676)
* add draft workloads api content for 1.8 (#5650)
* add draft workloads api content for 1.8
* edits per review, add tables, for 1.8 workloads api doc
* fix typo
* Minor fixes to kubeadm 1.8 upgrade guide. (#5678)
- The kubelet upgrade instructions should be done on every host, not
just worker nodes.
- We should just upgrade all packages, instead of calling out kubelet
specifically. This will also upgrade kubectl, kubeadm, and
kubernetes-cni, if installed.
- Draining nodes should also ignore daemonsets, and master errors can be
ignored.
- Make sure that the new kubeadm download is chmoded correctly.
- Add a step to run `kubeadm version` to verify after downloading.
- Manually approve new kubelet CSRs if rotation is enabled (known issue).
* Release 1.8 (#5680)
* Fix versions for 1.8 API ref docs
* Updates for 1.8 kubectl reference docs
* Kubeadm /docs/admin/kubeadm.md cleanup, editing. (#5681)
* Update docs/admin/kubeadm.md (mostly 1.8 related).
This is Fabrizio's work, which I'm committing along with my edits (in a commit on top of this).
* A few of my own edits to clarify and clean up some Markdown.
2017-09-29 04:46:51 +00:00
|
|
|
kubectl run mysql-client --image=mysql:5.7 -i --rm --restart=Never --\
|
2016-11-23 22:19:13 +00:00
|
|
|
mysql -h mysql-0.mysql <<EOF
|
|
|
|
CREATE DATABASE test;
|
|
|
|
CREATE TABLE test.messages (message VARCHAR(250));
|
|
|
|
INSERT INTO test.messages VALUES ('hello');
|
|
|
|
EOF
|
|
|
|
```
|
|
|
|
|
|
|
|
Use the hostname `mysql-read` to send test queries to any server that reports
|
|
|
|
being Ready:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl run mysql-client --image=mysql:5.7 -i -t --rm --restart=Never --\
|
|
|
|
mysql -h mysql-read -e "SELECT * FROM test.messages"
|
|
|
|
```
|
|
|
|
|
|
|
|
You should get output like this:
|
|
|
|
|
|
|
|
```
|
|
|
|
Waiting for pod default/mysql-client to be running, status is Pending, pod ready: false
|
|
|
|
+---------+
|
|
|
|
| message |
|
|
|
|
+---------+
|
|
|
|
| hello |
|
|
|
|
+---------+
|
|
|
|
pod "mysql-client" deleted
|
|
|
|
```
|
|
|
|
|
|
|
|
To demonstrate that the `mysql-read` Service distributes connections across
|
|
|
|
servers, you can run `SELECT @@server_id` in a loop:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl run mysql-client-loop --image=mysql:5.7 -i -t --rm --restart=Never --\
|
|
|
|
bash -ic "while sleep 1; do mysql -h mysql-read -e 'SELECT @@server_id,NOW()'; done"
|
|
|
|
```
|
|
|
|
|
2016-11-29 18:25:09 +00:00
|
|
|
You should see the reported `@@server_id` change randomly, because a different
|
|
|
|
endpoint might be selected upon each connection attempt:
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
```
|
|
|
|
+-------------+---------------------+
|
|
|
|
| @@server_id | NOW() |
|
|
|
|
+-------------+---------------------+
|
|
|
|
| 100 | 2006-01-02 15:04:05 |
|
|
|
|
+-------------+---------------------+
|
|
|
|
+-------------+---------------------+
|
|
|
|
| @@server_id | NOW() |
|
|
|
|
+-------------+---------------------+
|
|
|
|
| 102 | 2006-01-02 15:04:06 |
|
|
|
|
+-------------+---------------------+
|
|
|
|
+-------------+---------------------+
|
|
|
|
| @@server_id | NOW() |
|
|
|
|
+-------------+---------------------+
|
|
|
|
| 101 | 2006-01-02 15:04:07 |
|
|
|
|
+-------------+---------------------+
|
|
|
|
```
|
|
|
|
|
|
|
|
You can press **Ctrl+C** when you want to stop the loop, but it's useful to keep
|
|
|
|
it running in another window so you can see the effects of the following steps.
|
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
## Simulating Pod and Node downtime
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
To demonstrate the increased availability of reading from the pool of slaves
|
|
|
|
instead of a single server, keep the `SELECT @@server_id` loop from above
|
|
|
|
running while you force a Pod out of the Ready state.
|
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
### Break the Readiness Probe
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2017-09-25 06:45:50 +00:00
|
|
|
The [readiness probe](/docs/tasks/configure-pod-container/configure-liveness-readiness-probes/#define-readiness-probes)
|
2016-11-23 22:19:13 +00:00
|
|
|
for the `mysql` container runs the command `mysql -h 127.0.0.1 -e 'SELECT 1'`
|
|
|
|
to make sure the server is up and able to execute queries.
|
|
|
|
|
|
|
|
One way to force this readiness probe to fail is to break that command:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl exec mysql-2 -c mysql -- mv /usr/bin/mysql /usr/bin/mysql.off
|
|
|
|
```
|
|
|
|
|
|
|
|
This reaches into the actual container's filesystem for Pod `mysql-2` and
|
|
|
|
renames the `mysql` command so the readiness probe can't find it.
|
|
|
|
After a few seconds, the Pod should report one of its containers as not Ready,
|
|
|
|
which you can check by running:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl get pod mysql-2
|
|
|
|
```
|
|
|
|
|
|
|
|
Look for `1/2` in the `READY` column:
|
|
|
|
|
|
|
|
```
|
|
|
|
NAME READY STATUS RESTARTS AGE
|
|
|
|
mysql-2 1/2 Running 0 3m
|
|
|
|
```
|
|
|
|
|
|
|
|
At this point, you should see your `SELECT @@server_id` loop continue to run,
|
|
|
|
although it never reports `102` anymore.
|
|
|
|
Recall that the `init-mysql` script defined `server-id` as `100 + $ordinal`,
|
|
|
|
so server ID `102` corresponds to Pod `mysql-2`.
|
|
|
|
|
|
|
|
Now repair the Pod and it should reappear in the loop output
|
|
|
|
after a few seconds:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl exec mysql-2 -c mysql -- mv /usr/bin/mysql.off /usr/bin/mysql
|
|
|
|
```
|
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
### Delete Pods
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
The StatefulSet also recreates Pods if they're deleted, similar to what a
|
2016-11-29 18:04:17 +00:00
|
|
|
ReplicaSet does for stateless Pods.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl delete pod mysql-2
|
|
|
|
```
|
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
The StatefulSet controller notices that no `mysql-2` Pod exists anymore,
|
|
|
|
and creates a new one with the same name and linked to the same
|
2016-11-29 18:04:17 +00:00
|
|
|
PersistentVolumeClaim.
|
2016-11-23 22:19:13 +00:00
|
|
|
You should see server ID `102` disappear from the loop output for a while
|
|
|
|
and then return on its own.
|
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
### Drain a Node
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
If your Kubernetes cluster has multiple Nodes, you can simulate Node downtime
|
|
|
|
(such as when Nodes are upgraded) by issuing a
|
2018-04-27 22:02:19 +00:00
|
|
|
[drain](/docs/reference/generated/kubectl/kubectl-commands/#drain).
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
First determine which Node one of the MySQL Pods is on:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl get pod mysql-2 -o wide
|
|
|
|
```
|
|
|
|
|
|
|
|
The Node name should show up in the last column:
|
|
|
|
|
|
|
|
```
|
|
|
|
NAME READY STATUS RESTARTS AGE IP NODE
|
|
|
|
mysql-2 2/2 Running 0 15m 10.244.5.27 kubernetes-minion-group-9l2t
|
|
|
|
```
|
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
Then drain the Node by running the following command, which cordons it so
|
|
|
|
no new Pods may schedule there, and then evicts any existing Pods.
|
2016-11-23 22:19:13 +00:00
|
|
|
Replace `<node-name>` with the name of the Node you found in the last step.
|
|
|
|
|
2016-11-29 18:25:09 +00:00
|
|
|
This might impact other applications on the Node, so it's best to
|
2016-11-23 22:19:13 +00:00
|
|
|
**only do this in a test cluster**.
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl drain <node-name> --force --delete-local-data --ignore-daemonsets
|
|
|
|
```
|
|
|
|
|
|
|
|
Now you can watch as the Pod reschedules on a different Node:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl get pod mysql-2 -o wide --watch
|
|
|
|
```
|
|
|
|
|
|
|
|
It should look something like this:
|
|
|
|
|
|
|
|
```
|
|
|
|
NAME READY STATUS RESTARTS AGE IP NODE
|
|
|
|
mysql-2 2/2 Terminating 0 15m 10.244.1.56 kubernetes-minion-group-9l2t
|
|
|
|
[...]
|
|
|
|
mysql-2 0/2 Pending 0 0s <none> kubernetes-minion-group-fjlm
|
|
|
|
mysql-2 0/2 Init:0/2 0 0s <none> kubernetes-minion-group-fjlm
|
|
|
|
mysql-2 0/2 Init:1/2 0 20s 10.244.5.32 kubernetes-minion-group-fjlm
|
|
|
|
mysql-2 0/2 PodInitializing 0 21s 10.244.5.32 kubernetes-minion-group-fjlm
|
|
|
|
mysql-2 1/2 Running 0 22s 10.244.5.32 kubernetes-minion-group-fjlm
|
|
|
|
mysql-2 2/2 Running 0 30s 10.244.5.32 kubernetes-minion-group-fjlm
|
|
|
|
```
|
|
|
|
|
|
|
|
And again, you should see server ID `102` disappear from the
|
|
|
|
`SELECT @@server_id` loop output for a while and then return.
|
|
|
|
|
|
|
|
Now uncordon the Node to return it to a normal state:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl uncordon <node-name>
|
|
|
|
```
|
|
|
|
|
2017-01-18 18:18:37 +00:00
|
|
|
## Scaling the number of slaves
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
With MySQL replication, you can scale your read query capacity by adding slaves.
|
2016-11-29 18:04:17 +00:00
|
|
|
With StatefulSet, you can do this with a single command:
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
```shell
|
2017-09-25 06:45:50 +00:00
|
|
|
kubectl scale statefulset mysql --replicas=5
|
2016-11-23 22:19:13 +00:00
|
|
|
```
|
|
|
|
|
|
|
|
Watch the new Pods come up by running:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl get pods -l app=mysql --watch
|
|
|
|
```
|
|
|
|
|
|
|
|
Once they're up, you should see server IDs `103` and `104` start appearing in
|
|
|
|
the `SELECT @@server_id` loop output.
|
|
|
|
|
|
|
|
You can also verify that these new servers have the data you added before they
|
|
|
|
existed:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl run mysql-client --image=mysql:5.7 -i -t --rm --restart=Never --\
|
|
|
|
mysql -h mysql-3.mysql -e "SELECT * FROM test.messages"
|
|
|
|
```
|
|
|
|
|
|
|
|
```
|
|
|
|
Waiting for pod default/mysql-client to be running, status is Pending, pod ready: false
|
|
|
|
+---------+
|
|
|
|
| message |
|
|
|
|
+---------+
|
|
|
|
| hello |
|
|
|
|
+---------+
|
|
|
|
pod "mysql-client" deleted
|
|
|
|
```
|
|
|
|
|
|
|
|
Scaling back down is also seamless:
|
|
|
|
|
|
|
|
```shell
|
2017-09-25 06:45:50 +00:00
|
|
|
kubectl scale statefulset mysql --replicas=3
|
2016-11-23 22:19:13 +00:00
|
|
|
```
|
|
|
|
|
2016-11-29 18:04:17 +00:00
|
|
|
Note, however, that while scaling up creates new PersistentVolumeClaims
|
2016-11-23 22:19:13 +00:00
|
|
|
automatically, scaling down does not automatically delete these PVCs.
|
|
|
|
This gives you the choice to keep those initialized PVCs around to make
|
|
|
|
scaling back up quicker, or to extract data before deleting them.
|
|
|
|
|
|
|
|
You can see this by running:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl get pvc -l app=mysql
|
|
|
|
```
|
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
Which shows that all 5 PVCs still exist, despite having scaled the
|
2016-11-29 18:04:17 +00:00
|
|
|
StatefulSet down to 3:
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
```
|
|
|
|
NAME STATUS VOLUME CAPACITY ACCESSMODES AGE
|
|
|
|
data-mysql-0 Bound pvc-8acbf5dc-b103-11e6-93fa-42010a800002 10Gi RWO 20m
|
|
|
|
data-mysql-1 Bound pvc-8ad39820-b103-11e6-93fa-42010a800002 10Gi RWO 20m
|
|
|
|
data-mysql-2 Bound pvc-8ad69a6d-b103-11e6-93fa-42010a800002 10Gi RWO 20m
|
|
|
|
data-mysql-3 Bound pvc-50043c45-b1c5-11e6-93fa-42010a800002 10Gi RWO 2m
|
|
|
|
data-mysql-4 Bound pvc-500a9957-b1c5-11e6-93fa-42010a800002 10Gi RWO 2m
|
|
|
|
```
|
|
|
|
|
|
|
|
If you don't intend to reuse the extra PVCs, you can delete them:
|
|
|
|
|
|
|
|
```shell
|
|
|
|
kubectl delete pvc data-mysql-3
|
|
|
|
kubectl delete pvc data-mysql-4
|
|
|
|
```
|
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% /capture %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% capture cleanup %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
1. Cancel the `SELECT @@server_id` loop by pressing **Ctrl+C** in its terminal,
|
|
|
|
or running the following from another terminal:
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
```shell
|
|
|
|
kubectl delete pod mysql-client-loop --now
|
|
|
|
```
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
1. Delete the StatefulSet. This also begins terminating the Pods.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
```shell
|
|
|
|
kubectl delete statefulset mysql
|
|
|
|
```
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
1. Verify that the Pods disappear.
|
|
|
|
They might take some time to finish terminating.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
```shell
|
|
|
|
kubectl get pods -l app=mysql
|
|
|
|
```
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
You'll know the Pods have terminated when the above returns:
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
```
|
|
|
|
No resources found.
|
|
|
|
```
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
1. Delete the ConfigMap, Services, and PersistentVolumeClaims.
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
```shell
|
|
|
|
kubectl delete configmap,service,pvc -l app=mysql
|
|
|
|
```
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2016-12-02 21:27:43 +00:00
|
|
|
1. If you manually provisioned PersistentVolumes, you also need to manually
|
|
|
|
delete them, as well as release the underlying resources.
|
|
|
|
If you used a dynamic provisioner, it automatically deletes the
|
|
|
|
PersistentVolumes when it sees that you deleted the PersistentVolumeClaims.
|
|
|
|
Some dynamic provisioners (such as those for EBS and PD) also release the
|
|
|
|
underlying resources upon deleting the PersistentVolumes.
|
2016-11-29 18:40:15 +00:00
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% /capture %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% capture whatsnext %}}
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
* Look in the [Helm Charts repository](https://github.com/kubernetes/charts)
|
|
|
|
for other stateful application examples.
|
|
|
|
|
2018-05-05 16:00:51 +00:00
|
|
|
{{% /capture %}}
|
|
|
|
|
2016-11-23 22:19:13 +00:00
|
|
|
|
|
|
|
|