From ccd202a07aaf1e02ad75bb8192cbf5de5ae4be45 Mon Sep 17 00:00:00 2001 From: Jie Luo Date: Wed, 21 Dec 2016 14:33:37 +0800 Subject: [PATCH] Duplicated 'the' Signed-off-by: Jie Luo --- docs/admin/authentication.md | 2 +- docs/admin/daemons.md | 2 +- docs/admin/garbage-collection.md | 2 +- docs/admin/kube-controller-manager.md | 6 +++--- docs/user-guide/compute-resources.md | 2 +- 5 files changed, 7 insertions(+), 7 deletions(-) diff --git a/docs/admin/authentication.md b/docs/admin/authentication.md index ab41a6fd39..3ada61a5fd 100644 --- a/docs/admin/authentication.md +++ b/docs/admin/authentication.md @@ -31,7 +31,7 @@ to talk to the Kubernetes API. API requests are tied to either a normal user or a service account, or are treated as anonymous requests. This means every process inside or outside the cluster, from a human user typing `kubectl` on a workstation, to `kubelets` on nodes, to members -of the control plane, must authenticate when making requests to the the API server, +of the control plane, must authenticate when making requests to the API server, or be treated as an anonymous user. ## Authentication strategies diff --git a/docs/admin/daemons.md b/docs/admin/daemons.md index 7db42fe4e0..90637239b3 100644 --- a/docs/admin/daemons.md +++ b/docs/admin/daemons.md @@ -99,7 +99,7 @@ Some possible patterns for communicating with pods in a DaemonSet are: - **Push**: Pods in the Daemon Set are configured to send updates to another service, such as a stats database. They do not have clients. - **NodeIP and Known Port**: Pods in the Daemon Set use a `hostPort`, so that the pods are reachable - via the node IPs. Clients knows the the list of nodes ips somehow, and know the port by convention. + via the node IPs. Clients knows the list of nodes ips somehow, and know the port by convention. - **DNS**: Create a [headless service](/docs/user-guide/services/#headless-services) with the same pod selector, and then discover DaemonSets using the `endpoints` resource or retrieve multiple A records from DNS. diff --git a/docs/admin/garbage-collection.md b/docs/admin/garbage-collection.md index 0276596f6c..0492f9f277 100644 --- a/docs/admin/garbage-collection.md +++ b/docs/admin/garbage-collection.md @@ -17,7 +17,7 @@ kubernetes manages lifecycle of all images through imageManager, with the cooper of cadvisor. The policy for garbage collecting images takes two factors into consideration: -`HighThresholdPercent` and `LowThresholdPercent`. Disk usage above the the high threshold +`HighThresholdPercent` and `LowThresholdPercent`. Disk usage above the high threshold will trigger garbage collection. The garbage collection will delete least recently used images until the low threshold has been met. diff --git a/docs/admin/kube-controller-manager.md b/docs/admin/kube-controller-manager.md index f6f11c5f37..4b158fe4e4 100644 --- a/docs/admin/kube-controller-manager.md +++ b/docs/admin/kube-controller-manager.md @@ -45,7 +45,7 @@ kube-controller-manager --concurrent_rc_syncs int32 The number of replication controllers that are allowed to sync concurrently. Larger number = more responsive replica management, but more CPU (and network) load (default 5) --configure-cloud-routes Should CIDRs allocated by allocate-node-cidrs be configured on the cloud provider. (default true) --controller-start-interval duration Interval between starting controller managers. - --daemonset-lookup-cache-size int32 The the size of lookup cache for daemonsets. Larger number = more responsive daemonsets, but more MEM load. (default 1024) + --daemonset-lookup-cache-size int32 The size of lookup cache for daemonsets. Larger number = more responsive daemonsets, but more MEM load. (default 1024) --deployment-controller-sync-period duration Period for syncing the deployments. (default 30s) --enable-dynamic-provisioning Enable dynamic provisioning for environments that support it. (default true) --enable-garbage-collector Enables the generic garbage collector. MUST be synced with the corresponding flag of the kube-apiserver. (default true) @@ -89,8 +89,8 @@ StreamingProxyRedirects=true|false (ALPHA - default=false) --pv-recycler-pod-template-filepath-nfs string The file path to a pod definition used as a template for NFS persistent volume recycling --pv-recycler-timeout-increment-hostpath int32 the increment of time added per Gi to ActiveDeadlineSeconds for a HostPath scrubber pod. This is for development and testing only and will not work in a multi-node cluster. (default 30) --pvclaimbinder-sync-period duration The period for syncing persistent volumes and persistent volume claims (default 15s) - --replicaset-lookup-cache-size int32 The the size of lookup cache for replicatsets. Larger number = more responsive replica management, but more MEM load. (default 4096) - --replication-controller-lookup-cache-size int32 The the size of lookup cache for replication controllers. Larger number = more responsive replica management, but more MEM load. (default 4096) + --replicaset-lookup-cache-size int32 The size of lookup cache for replicatsets. Larger number = more responsive replica management, but more MEM load. (default 4096) + --replication-controller-lookup-cache-size int32 The size of lookup cache for replication controllers. Larger number = more responsive replica management, but more MEM load. (default 4096) --resource-quota-sync-period duration The period for syncing quota usage status in the system (default 5m0s) --root-ca-file string If set, this root certificate authority will be included in service account's token secret. This must be a valid PEM-encoded CA bundle. --route-reconciliation-period duration The period for reconciling routes created for Nodes by cloud provider. (default 10s) diff --git a/docs/user-guide/compute-resources.md b/docs/user-guide/compute-resources.md index 2aac91d0ba..2e524e9117 100644 --- a/docs/user-guide/compute-resources.md +++ b/docs/user-guide/compute-resources.md @@ -328,7 +328,7 @@ Host: k8s-master:8080 ``` To consume opaque resources in pods, include the name of the opaque -resource as a key in the the `spec.containers[].resources.requests` map. +resource as a key in the `spec.containers[].resources.requests` map. The pod will be scheduled only if all of the resource requests are satisfied (including cpu, memory and any opaque resources.) The pod will