Merge pull request #43961 from yanggangtony/fix-some-typos
Fix some typos in [en] docs.pull/44055/head
commit
eb185751a7
|
@ -20,7 +20,7 @@ Real time visualization is a strength that UI’s have over CLI’s, and with 1.
|
|||
Based on user research with Kubernetes’ predecessor [Borg](http://research.google.com/pubs/pub43438.html) and continued community feedback, we know logs are tremendously important to users. For this reason we’re constantly looking for ways to improve these features in Dashboard. This release includes a fix for an issue wherein large numbers of logs would crash the system, as well as the introduction of the ability to view logs by date.
|
||||
|
||||
**Showing More Resources**
|
||||
The previous release brought all workloads to Dashboard: Pods, Pet Sets, Daemon Sets, Replication Controllers, Replica Set, Services, & Deployments. With 1.4, we expand upon that set of objects by including Services, Ingresses, Persistent Volume Claims, Secrets, & Config Maps. We’ve also introduced an “Admin” section with the Namespace-independent global objects of Namespaces, Nodes, and Persistent Volumes. With the addition of roles, these will be shown only to cluster operators, and developers’ side nav will begin with the Namespace dropdown.
|
||||
The previous release brought all workloads to Dashboard: Pods, Pet Sets, Daemon Sets, Replication Controllers, Replica Set, Services, & Deployments. With 1.4, we expand upon that set of objects by including Services, Ingresses, Persistent Volume Claims, Secrets, & ConfigMaps. We’ve also introduced an “Admin” section with the Namespace-independent global objects of Namespaces, Nodes, and Persistent Volumes. With the addition of roles, these will be shown only to cluster operators, and developers’ side nav will begin with the Namespace dropdown.
|
||||
|
||||
Like glue binding together a loose stack of papers into a book, we needed some way to impose order on these resources for their value to be realized, so one of the features we’re most excited to announce in 1.4 is navigation.
|
||||
|
||||
|
|
|
@ -97,7 +97,7 @@ Eviction signal value is calculated periodically and does NOT enforce the limit.
|
|||
PID limiting - per Pod and per Node sets the hard limit.
|
||||
Once the limit is hit, workload will start experiencing failures when trying to get a new PID.
|
||||
It may or may not lead to rescheduling of a Pod,
|
||||
depending on how workload reacts on these failures and how liveleness and readiness
|
||||
depending on how workload reacts on these failures and how liveness and readiness
|
||||
probes are configured for the Pod. However, if limits were set correctly,
|
||||
you can guarantee that other Pods workload and system processes will not run out of PIDs
|
||||
when one Pod is misbehaving.
|
||||
|
|
|
@ -438,7 +438,7 @@ The two options are discussed in more detail in the following sections.
|
|||
|
||||
As previously mentioned, you should consider isolating each workload in its own namespace, even if
|
||||
you are using dedicated clusters or virtualized control planes. This ensures that each workload
|
||||
only has access to its own resources, such as Config Maps and Secrets, and allows you to tailor
|
||||
only has access to its own resources, such as ConfigMaps and Secrets, and allows you to tailor
|
||||
dedicated security policies for each workload. In addition, it is a best practice to give each
|
||||
namespace names that are unique across your entire fleet (that is, even if they are in separate
|
||||
clusters), as this gives you the flexibility to switch between dedicated and shared clusters in
|
||||
|
|
Loading…
Reference in New Issue