629 lines
25 KiB
Markdown
629 lines
25 KiB
Markdown
---
|
||
title: 审计
|
||
content_type: concept
|
||
---
|
||
<!--
|
||
reviewers:
|
||
- soltysh
|
||
- sttts
|
||
- ericchiang
|
||
content_type: concept
|
||
title: Auditing
|
||
-->
|
||
<!-- overview -->
|
||
|
||
{{< feature-state state="beta" >}}
|
||
|
||
<!--
|
||
Kubernetes auditing provides a security-relevant chronological set of records documenting
|
||
the sequence of activities that have affected system by individual users, administrators
|
||
or other components of the system. It allows cluster administrator to
|
||
answer the following questions:
|
||
-->
|
||
Kubernetes 审计功能提供了与安全相关的按时间顺序排列的记录集,记录每个用户、管理员
|
||
或系统其他组件影响系统的活动顺序。
|
||
它能帮助集群管理员处理以下问题:
|
||
|
||
<!--
|
||
- what happened?
|
||
- when did it happen?
|
||
- who initiated it?
|
||
- on what did it happen?
|
||
- where was it observed?
|
||
- from where was it initiated?
|
||
- to where was it going?
|
||
-->
|
||
- 发生了什么?
|
||
- 什么时候发生的?
|
||
- 谁触发的?
|
||
- 活动发生在哪个(些)对象上?
|
||
- 在哪观察到的?
|
||
- 它从哪触发的?
|
||
- 活动的后续处理行为是什么?
|
||
|
||
<!-- body -->
|
||
|
||
<!--
|
||
Audit records begin their lifecycle inside the
|
||
[kube-apiserver](/docs/reference/command-line-tools-reference/kube-apiserver/)
|
||
component. Each request on each stage
|
||
of its execution generates an audit event, which is then pre-processed according to
|
||
a certain policy and written to a backend. The policy determines what's recorded
|
||
and the backends persist the records. The current backend implementations
|
||
include logs files and webhooks.
|
||
-->
|
||
审计记录最初产生于
|
||
[kube-apiserver](/zh/docs/reference/command-line-tools-reference/kube-apiserver/)
|
||
内部。每个请求在不同执行阶段都会生成审计事件;这些审计事件会根据特定策略
|
||
被预处理并写入后端。策略确定要记录的内容和用来存储记录的后端。
|
||
当前的后端支持日志文件和 webhook。
|
||
|
||
<!--
|
||
Each request can be recorded with an associated "stage". The known stages are:
|
||
|
||
- `RequestReceived` - The stage for events generated as soon as the audit
|
||
handler receives the request, and before it is delegated down the handler
|
||
chain.
|
||
- `ResponseStarted` - Once the response headers are sent, but before the
|
||
response body is sent. This stage is only generated for long-running requests
|
||
(e.g. watch).
|
||
- `ResponseComplete` - The response body has been completed and no more bytes
|
||
will be sent.
|
||
- `Panic` - Events generated when a panic occurred.
|
||
-->
|
||
每个请求都可以用相关的 "stage" 记录。已知的 stage 有:
|
||
|
||
- `RequestReceived` - 事件的 stage 将在审计处理器接收到请求后,并且在委托给其余处理器之前生成。
|
||
- `ResponseStarted` - 在响应消息的头部发送后,但是响应消息体发送前。
|
||
这个阶段仅为长时间运行的请求生成(例如 watch)。
|
||
- `ResponseComplete` - 当响应消息体完成并且没有更多数据需要传输的时候。
|
||
- `Panic` - 当 panic 发生时生成。
|
||
|
||
<!--
|
||
The audit logging feature increases the memory consumption of the API
|
||
server because some context required for auditing is stored for each request.
|
||
Additionally, memory consumption depends on the audit logging configuration.
|
||
-->
|
||
{{< note >}}
|
||
审计日志记录功能会增加 API server 的内存消耗,因为需要为每个请求存储审计所需的某些上下文。
|
||
此外,内存消耗取决于审计日志记录的配置。
|
||
{{< /note >}}
|
||
|
||
<!--
|
||
## Audit Policy
|
||
|
||
Audit policy defines rules about what events should be recorded and what data
|
||
they should include. The audit policy object structure is defined in the
|
||
[`audit.k8s.io` API group][auditing-api]. When an event is processed, it's
|
||
compared against the list of rules in order. The first matching rule sets the
|
||
"audit level" of the event. The known audit levels are:
|
||
-->
|
||
## 审计策略 {#audit-policy}
|
||
|
||
审计政策定义了关于应记录哪些事件以及应包含哪些数据的规则。
|
||
审计策略对象结构定义在
|
||
[`audit.k8s.io` API 组](https://github.com/kubernetes/kubernetes/blob/{{< param "githubbranch" >}}/staging/src/k8s.io/apiserver/pkg/apis/audit/v1/types.go)
|
||
处理事件时,将按顺序与规则列表进行比较。第一个匹配规则设置事件的
|
||
"审计级别"。已知的审计级别有:
|
||
|
||
<!--
|
||
- `None` - don't log events that match this rule.
|
||
- `Metadata` - log request metadata (requesting user, timestamp, resource,
|
||
verb, etc.) but not request or response body.
|
||
- `Request` - log event metadata and request body but not response body.
|
||
This does not apply for non-resource requests.
|
||
- `RequestResponse` - log event metadata, request and response bodies.
|
||
This does not apply for non-resource requests.
|
||
-->
|
||
- `None` - 符合这条规则的日志将不会记录。
|
||
- `Metadata` - 记录请求的元数据(请求的用户、时间戳、资源、动词等等),
|
||
但是不记录请求或者响应的消息体。
|
||
- `Request` - 记录事件的元数据和请求的消息体,但是不记录响应的消息体。
|
||
这不适用于非资源类型的请求。
|
||
- `RequestResponse` - 记录事件的元数据,请求和响应的消息体。这不适用于非资源类型的请求。
|
||
|
||
<!--
|
||
You can pass a file with the policy to [kube-apiserver][kube-apiserver]
|
||
using the `--audit-policy-file` flag. If the flag is omitted, no events are logged.
|
||
Note that the `rules` field __must__ be provided in the audit policy file.
|
||
A policy with no (0) rules is treated as illegal.
|
||
|
||
Below is an example audit policy file:
|
||
-->
|
||
你可以使用 `--audit-policy-file` 标志将包含策略的文件传递给 `kube-apiserver`。
|
||
如果不设置该标志,则不记录事件。
|
||
注意 `rules` 字段 __必须__ 在审计策略文件中提供。没有(0)规则的策略将被视为非法配置。
|
||
|
||
以下是一个审计策略文件的示例:
|
||
|
||
{{< codenew file="audit/audit-policy.yaml" >}}
|
||
|
||
<!--
|
||
You can use a minimal audit policy file to log all requests at the `Metadata` level:
|
||
-->
|
||
你可以使用最低限度的审计策略文件在 `Metadata` 级别记录所有请求:
|
||
|
||
```yaml
|
||
# 在 Metadata 级别为所有请求生成日志
|
||
apiVersion: audit.k8s.io/v1beta1
|
||
kind: Policy
|
||
rules:
|
||
- level: Metadata
|
||
```
|
||
|
||
<!--
|
||
The audit profile used by GCE should be used as reference by admins constructing their own audit profiles. You can check the
|
||
[configure-helper.sh](https://github.com/kubernetes/kubernetes/blob/{{< param "githubbranch" >}}/cluster/gce/gci/configure-helper.sh)
|
||
script, which generates the audit policy file. You can see most of the audit policy file by looking directly at the script.
|
||
-->
|
||
管理员构建自己的审计配置文件时,可参考
|
||
[configure-helper.sh](https://github.com/kubernetes/kubernetes/blob/{{< param "githubbranch" >}}/cluster/gce/gci/configure-helper.sh)
|
||
脚本,该脚本生成审计策略文件。你可以直接在脚本中看到审计策略的绝大部份内容。
|
||
[]。
|
||
|
||
<!--
|
||
## Audit backends
|
||
|
||
Audit backends persist audit events to an external storage.
|
||
Out of the box, the kube-apiserver provides two backends:
|
||
|
||
- Log backend, which writes events to a disk
|
||
- Webhook backend, which sends events to an external API
|
||
|
||
In both cases, audit events structure is defined by the API in the
|
||
`audit.k8s.io` API group. The current version of the API is
|
||
[`v1`](https://github.com/kubernetes/kubernetes/blob/{{< param "githubbranch" >}}/staging/src/k8s.io/apiserver/pkg/apis/audit/v1/types.go).
|
||
-->
|
||
## 审计后端 {#audit-backends}
|
||
|
||
审计后端实现将审计事件导出到外部存储。`Kube-apiserver` 默认提供两个后端:
|
||
|
||
- Log 后端,将事件写入到磁盘
|
||
- Webhook 后端,将事件发送到外部 API
|
||
|
||
在这两种情况下,审计事件结构均由 `audit.k8s.io` API 组中的 API 定义。
|
||
当前版本的 API 是
|
||
[`v1`](https://github.com/kubernetes/kubernetes/blob/{{< param "githubbranch" >}}/staging/src/k8s.io/apiserver/pkg/apis/audit/v1/types.go).
|
||
|
||
<!--
|
||
In case of patches, request body is a JSON array with patch operations, not a JSON object
|
||
with an appropriate Kubernetes API object. For example, the following request body is a valid patch
|
||
request to `/apis/batch/v1/namespaces/some-namespace/jobs/some-job-name`.
|
||
-->
|
||
{{< note >}}
|
||
在 patch 请求的情况下,请求的消息体需要是一个 JSON 串指定 patch 操作,
|
||
而不是一个完整的 Kubernetes API 对象 JSON 串。
|
||
例如,以下的示例是一个合法的 patch 请求消息体,该请求对应
|
||
`/apis/batch/v1/namespaces/some-namespace/jobs/some-job-name`。
|
||
|
||
```json
|
||
[
|
||
{
|
||
"op": "replace",
|
||
"path": "/spec/parallelism",
|
||
"value": 0
|
||
},
|
||
{
|
||
"op": "remove",
|
||
"path": "/spec/template/spec/containers/0/terminationMessagePolicy"
|
||
}
|
||
]
|
||
```
|
||
{{< /note >}}
|
||
|
||
<!--
|
||
### Log backend
|
||
|
||
Log backend writes audit events to a file in JSON format. You can configure
|
||
log audit backend using the following [kube-apiserver][kube-apiserver] flags:
|
||
-->
|
||
### Log 后端
|
||
|
||
Log 后端将审计事件写入 JSON 格式的文件。你可以使用以下 `kube-apiserver` 标志配置
|
||
Log 审计后端:
|
||
|
||
<!--
|
||
- `--audit-log-path` specifies the log file path that log backend uses to write
|
||
audit events. Not specifying this flag disables log backend. `-` means standard out
|
||
- `--audit-log-maxage` defined the maximum number of days to retain old audit log files
|
||
- `--audit-log-maxbackup` defines the maximum number of audit log files to retain
|
||
- `--audit-log-maxsize` defines the maximum size in megabytes of the audit log file before it gets rotated
|
||
-->
|
||
- `--audit-log-path` 指定用来写入审计事件的日志文件路径。不指定此标志会禁用日志后端。`-` 意味着标准化
|
||
- `--audit-log-maxage` 定义了保留旧审计日志文件的最大天数
|
||
- `--audit-log-maxbackup` 定义了要保留的审计日志文件的最大数量
|
||
- `--audit-log-maxsize` 定义审计日志文件的最大大小(兆字节)
|
||
|
||
<!--
|
||
In case kube-apiserver is configured as a Pod,remember to mount the hostPath to the location of the policy file and log file. For example,
|
||
`
|
||
--audit-policy-file=/etc/kubernetes/audit-policy.yaml
|
||
--audit-log-path=/var/log/audit.log
|
||
`
|
||
then mount the volumes:
|
||
-->
|
||
如果 `kube-apiserver` 被配置为运行在 Pod 中,请记得将包含策略文件和日志文件的
|
||
位置用 `hostPath` 挂载到 Pod 中。例如,
|
||
|
||
```
|
||
--audit-policy-file=/etc/kubernetes/audit-policy.yaml
|
||
--audit-log-path=/var/log/audit.log
|
||
```
|
||
|
||
接下来挂载数据卷:
|
||
|
||
```
|
||
volumeMounts:
|
||
- mountPath: /etc/kubernetes/audit-policy.yaml
|
||
name: audit
|
||
readOnly: true
|
||
- mountPath: /var/log/audit.log
|
||
name: audit-log
|
||
readOnly: false
|
||
```
|
||
|
||
<!-- finally the hostPath: -->
|
||
下面是 hostPath 卷本身。
|
||
|
||
```
|
||
- name: audit
|
||
hostPath:
|
||
path: /etc/kubernetes/audit-policy.yaml
|
||
type: File
|
||
|
||
- name: audit-log
|
||
hostPath:
|
||
path: /var/log/audit.log
|
||
type: FileOrCreate
|
||
|
||
```
|
||
|
||
<!--
|
||
### Webhook backend
|
||
|
||
Webhook backend sends audit events to a remote API, which is assumed to be the
|
||
same API as [kube-apiserver][kube-apiserver] exposes. You can configure webhook
|
||
audit backend using the following kube-apiserver flags:
|
||
-->
|
||
### Webhook 后端 {#webhook-backend}
|
||
|
||
Webhook 后端将审计事件发送到远程 API,该远程 API 应该暴露与 `kube-apiserver` 相同的API。
|
||
你可以使用如下 kube-apiserver 标志来配置 webhook 审计后端:
|
||
|
||
<!--
|
||
- `--audit-webhook-config-file` specifies the path to a file with a webhook
|
||
configuration. Webhook configuration is effectively a [kubeconfig][kubeconfig].
|
||
- `--audit-webhook-initial-backoff` specifies the amount of time to wait after the first failed
|
||
request before retrying. Subsequent requests are retried with exponential backoff.
|
||
|
||
The webhook config file uses the kubeconfig format to specify the remote address of
|
||
the service and credentials used to connect to it.
|
||
-->
|
||
- `--audit-webhook-config-file` webhook 配置文件的路径。Webhook 配置文件实际上是一个
|
||
[kubeconfig 文件](/zh/docs/concepts/configuration/organize-cluster-access-kubeconfig/)
|
||
- `--audit-webhook-initial-backoff` 指定在第一次失败后重发请求等待的时间。随后的请求将以指数退避重试。
|
||
|
||
webhook 配置文件使用 kubeconfig 格式指定服务的远程地址和用于连接它的凭据。
|
||
|
||
<!--
|
||
### Batching
|
||
|
||
Both log and webhook backends support batching. Using webhook as an example, here's the list of
|
||
available flags. To get the same flag for log backend, replace `webhook` with `log` in the flag
|
||
name. By default, batching is enabled in `webhook` and disabled in `log`. Similarly, by default
|
||
throttling is enabled in `webhook` and disabled in `log`.
|
||
|
||
- `--audit-webhook-mode` defines the buffering strategy. One of the following:
|
||
- `batch` - buffer events and asynchronously process them in batches. This is the default.
|
||
- `blocking` - block API server responses on processing each individual event.
|
||
- `blocking-strict` - Same as blocking, but when there is a failure during audit logging at RequestReceived stage, the whole request to apiserver will fail.
|
||
-->
|
||
### 批处理 {#batching}
|
||
|
||
log 和 webhook 后端都支持批处理。以 webhook 为例,以下是可用参数列表。要获取 log 后端的同样参数,
|
||
请在参数名称中将 `webhook` 替换为 `log`。
|
||
默认情况下,在 `webhook` 中启用批处理,在 `log` 中禁用批处理。
|
||
同样,默认情况下,在 `webhook` 中启用带宽限制,在 `log` 中禁用带宽限制。
|
||
|
||
- `--audit-webhook-mode` 定义缓存策略,可选值如下:
|
||
- `batch` - 以批处理缓存事件和异步的过程。这是默认值。
|
||
- `blocking` - 在 API 服务器处理每个单独事件时,阻塞其响应。
|
||
- `blocking-strict` - 与 `blocking` 相同,不过当审计日志在 RequestReceived 阶段
|
||
失败时,整个 API 服务请求会失效。
|
||
|
||
<!--
|
||
The following flags are used only in the `batch` mode.
|
||
|
||
- `--audit-webhook-batch-buffer-size` defines the number of events to buffer before batching.
|
||
If the rate of incoming events overflows the buffer, events are dropped.
|
||
- `--audit-webhook-batch-max-size` defines the maximum number of events in one batch.
|
||
- `--audit-webhook-batch-max-wait` defines the maximum amount of time to wait before unconditionally
|
||
batching events in the queue.
|
||
- `--audit-webhook-batch-throttle-qps` defines the maximum average number of batches generated
|
||
per second.
|
||
- `--audit-webhook-batch-throttle-burst` defines the maximum number of batches generated at the same
|
||
moment if the allowed QPS was underutilized previously.
|
||
-->
|
||
以下参数仅用于 `batch` 模式。
|
||
|
||
- `--audit-webhook-batch-buffer-size` 定义 batch 之前要缓存的事件数。
|
||
如果传入事件的速率溢出缓存区,则会丢弃事件。
|
||
- `--audit-webhook-batch-max-size` 定义一个 batch 中的最大事件数。
|
||
- `--audit-webhook-batch-max-wait` 无条件 batch 队列中的事件前等待的最大事件。
|
||
- `--audit-webhook-batch-throttle-qps` 每秒生成的最大批次数。
|
||
- `--audit-webhook-batch-throttle-burst` 在达到允许的 QPS 前,同一时刻允许存在的最大 batch 生成数。
|
||
|
||
<!--
|
||
#### Parameter tuning
|
||
|
||
Parameters should be set to accommodate the load on the apiserver.
|
||
|
||
For example, if kube-apiserver receives 100 requests each second, and each request is audited only
|
||
on `ResponseStarted` and `ResponseComplete` stages, you should account for ~200 audit
|
||
events being generated each second. Assuming that there are up to 100 events in a batch,
|
||
you should set throttling level at least 2 QPS. Assuming that the backend can take up to
|
||
5 seconds to write events, you should set the buffer size to hold up to 5 seconds of events, i.e.
|
||
10 batches, i.e. 1000 events.
|
||
-->
|
||
#### 参数调整
|
||
|
||
需要设置参数以适应 apiserver 上的负载。
|
||
|
||
|
||
例如,如果 kube-apiserver 每秒收到 100 个请求,并且每个请求仅在 `ResponseStarted` 和 `ResponseComplete` 阶段进行审计,则应该考虑每秒生成约 200 个审计事件。
|
||
假设批处理中最多有 100 个事件,则应将限制级别设置为至少 2 个 QPS。
|
||
假设后端最多需要 5 秒钟来写入事件,你应该设置缓冲区大小以容纳最多 5 秒的事件,即 10 个 batch,即 1000 个事件。
|
||
|
||
<!--
|
||
In most cases however, the default parameters should be sufficient and you don't have to worry about
|
||
setting them manually. You can look at the following Prometheus metrics exposed by kube-apiserver
|
||
and in the logs to monitor the state of the auditing subsystem.
|
||
|
||
- `apiserver_audit_event_total` metric contains the total number of audit events exported.
|
||
- `apiserver_audit_error_total` metric contains the total number of events dropped due to an error
|
||
during exporting.
|
||
-->
|
||
但是,在大多数情况下,默认参数应该足够了,你不必手动设置它们。
|
||
你可以查看 kube-apiserver 公开的以下 Prometheus 指标,并在日志中监控审计子系统的状态。
|
||
|
||
- `apiserver_audit_event_total` 包含所有暴露的审计事件数量的指标。
|
||
- `apiserver_audit_error_total` 在暴露时由于发生错误而被丢弃的事件的数量。
|
||
|
||
<!--
|
||
## Setup for multiple API servers
|
||
|
||
If you're extending the Kubernetes API with the [aggregation layer][kube-aggregator], you can also
|
||
set up audit logging for the aggregated apiserver. To do this, pass the configuration options in the
|
||
same format as described above to the aggregated apiserver and set up the log ingesting pipeline
|
||
to pick up audit logs. Different apiservers can have different audit configurations and different
|
||
audit policies.
|
||
-->
|
||
## 多 API 服务器的配置
|
||
|
||
如果你通过[聚合层](/zh/docs/concepts/extend-kubernetes/api-extension/apiserver-aggregation/)
|
||
对 Kubernetes API 进行扩展,那么你也可以为聚合的 API 服务器设置审计日志。
|
||
想要这么做,你需要以上述的格式给聚合的 API 服务器配置参数,并且配置日志管道以采用审计日志。
|
||
不同的 API 服务器可以配置不同的审计配置和策略。
|
||
|
||
<!--
|
||
## Log Collector Examples
|
||
|
||
### Use fluentd to collect and distribute audit events from log file
|
||
|
||
[Fluentd](https://www.fluentd.org/) is an open source data collector for unified logging layer.
|
||
In this example, we will use fluentd to split audit events by different namespaces.
|
||
-->
|
||
## 日志收集器示例
|
||
|
||
### 使用 fluentd 从日志文件中选择并且分发审计日志
|
||
|
||
[Fluentd](https://www.fluentd.org/) 是一个开源的数据采集器,可以从统一的日志层中采集。
|
||
在以下示例中,我们将使用 fluentd 来按照命名空间划分审计事件。
|
||
|
||
<!--
|
||
The `fluent-plugin-forest` and `fluent-plugin-rewrite-tag-filter` are plugins for fluentd.
|
||
You can get details about plugin installation from
|
||
[fluentd plugin-management](https://docs.fluentd.org/v1.0/articles/plugin-management).
|
||
-->
|
||
{{< note >}}
|
||
`fluent-plugin-forest` 和 `fluent-plugin-rewrite-tag-filter` fluentd 的插件。
|
||
你可以在 [fluentd plugin-management](https://docs.fluentd.org/v1.0/articles/plugin-management)
|
||
了解安装插件相关的细节。
|
||
{{< /note >}}
|
||
|
||
<!--
|
||
1. Install [`fluentd`](https://docs.fluentd.org/v1.0/articles/quickstart#step-1:-installing-fluentd),
|
||
`fluent-plugin-forest` and `fluent-plugin-rewrite-tag-filter` in the kube-apiserver node
|
||
|
||
1. Create a config file for fluentd
|
||
-->
|
||
1. 在 kube-apiserver 节点上安装 [`fluentd`](https://docs.fluentd.org/v1.0/articles/quickstart#step-1:-installing-fluentd)、
|
||
, `fluent-plugin-forest` 和 `fluent-plugin-rewrite-tag-filter`。
|
||
|
||
1. 为 fluentd 创建一个配置文件
|
||
|
||
```none
|
||
$ cat <<EOF > /etc/fluentd/config
|
||
# fluentd 运行在 kube-apiserver 相同的主机上
|
||
<source>
|
||
@type tail
|
||
# kube-apiserver 审计日志路径
|
||
path /var/log/audit
|
||
pos_file /var/log/audit.pos
|
||
format json
|
||
time_key time
|
||
time_format %Y-%m-%dT%H:%M:%S.%N%z
|
||
tag audit
|
||
</source>
|
||
|
||
<filter audit>
|
||
#https://github.com/fluent/fluent-plugin-rewrite-tag-filter/issues/13
|
||
type record_transformer
|
||
enable_ruby
|
||
<record>
|
||
namespace ${record["objectRef"].nil? ? "none":(record["objectRef"]["namespace"].nil? ? "none":record["objectRef"]["namespace"])}
|
||
</record>
|
||
</filter>
|
||
|
||
<match audit>
|
||
# 根据上下文中的名字空间元素对审计进行路由
|
||
@type rewrite_tag_filter
|
||
rewriterule1 namespace ^(.+) ${tag}.$1
|
||
</match>
|
||
|
||
<filter audit.**>
|
||
@type record_transformer
|
||
remove_keys namespace
|
||
</filter>
|
||
|
||
<match audit.**>
|
||
@type forest
|
||
subtype file
|
||
remove_prefix audit
|
||
<template>
|
||
time_slice_format %Y%m%d%H
|
||
compress gz
|
||
path /var/log/audit-${tag}.*.log
|
||
format json
|
||
include_time_key true
|
||
</template>
|
||
</match>
|
||
```
|
||
|
||
<!--
|
||
1. start fluentd
|
||
-->
|
||
3. 启动 fluentd
|
||
|
||
```shell
|
||
fluentd -c /etc/fluentd/config -vv
|
||
```
|
||
|
||
<!--
|
||
1. start kube-apiserver with the following options:
|
||
-->
|
||
4. 给 kube-apiserver 配置以下参数并启动:
|
||
|
||
```shell
|
||
--audit-policy-file=/etc/kubernetes/audit-policy.yaml --audit-log-path=/var/log/kube-audit --audit-log-format=json
|
||
```
|
||
|
||
<!--
|
||
1. check audits for different namespaces in `/var/log/audit-*.log`
|
||
-->
|
||
5. 在 `/var/log/audit-*.log` 文件中检查不同命名空间的审计事件
|
||
|
||
<!--
|
||
### Use logstash to collect and distribute audit events from webhook backend
|
||
|
||
|
||
[Logstash](https://www.elastic.co/products/logstash)
|
||
is an open source, server-side data processing tool. In this example,
|
||
we will use logstash to collect audit events from webhook backend, and save events of
|
||
different users into different files.
|
||
|
||
1. install [logstash](https://www.elastic.co/guide/en/logstash/current/installing-logstash.html)
|
||
1. create config file for logstash
|
||
-->
|
||
### 使用 logstash 采集并分发 webhook 后端的审计事件
|
||
|
||
[Logstash](https://www.elastic.co/products/logstash) 是一个开源的、服务器端的数据处理工具。
|
||
在下面的示例中,我们将使用 logstash 采集 webhook 后端的审计事件,并且将来自不同用户的事件存入不同的文件。
|
||
|
||
1. 安装 [logstash](https://www.elastic.co/guide/en/logstash/current/installing-logstash.html)
|
||
|
||
1. 为 logstash 创建配置文件
|
||
|
||
```none
|
||
cat <<EOF > /etc/logstash/config
|
||
input{
|
||
http{
|
||
#TODO, figure out a way to use kubeconfig file to authenticate to logstash
|
||
#https://www.elastic.co/guide/en/logstash/current/plugins-inputs-http.html#plugins-inputs-http-ssl
|
||
port=>8888
|
||
}
|
||
}
|
||
filter{
|
||
split{
|
||
# Webhook 审计后端与 EventList 一起发送若干事件
|
||
# 对事件进行分割
|
||
field=>[items]
|
||
# 我们只需要 event 子元素,去掉其他元素
|
||
remove_field=>[headers, metadata, apiVersion, "@timestamp", kind, "@version", host]
|
||
}
|
||
mutate{
|
||
rename => {items=>event}
|
||
}
|
||
}
|
||
output{
|
||
file{
|
||
# 来自不同用户的审计事件会被保存到不同文件中
|
||
path=>"/var/log/kube-audit-%{[event][user][username]}/audit"
|
||
}
|
||
}
|
||
```
|
||
|
||
<!--
|
||
1. start logstash
|
||
-->
|
||
3. 启动 logstash
|
||
|
||
```shell
|
||
bin/logstash -f /etc/logstash/config --path.settings /etc/logstash/
|
||
```
|
||
|
||
<!--
|
||
1. create a [kubeconfig file](/docs/tasks/access-application-cluster/configure-access-multiple-clusters/) for kube-apiserver webhook audit backend
|
||
-->
|
||
4. 为 kube-apiserver webhook 审计后端创建一个
|
||
[kubeconfig 文件](/zh/docs/tasks/access-application-cluster/configure-access-multiple-clusters/):
|
||
|
||
```bash
|
||
cat <<EOF > /etc/kubernetes/audit-webhook-kubeconfig
|
||
apiVersion: v1
|
||
clusters:
|
||
- cluster:
|
||
server: http://<ip_of_logstash>:8888
|
||
name: logstash
|
||
contexts:
|
||
- context:
|
||
cluster: logstash
|
||
user: ""
|
||
name: default-context
|
||
current-context: default-context
|
||
kind: Config
|
||
preferences: {}
|
||
users: []
|
||
EOF
|
||
```
|
||
|
||
<!--
|
||
1. start kube-apiserver with the following options:
|
||
-->
|
||
5. 为 kube-apiserver 配置以下参数并启动:
|
||
|
||
```shell
|
||
--audit-policy-file=/etc/kubernetes/audit-policy.yaml --audit-webhook-config-file=/etc/kubernetes/audit-webhook-kubeconfig
|
||
```
|
||
|
||
<!--
|
||
1. check audits in logstash node's directories `/var/log/kube-audit-*/audit`
|
||
|
||
Note that in addition to file output plugin, logstash has a variety of outputs that
|
||
let users route data where they want. For example, users can emit audit events to elasticsearch
|
||
plugin which supports full-text search and analytics.
|
||
-->
|
||
6. 在 logstash 节点的 `/var/log/kube-audit-*/audit` 目录中检查审计事件
|
||
|
||
请注意,除了文件输出插件外,logstash 还有其它多种输出可以让用户路由不同的数据。
|
||
例如,用户可以将审计事件发送给支持全文搜索和分析的 elasticsearch 插件。
|
||
|
||
## {{% heading "whatsnext" %}}
|
||
|
||
<!--
|
||
Learn about [Mutating webhook auditing annotations](/docs/reference/access-authn-authz/extensible-admission-controllers/#mutating-webhook-auditing-annotations).
|
||
-->
|
||
|
||
了解 [Mutating webhook 审计注解](/zh/docs/reference/access-authn-authz/extensible-admission-controllers/#mutating-webhook-auditing-annotations)。
|
||
|