From 0538dddd9d8dd19ce4442fecb5bd30c8a1fd0e85 Mon Sep 17 00:00:00 2001 From: Priya Wadhwa Date: Mon, 3 Aug 2020 13:48:09 -0400 Subject: [PATCH 1/7] Upgrade storage provisioner to v2 --- Makefile | 5 ++-- .../storage-provisioner.yaml.tmpl | 2 +- pkg/minikube/assets/addons.go | 23 +++++++++++-------- pkg/minikube/bootstrapper/images/images.go | 3 ++- .../bootstrapper/images/images_test.go | 4 ++-- .../bootstrapper/images/kubeadm_test.go | 12 +++++----- pkg/minikube/cruntime/containerd_test.go | 2 +- pkg/version/version.go | 8 +++++++ 8 files changed, 35 insertions(+), 24 deletions(-) diff --git a/Makefile b/Makefile index 05c3895fc4..b5a0a1b9e4 100644 --- a/Makefile +++ b/Makefile @@ -87,8 +87,7 @@ SHA512SUM=$(shell command -v sha512sum || echo "shasum -a 512") GVISOR_TAG ?= latest # storage provisioner tag to push changes to -# to update minikubes default, update pkg/minikube/bootstrapper/images -STORAGE_PROVISIONER_TAG ?= latest +STORAGE_PROVISIONER_TAG ?= v2 # TODO: multi-arch manifest ifeq ($(GOARCH),amd64) @@ -98,7 +97,7 @@ STORAGE_PROVISIONER_IMAGE ?= $(REGISTRY)/storage-provisioner-$(GOARCH):$(STORAGE endif # Set the version information for the Kubernetes servers -MINIKUBE_LDFLAGS := -X k8s.io/minikube/pkg/version.version=$(VERSION) -X k8s.io/minikube/pkg/version.isoVersion=$(ISO_VERSION) -X k8s.io/minikube/pkg/version.isoPath=$(ISO_BUCKET) -X k8s.io/minikube/pkg/version.gitCommitID=$(COMMIT) +MINIKUBE_LDFLAGS := -X k8s.io/minikube/pkg/version.version=$(VERSION) -X k8s.io/minikube/pkg/version.isoVersion=$(ISO_VERSION) -X k8s.io/minikube/pkg/version.isoPath=$(ISO_BUCKET) -X k8s.io/minikube/pkg/version.gitCommitID=$(COMMIT) -X k8s.io/minikube/pkg/version.storageProvisionerVersion=$(STORAGE_PROVISIONER_TAG) PROVISIONER_LDFLAGS := "-X k8s.io/minikube/pkg/storage.version=$(STORAGE_PROVISIONER_TAG) -s -w -extldflags '-static'" MINIKUBEFILES := ./cmd/minikube/ diff --git a/deploy/addons/storage-provisioner/storage-provisioner.yaml.tmpl b/deploy/addons/storage-provisioner/storage-provisioner.yaml.tmpl index 7f137aa374..1c1181f6f5 100644 --- a/deploy/addons/storage-provisioner/storage-provisioner.yaml.tmpl +++ b/deploy/addons/storage-provisioner/storage-provisioner.yaml.tmpl @@ -98,7 +98,7 @@ spec: hostNetwork: true containers: - name: storage-provisioner - image: {{default "gcr.io/k8s-minikube" .ImageRepository}}/storage-provisioner{{.ExoticArch}}:v1.8.1 + image: {{default "gcr.io/k8s-minikube" .ImageRepository}}/storage-provisioner{{.ExoticArch}}:{{.StorageProvisionerVersion}} command: ["/storage-provisioner"] imagePullPolicy: IfNotPresent volumeMounts: diff --git a/pkg/minikube/assets/addons.go b/pkg/minikube/assets/addons.go index 73b2d46854..e557a96b2c 100644 --- a/pkg/minikube/assets/addons.go +++ b/pkg/minikube/assets/addons.go @@ -22,6 +22,7 @@ import ( "k8s.io/minikube/pkg/minikube/config" "k8s.io/minikube/pkg/minikube/constants" "k8s.io/minikube/pkg/minikube/vmpath" + "k8s.io/minikube/pkg/version" ) // Addon is a named list of assets, that can be enabled @@ -452,17 +453,19 @@ func GenerateTemplateData(cfg config.KubernetesConfig) interface{} { ea = "-" + runtime.GOARCH } opts := struct { - Arch string - ExoticArch string - ImageRepository string - LoadBalancerStartIP string - LoadBalancerEndIP string + Arch string + ExoticArch string + ImageRepository string + LoadBalancerStartIP string + LoadBalancerEndIP string + StorageProvisionerVersion string }{ - Arch: a, - ExoticArch: ea, - ImageRepository: cfg.ImageRepository, - LoadBalancerStartIP: cfg.LoadBalancerStartIP, - LoadBalancerEndIP: cfg.LoadBalancerEndIP, + Arch: a, + ExoticArch: ea, + ImageRepository: cfg.ImageRepository, + LoadBalancerStartIP: cfg.LoadBalancerStartIP, + LoadBalancerEndIP: cfg.LoadBalancerEndIP, + StorageProvisionerVersion: version.GetStorageProvisionerVersion(), } return opts diff --git a/pkg/minikube/bootstrapper/images/images.go b/pkg/minikube/bootstrapper/images/images.go index 1092e2abdf..b066dcc8bb 100644 --- a/pkg/minikube/bootstrapper/images/images.go +++ b/pkg/minikube/bootstrapper/images/images.go @@ -23,6 +23,7 @@ import ( "runtime" "github.com/blang/semver" + "k8s.io/minikube/pkg/version" ) // Pause returns the image name to pull for a given Kubernetes version @@ -127,7 +128,7 @@ func auxiliary(mirror string) []string { // storageProvisioner returns the minikube storage provisioner image func storageProvisioner(mirror string) string { - return path.Join(minikubeRepo(mirror), "storage-provisioner"+archTag(false)+"v1.8.1") + return path.Join(minikubeRepo(mirror), "storage-provisioner"+archTag(false)+version.GetStorageProvisionerVersion()) } // dashboardFrontend returns the image used for the dashboard frontend diff --git a/pkg/minikube/bootstrapper/images/images_test.go b/pkg/minikube/bootstrapper/images/images_test.go index 657d43a51c..2bc369b87c 100644 --- a/pkg/minikube/bootstrapper/images/images_test.go +++ b/pkg/minikube/bootstrapper/images/images_test.go @@ -24,7 +24,7 @@ import ( func TestAuxiliary(t *testing.T) { want := []string{ - "gcr.io/k8s-minikube/storage-provisioner:v1.8.1", + "gcr.io/k8s-minikube/storage-provisioner:v2", "kubernetesui/dashboard:v2.0.1", "kubernetesui/metrics-scraper:v1.0.4", } @@ -36,7 +36,7 @@ func TestAuxiliary(t *testing.T) { func TestAuxiliaryMirror(t *testing.T) { want := []string{ - "test.mirror/storage-provisioner:v1.8.1", + "test.mirror/storage-provisioner:v2", "test.mirror/dashboard:v2.0.1", "test.mirror/metrics-scraper:v1.0.4", } diff --git a/pkg/minikube/bootstrapper/images/kubeadm_test.go b/pkg/minikube/bootstrapper/images/kubeadm_test.go index d95ca2fd45..f1348ce901 100644 --- a/pkg/minikube/bootstrapper/images/kubeadm_test.go +++ b/pkg/minikube/bootstrapper/images/kubeadm_test.go @@ -37,7 +37,7 @@ func TestKubeadmImages(t *testing.T) { "k8s.gcr.io/coredns:1.6.5", "k8s.gcr.io/etcd:3.4.3-0", "k8s.gcr.io/pause:3.1", - "gcr.io/k8s-minikube/storage-provisioner:v1.8.1", + "gcr.io/k8s-minikube/storage-provisioner:v2", "kubernetesui/dashboard:v2.0.1", "kubernetesui/metrics-scraper:v1.0.4", }}, @@ -49,7 +49,7 @@ func TestKubeadmImages(t *testing.T) { "mirror.k8s.io/coredns:1.6.2", "mirror.k8s.io/etcd:3.3.15-0", "mirror.k8s.io/pause:3.1", - "mirror.k8s.io/storage-provisioner:v1.8.1", + "mirror.k8s.io/storage-provisioner:v2", "mirror.k8s.io/dashboard:v2.0.1", "mirror.k8s.io/metrics-scraper:v1.0.4", }}, @@ -61,7 +61,7 @@ func TestKubeadmImages(t *testing.T) { "k8s.gcr.io/coredns:1.3.1", "k8s.gcr.io/etcd:3.3.10", "k8s.gcr.io/pause:3.1", - "gcr.io/k8s-minikube/storage-provisioner:v1.8.1", + "gcr.io/k8s-minikube/storage-provisioner:v2", "kubernetesui/dashboard:v2.0.1", "kubernetesui/metrics-scraper:v1.0.4", }}, @@ -73,7 +73,7 @@ func TestKubeadmImages(t *testing.T) { "k8s.gcr.io/coredns:1.3.1", "k8s.gcr.io/etcd:3.3.10", "k8s.gcr.io/pause:3.1", - "gcr.io/k8s-minikube/storage-provisioner:v1.8.1", + "gcr.io/k8s-minikube/storage-provisioner:v2", "kubernetesui/dashboard:v2.0.1", "kubernetesui/metrics-scraper:v1.0.4", }}, @@ -85,7 +85,7 @@ func TestKubeadmImages(t *testing.T) { "k8s.gcr.io/coredns:1.2.6", "k8s.gcr.io/etcd:3.2.24", "k8s.gcr.io/pause:3.1", - "gcr.io/k8s-minikube/storage-provisioner:v1.8.1", + "gcr.io/k8s-minikube/storage-provisioner:v2", "kubernetesui/dashboard:v2.0.1", "kubernetesui/metrics-scraper:v1.0.4", }}, @@ -97,7 +97,7 @@ func TestKubeadmImages(t *testing.T) { "k8s.gcr.io/coredns:1.2.2", "k8s.gcr.io/etcd:3.2.24", "k8s.gcr.io/pause:3.1", - "gcr.io/k8s-minikube/storage-provisioner:v1.8.1", + "gcr.io/k8s-minikube/storage-provisioner:v2", "kubernetesui/dashboard:v2.0.1", "kubernetesui/metrics-scraper:v1.0.4", }}, diff --git a/pkg/minikube/cruntime/containerd_test.go b/pkg/minikube/cruntime/containerd_test.go index 5f99033daa..5b1e6414f9 100644 --- a/pkg/minikube/cruntime/containerd_test.go +++ b/pkg/minikube/cruntime/containerd_test.go @@ -27,7 +27,7 @@ func TestAddRepoTagToImageName(t *testing.T) { }{ {"kubernetesui/dashboard:v2.0.1", "docker.io/kubernetesui/dashboard:v2.0.1"}, {"kubernetesui/metrics-scraper:v1.0.4", "docker.io/kubernetesui/metrics-scraper:v1.0.4"}, - {"gcr.io/k8s-minikube/storage-provisioner:v1.8.1", "gcr.io/k8s-minikube/storage-provisioner:v1.8.1"}, + {"gcr.io/k8s-minikube/storage-provisioner:v2", "gcr.io/k8s-minikube/storage-provisioner:v2"}, } for _, tc := range tests { t.Run(tc.imgName, func(t *testing.T) { diff --git a/pkg/version/version.go b/pkg/version/version.go index a3b921c5ca..0faeb4964f 100644 --- a/pkg/version/version.go +++ b/pkg/version/version.go @@ -34,6 +34,9 @@ var gitCommitID = "" // isoVersion is a private field and should be set when compiling with --ldflags="-X k8s.io/minikube/pkg/version.isoVersion=vX.Y.Z" var isoVersion = "v0.0.0-unset" +// storageProvisionerVersion is a private field and should be set when compiling with --ldflags="-X k8s.io/minikube/pkg/version.storageProvisionerVersion=" +var storageProvisionerVersion = "" + // GetVersion returns the current minikube version func GetVersion() string { return version @@ -53,3 +56,8 @@ func GetISOVersion() string { func GetSemverVersion() (semver.Version, error) { return semver.Make(strings.TrimPrefix(GetVersion(), VersionPrefix)) } + +// GetStorageProvisionerVersion returns the storage provisioner version +func GetStorageProvisionerVersion() string { + return storageProvisionerVersion +} From 52bee6dcedbecf3cb2a0794f83f545f8380986ae Mon Sep 17 00:00:00 2001 From: Priya Wadhwa Date: Mon, 3 Aug 2020 13:50:22 -0400 Subject: [PATCH 2/7] Bump up preload version for new storage provisioner image --- pkg/minikube/download/preload.go | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pkg/minikube/download/preload.go b/pkg/minikube/download/preload.go index 570a7d732b..4429dabcc1 100644 --- a/pkg/minikube/download/preload.go +++ b/pkg/minikube/download/preload.go @@ -40,7 +40,7 @@ const ( // PreloadVersion is the current version of the preloaded tarball // // NOTE: You may need to bump this version up when upgrading auxiliary docker images - PreloadVersion = "v4" + PreloadVersion = "v5" // PreloadBucket is the name of the GCS bucket where preloaded volume tarballs exist PreloadBucket = "minikube-preloaded-volume-tarballs" ) From 5e899a81cc4f0c84cfedb6bd2dd46703dad1f0dd Mon Sep 17 00:00:00 2001 From: Priya Wadhwa Date: Mon, 3 Aug 2020 15:23:45 -0400 Subject: [PATCH 3/7] Make sure cluster roles are applied to the service account the storage provisioner runs under --- .../storage-provisioner/storage-provisioner.yaml.tmpl | 8 +++++--- test/integration/testdata/pvc.yaml | 2 +- 2 files changed, 6 insertions(+), 4 deletions(-) diff --git a/deploy/addons/storage-provisioner/storage-provisioner.yaml.tmpl b/deploy/addons/storage-provisioner/storage-provisioner.yaml.tmpl index 1c1181f6f5..60be98e409 100644 --- a/deploy/addons/storage-provisioner/storage-provisioner.yaml.tmpl +++ b/deploy/addons/storage-provisioner/storage-provisioner.yaml.tmpl @@ -40,7 +40,7 @@ subjects: apiVersion: rbac.authorization.k8s.io/v1 kind: Role metadata: - name: system::leader-locking-storage-provisioner + name: system:persistent-volume-provisioner namespace: kube-system labels: addonmanager.kubernetes.io/mode: EnsureExists @@ -51,6 +51,7 @@ rules: - endpoints verbs: - watch + - create - apiGroups: - "" resourceNames: @@ -60,18 +61,19 @@ rules: verbs: - get - update + - create --- apiVersion: rbac.authorization.k8s.io/v1 kind: RoleBinding metadata: - name: leader-locking-storage-provisioner + name: system:persistent-volume-provisioner namespace: kube-system labels: addonmanager.kubernetes.io/mode: EnsureExists roleRef: apiGroup: rbac.authorization.k8s.io kind: Role - name: system::leader-locking-storage-provisioner + name: system:persistent-volume-provisioner subjects: - kind: ServiceAccount name: storage-provisioner diff --git a/test/integration/testdata/pvc.yaml b/test/integration/testdata/pvc.yaml index 93f487d03d..2c0e103fa4 100644 --- a/test/integration/testdata/pvc.yaml +++ b/test/integration/testdata/pvc.yaml @@ -7,4 +7,4 @@ spec: - ReadWriteOnce resources: requests: - storage: 2Gi \ No newline at end of file + storage: 500Mi \ No newline at end of file From ae7fbaf8f96b6101829e31bf4900c81441f2ae5d Mon Sep 17 00:00:00 2001 From: Priya Wadhwa Date: Mon, 3 Aug 2020 15:50:42 -0400 Subject: [PATCH 4/7] make sure files are actually being maintained in PVCs in our integration test --- test/integration/fn_pvc_test.go | 47 +++++++++++++++++-- .../testdata/storage-provisioner/pod.yaml | 17 +++++++ .../{ => storage-provisioner}/pvc.yaml | 7 +-- 3 files changed, 65 insertions(+), 6 deletions(-) create mode 100644 test/integration/testdata/storage-provisioner/pod.yaml rename test/integration/testdata/{ => storage-provisioner}/pvc.yaml (66%) diff --git a/test/integration/fn_pvc_test.go b/test/integration/fn_pvc_test.go index 18f573d3bd..392ace7ea6 100644 --- a/test/integration/fn_pvc_test.go +++ b/test/integration/fn_pvc_test.go @@ -24,7 +24,8 @@ import ( "encoding/json" "fmt" "os/exec" - "path/filepath" + "path" + "strings" "testing" "time" @@ -64,13 +65,14 @@ func validatePersistentVolumeClaim(ctx context.Context, t *testing.T, profile st } // Now create a testpvc - rr, err := Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "apply", "-f", filepath.Join(*testdataDir, "pvc.yaml"))) + rr, err := Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "apply", "-f", path.Join(*testdataDir, "storage-provisioner", "pvc.yaml"))) if err != nil { t.Fatalf("kubectl apply pvc.yaml failed: args %q: %v", rr.Command(), err) } + // make sure the pvc is Bound checkStoragePhase := func() error { - rr, err := Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "get", "pvc", "testpvc", "-o=json")) + rr, err := Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "get", "pvc", "myclaim", "-o=json")) if err != nil { return err } @@ -88,4 +90,43 @@ func validatePersistentVolumeClaim(ctx context.Context, t *testing.T, profile st if err := retry.Expo(checkStoragePhase, 2*time.Second, Minutes(4)); err != nil { t.Fatalf("failed to check storage phase: %v", err) } + + // create a test pod that will mount the persistent volume + createPVTestPod(t, ctx, profile) + + // write to the persistent volume + podName := "sp-pod" + rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "exec", podName, "--", "touch", "/tmp/mount/foo")) + if err != nil { + t.Fatalf("creating file in pv: args %q: %v", rr.Command(), err) + } + + // kill the pod + rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "delete", "-f", path.Join(*testdataDir, "storage-provisioner", "pod.yaml"))) + if err != nil { + t.Fatalf("kubectl delete pod.yaml failed: args %q: %v", rr.Command(), err) + } + // recreate the pod + createPVTestPod(t, ctx, profile) + + // make sure the file we previously wrote to the persistent volume is still there + rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "exec", podName, "--", "ls", "/tmp/mount")) + if err != nil { + t.Fatalf("creating file in pv: args %q: %v", rr.Command(), err) + } + if !strings.Contains(rr.Output(), "foo") { + t.Fatalf("expected file foo to persist in pvc, instead got [%v] as files in pv", rr.Output()) + } +} + +func createPVTestPod(t *testing.T, ctx context.Context, profile string) { + // Deploy a pod that will mount the PV + rr, err := Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "apply", "-f", path.Join(*testdataDir, "storage-provisioner", "pod.yaml"))) + if err != nil { + t.Fatalf("kubectl apply pvc.yaml failed: args %q: %v", rr.Command(), err) + } + // wait for pod to be running + if _, err := PodWait(ctx, t, profile, "default", "test=storage-provisioner", Minutes(1)); err != nil { + t.Fatalf("failed waiting for pod: %v", err) + } } diff --git a/test/integration/testdata/storage-provisioner/pod.yaml b/test/integration/testdata/storage-provisioner/pod.yaml new file mode 100644 index 0000000000..f2376b5f49 --- /dev/null +++ b/test/integration/testdata/storage-provisioner/pod.yaml @@ -0,0 +1,17 @@ +apiVersion: v1 +kind: Pod +metadata: + name: sp-pod + labels: + test: storage-provisioner +spec: + containers: + - name: myfrontend + image: nginx + volumeMounts: + - mountPath: "/tmp/mount" + name: mypd + volumes: + - name: mypd + persistentVolumeClaim: + claimName: myclaim diff --git a/test/integration/testdata/pvc.yaml b/test/integration/testdata/storage-provisioner/pvc.yaml similarity index 66% rename from test/integration/testdata/pvc.yaml rename to test/integration/testdata/storage-provisioner/pvc.yaml index 2c0e103fa4..907dfdfd6b 100644 --- a/test/integration/testdata/pvc.yaml +++ b/test/integration/testdata/storage-provisioner/pvc.yaml @@ -1,10 +1,11 @@ -kind: PersistentVolumeClaim apiVersion: v1 +kind: PersistentVolumeClaim metadata: - name: testpvc + name: myclaim spec: accessModes: - ReadWriteOnce + volumeMode: Filesystem resources: requests: - storage: 500Mi \ No newline at end of file + storage: 500Mi From 951be82fa966b947cb3223f728a5e25dffb9b8a0 Mon Sep 17 00:00:00 2001 From: Priya Wadhwa Date: Mon, 3 Aug 2020 15:57:52 -0400 Subject: [PATCH 5/7] fix lint --- test/integration/fn_pvc_test.go | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/test/integration/fn_pvc_test.go b/test/integration/fn_pvc_test.go index 392ace7ea6..32bb9a4cc1 100644 --- a/test/integration/fn_pvc_test.go +++ b/test/integration/fn_pvc_test.go @@ -92,7 +92,7 @@ func validatePersistentVolumeClaim(ctx context.Context, t *testing.T, profile st } // create a test pod that will mount the persistent volume - createPVTestPod(t, ctx, profile) + createPVTestPod(ctx, t, profile) // write to the persistent volume podName := "sp-pod" @@ -107,7 +107,7 @@ func validatePersistentVolumeClaim(ctx context.Context, t *testing.T, profile st t.Fatalf("kubectl delete pod.yaml failed: args %q: %v", rr.Command(), err) } // recreate the pod - createPVTestPod(t, ctx, profile) + createPVTestPod(ctx, t, profile) // make sure the file we previously wrote to the persistent volume is still there rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "exec", podName, "--", "ls", "/tmp/mount")) @@ -119,7 +119,7 @@ func validatePersistentVolumeClaim(ctx context.Context, t *testing.T, profile st } } -func createPVTestPod(t *testing.T, ctx context.Context, profile string) { +func createPVTestPod(ctx context.Context, t *testing.T, profile string) { // Deploy a pod that will mount the PV rr, err := Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "apply", "-f", path.Join(*testdataDir, "storage-provisioner", "pod.yaml"))) if err != nil { From 43de628f811521d98f09e472240d6a61e2672262 Mon Sep 17 00:00:00 2001 From: Priya Wadhwa Date: Mon, 3 Aug 2020 16:18:14 -0400 Subject: [PATCH 6/7] update comment --- test/integration/fn_pvc_test.go | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/test/integration/fn_pvc_test.go b/test/integration/fn_pvc_test.go index 32bb9a4cc1..fce583d6b6 100644 --- a/test/integration/fn_pvc_test.go +++ b/test/integration/fn_pvc_test.go @@ -109,7 +109,7 @@ func validatePersistentVolumeClaim(ctx context.Context, t *testing.T, profile st // recreate the pod createPVTestPod(ctx, t, profile) - // make sure the file we previously wrote to the persistent volume is still there + // make sure the file we previously wrote to the persistent volume still exists rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "exec", podName, "--", "ls", "/tmp/mount")) if err != nil { t.Fatalf("creating file in pv: args %q: %v", rr.Command(), err) From 592363409551496aaae17341849b99a47ea0ae10 Mon Sep 17 00:00:00 2001 From: Sharif Elgamal Date: Mon, 3 Aug 2020 14:22:13 -0700 Subject: [PATCH 7/7] don't set event log on deletion --- cmd/minikube/cmd/delete.go | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/cmd/minikube/cmd/delete.go b/cmd/minikube/cmd/delete.go index c42d38cd0d..98ed5df2e3 100644 --- a/cmd/minikube/cmd/delete.go +++ b/cmd/minikube/cmd/delete.go @@ -126,7 +126,7 @@ func runDelete(cmd *cobra.Command, args []string) { if len(args) > 0 { exit.UsageT("Usage: minikube delete") } - register.SetEventLogPath(localpath.EventLog(ClusterFlagValue())) + //register.SetEventLogPath(localpath.EventLog(ClusterFlagValue())) register.Reg.SetStep(register.Deleting) validProfiles, invalidProfiles, err := config.ListProfiles()