"Kubectl get events"의 두 판 사이의 차이

잔글 (봇: 자동으로 텍스트 교체 (-</source> +</syntaxhighlight>, -<source +<syntaxhighlight ))
 
(다른 사용자 한 명의 중간 판 14개는 보이지 않습니다)
3번째 줄: 3번째 줄:
;kubectl get events
;kubectl get events
;kubectl get ev
;kubectl get ev
<source lang='console'>
 
==예시 v1.25==
<syntaxhighlight lang='console'>
$ kubectl get ev
LAST SEEN  TYPE      REASON                      OBJECT                MESSAGE
5m5s        Normal    NodeHasSufficientMemory    node/cluster1-node1  Node cluster1-node1 status is now: NodeHasSufficientMemory
57m        Warning  FreeDiskSpaceFailed        node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1268858880 bytes, but freed 0 bytes
57m        Warning  ImageGCFailed              node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1268858880 bytes, but freed 0 bytes
52m        Warning  FreeDiskSpaceFailed        node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1270890496 bytes, but freed 0 bytes
52m        Warning  ImageGCFailed              node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1270890496 bytes, but freed 0 bytes
47m        Warning  FreeDiskSpaceFailed        node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1272733696 bytes, but freed 0 bytes
47m        Warning  ImageGCFailed              node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1272733696 bytes, but freed 0 bytes
32m        Warning  FreeDiskSpaceFailed        node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 930242560 bytes, but freed 0 bytes
27m        Warning  FreeDiskSpaceFailed        node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1151131648 bytes, but freed 0 bytes
27m        Warning  ImageGCFailed              node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1151131648 bytes, but freed 0 bytes
22m        Warning  FreeDiskSpaceFailed        node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1152888832 bytes, but freed 0 bytes
22m        Warning  ImageGCFailed              node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1152888832 bytes, but freed 0 bytes
17m        Warning  FreeDiskSpaceFailed        node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1154555904 bytes, but freed 0 bytes
17m        Warning  ImageGCFailed              node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1154555904 bytes, but freed 0 bytes
12m        Warning  FreeDiskSpaceFailed        node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1146241024 bytes, but freed 0 bytes
12m        Warning  ImageGCFailed              node/cluster1-node1  failed to garbage collect required amount of images. Wanted to free 1146241024 bytes, but freed 0 bytes
10m        Warning  EvictionThresholdMet        node/cluster1-node1  Attempting to reclaim memory
10m        Normal    NodeHasInsufficientMemory  node/cluster1-node1  Node cluster1-node1 status is now: NodeHasInsufficientMemory
</syntaxhighlight>
 
==기타==
<syntaxhighlight lang='console'>
$ kubectl get events
$ kubectl get events
LASTSEEN  FIRSTSEEN  COUNT    NAME            KIND      SUBOBJECT                        TYPE      REASON
LASTSEEN  FIRSTSEEN  COUNT    NAME            KIND      SUBOBJECT                        TYPE      REASON
0s        0s          1        dapi-test-pod  Pod                                        Warning  InvalidEnvironmentVariableNames  kubelet, 127.0.0.1      Keys [1badkey, 2alsobad] from the EnvFrom secret default/mysecret were skipped since they are considered invalid environment variable names.
0s        0s          1        dapi-test-pod  Pod                                        Warning  InvalidEnvironmentVariableNames  kubelet, 127.0.0.1      Keys [1badkey, 2alsobad] from the EnvFrom secret default/mysecret were skipped since they are considered invalid environment variable names.
</source>
</syntaxhighlight>
 
<syntaxhighlight lang='console'>
$ kubectl get events
LASTSEEN  FIRSTSEEN  COUNT    NAME                        KIND        SUBOBJECT                    TYPE      REASON                  SOURCE                  MESSAGE
3m        3m          1        guestbook-75786d799f-r6mxl  Pod                                      Normal    Scheduled              default-scheduler      Successfully assigned guestbook-75786d799f-r6mxl to 10.77.155.84
3m        3m          1        guestbook-75786d799f-r6mxl  Pod                                      Normal    SuccessfulMountVolume  kubelet, 10.77.155.84  MountVolume.SetUp succeeded for volume "default-token-5rlxc"
3m        3m          1        guestbook-75786d799f-r6mxl  Pod          spec.containers{guestbook}  Normal    Pulled                  kubelet, 10.77.155.84  Container image "ibmcom/guestbook:v1" already present on machine
3m        3m          1        guestbook-75786d799f-r6mxl  Pod          spec.containers{guestbook}  Normal    Created                kubelet, 10.77.155.84  Created container
3m        3m          1        guestbook-75786d799f-r6mxl  Pod          spec.containers{guestbook}  Normal    Started                kubelet, 10.77.155.84  Started container
3m        3m          1        guestbook-75786d799f-xvpvv  Pod          spec.containers{guestbook}  Normal    Killing                kubelet, 10.77.155.84  Killing container with id docker://guestbook:Need to kill Pod
3m        3m          1        guestbook-75786d799f        ReplicaSet                                Normal    SuccessfulDelete        replicaset-controller  Deleted pod: guestbook-75786d799f-xvpvv
3m        3m          1        guestbook-75786d799f        ReplicaSet                                Normal    SuccessfulCreate        replicaset-controller  Created pod: guestbook-75786d799f-r6mxl
3m        3m          1        guestbook                    Deployment                                Normal    ScalingReplicaSet      deployment-controller  Scaled down replica set guestbook-75786d799f to 0
3m        3m          1        guestbook                    Deployment                                Normal    ScalingReplicaSet      deployment-controller  Scaled up replica set guestbook-75786d799f to 1
</syntaxhighlight>


<source lang='console'>
<syntaxhighlight lang='console'>
$ kubectl get events
$ kubectl get events
    LASTSEEN  FIRSTSEEN  COUNT    NAME                        KIND        SUBOBJECT                    TYPE      REASON                  SOURCE                  MESSAGE
LASTSEEN  FIRSTSEEN  COUNT    NAME                        KIND        SUBOBJECT                    TYPE      REASON                  SOURCE                  MESSAGE
    3m        3m          1        guestbook-75786d799f-r6mxl  Pod                                      Normal    Scheduled              default-scheduler      Successfully assigned guestbook-75786d799f-r6mxl to 10.77.155.84
3m        3m          1        guestbook-75786d799f-r6mxl  Pod                                      Normal    Scheduled              default-scheduler      Successfully assigned guestbook-75786d799f-r6mxl to 10.77.155.84
    3m        3m          1        guestbook-75786d799f-r6mxl  Pod                                      Normal    SuccessfulMountVolume  kubelet, 10.77.155.84  MountVolume.SetUp succeeded for volume "default-token-5rlxc"
3m        3m          1        guestbook-75786d799f-r6mxl  Pod                                      Normal    SuccessfulMountVolume  kubelet, 10.77.155.84  MountVolume.SetUp succeeded for volume "default-token-5rlxc"
    3m        3m          1        guestbook-75786d799f-r6mxl  Pod          spec.containers{guestbook}  Normal    Pulled                  kubelet, 10.77.155.84  Container image "ibmcom/guestbook:v1" already present on machine
3m        3m          1        guestbook-75786d799f-r6mxl  Pod          spec.containers{guestbook}  Normal    Pulled                  kubelet, 10.77.155.84  Container image "ibmcom/guestbook:v1" already present on machine
    3m        3m          1        guestbook-75786d799f-r6mxl  Pod          spec.containers{guestbook}  Normal    Created                kubelet, 10.77.155.84  Created container
3m        3m          1        guestbook-75786d799f-r6mxl  Pod          spec.containers{guestbook}  Normal    Created                kubelet, 10.77.155.84  Created container
    3m        3m          1        guestbook-75786d799f-r6mxl  Pod          spec.containers{guestbook}  Normal    Started                kubelet, 10.77.155.84  Started container
3m        3m          1        guestbook-75786d799f-r6mxl  Pod          spec.containers{guestbook}  Normal    Started                kubelet, 10.77.155.84  Started container
    3m        3m          1        guestbook-75786d799f-xvpvv  Pod          spec.containers{guestbook}  Normal    Killing                kubelet, 10.77.155.84  Killing container with id docker://guestbook:Need to kill Pod
3m        3m          1        guestbook-75786d799f-xvpvv  Pod          spec.containers{guestbook}  Normal    Killing                kubelet, 10.77.155.84  Killing container with id docker://guestbook:Need to kill Pod
    3m        3m          1        guestbook-75786d799f        ReplicaSet                                Normal    SuccessfulDelete        replicaset-controller  Deleted pod: guestbook-75786d799f-xvpvv
3m        3m          1        guestbook-75786d799f        ReplicaSet                                Normal    SuccessfulDelete        replicaset-controller  Deleted pod: guestbook-75786d799f-xvpvv
    3m        3m          1        guestbook-75786d799f        ReplicaSet                                Normal    SuccessfulCreate        replicaset-controller  Created pod: guestbook-75786d799f-r6mxl
3m        3m          1        guestbook-75786d799f        ReplicaSet                                Normal    SuccessfulCreate        replicaset-controller  Created pod: guestbook-75786d799f-r6mxl
    3m        3m          1        guestbook                    Deployment                                Normal    ScalingReplicaSet      deployment-controller  Scaled down replica set guestbook-75786d799f to 0
3m        3m          1        guestbook                    Deployment                                Normal    ScalingReplicaSet      deployment-controller  Scaled down replica set guestbook-75786d799f to 0
    3m        3m          1        guestbook                    Deployment                                Normal    ScalingReplicaSet      deployment-controller  Scaled up replica set guestbook-75786d799f to 1
3m        3m          1        guestbook                    Deployment                                Normal    ScalingReplicaSet      deployment-controller  Scaled up replica set guestbook-75786d799f to 1
</source>
</syntaxhighlight>
<syntaxhighlight lang='console'>
$ kubectl get events --all-namespaces
NAMESPACE    LAST SEEN  TYPE      REASON                    OBJECT                                          MESSAGE
default      36m        Normal    NodeHasSufficientMemory  node/master                                      Node master status is now: NodeHasSufficientMemory
default      36m        Normal    NodeHasNoDiskPressure    node/master                                      Node master status is now: NodeHasNoDiskPressure
default      36m        Normal    NodeHasSufficientPID      node/master                                      Node master status is now: NodeHasSufficientPID
default      35m        Normal    RegisteredNode            node/master                                      Node master event: Registered Node master in Controller
default      35m        Normal    Starting                  node/master                                      Starting kube-proxy.
default      35m        Normal    RegisteredNode            node/node01                                      Node node01 event: Registered Node node01 in Controller
default      35m        Normal    Starting                  node/node01                                      Starting kubelet.
default      35m        Normal    NodeHasSufficientMemory  node/node01                                      Node node01 status is now: NodeHasSufficientMemory
default      35m        Normal    NodeHasNoDiskPressure    node/node01                                      Node node01 status is now: NodeHasNoDiskPressure
default      35m        Normal    NodeHasSufficientPID      node/node01                                      Node node01 status is now: NodeHasSufficientPID
default      35m        Normal    NodeAllocatableEnforced  node/node01                                      Updated Node Allocatable limit across pods
default      35m        Normal    Starting                  node/node01                                      Starting kube-proxy.
default      35m        Normal    NodeReady                node/node01                                      Node node01 status is now: NodeReady
kube-system  35m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  33m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  32m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  30m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  29m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  27m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  26m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  21m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  15m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  13m        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  7m8s        Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  5m40s      Normal    LeaderElection            endpoints/cloud-controller-manager              katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system  35m        Warning  FailedScheduling          pod/coredns-fb8b8dccf-wcmr6                      0/1 nodes are available: 1 node(s) had taints that the pod didn't tolerate.
kube-system  35m        Warning  FailedScheduling          pod/coredns-fb8b8dccf-wcmr6                      0/2 nodes are available: 2 node(s) had taints that the pod didn't tolerate.
kube-system  35m        Normal    Scheduled                pod/coredns-fb8b8dccf-wcmr6                      Successfully assigned kube-system/coredns-fb8b8dccf-wcmr6 to node01
kube-system  35m        Normal    Pulled                    pod/coredns-fb8b8dccf-wcmr6                      Container image "k8s.gcr.io/coredns:1.3.1" already present on machine
kube-system  35m        Normal    Created                  pod/coredns-fb8b8dccf-wcmr6                      Created container coredns
kube-system  35m        Normal    Started                  pod/coredns-fb8b8dccf-wcmr6                      Started container coredns
kube-system  35m        Warning  FailedScheduling          pod/coredns-fb8b8dccf-xs6lx                      0/1 nodes are available: 1 node(s) had taints that the pod didn't tolerate.
kube-system  35m        Warning  FailedScheduling          pod/coredns-fb8b8dccf-xs6lx                      0/2 nodes are available: 2 node(s) had taints that the pod didn't tolerate.
kube-system  35m        Normal    Scheduled                pod/coredns-fb8b8dccf-xs6lx                      Successfully assigned kube-system/coredns-fb8b8dccf-xs6lx to node01
kube-system  35m        Normal    Pulled                    pod/coredns-fb8b8dccf-xs6lx                      Container image "k8s.gcr.io/coredns:1.3.1" already present on machine
kube-system  35m        Normal    Created                  pod/coredns-fb8b8dccf-xs6lx                      Created container coredns
kube-system  35m        Normal    Started                  pod/coredns-fb8b8dccf-xs6lx                      Started container coredns
kube-system  35m        Normal    SuccessfulCreate          replicaset/coredns-fb8b8dccf                    Created pod: coredns-fb8b8dccf-xs6lx
kube-system  35m        Normal    SuccessfulCreate          replicaset/coredns-fb8b8dccf                    Created pod: coredns-fb8b8dccf-wcmr6
kube-system  35m        Normal    ScalingReplicaSet        deployment/coredns                              Scaled up replica set coredns-fb8b8dccf to 2
kube-system  21m        Normal    Scheduled                pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Successfully assigned kube-system/dash-kubernetes-dashboard-dff4ccb96-h5pbw to node01
kube-system  21m        Normal    Pulling                  pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Pulling image "k8s.gcr.io/kubernetes-dashboard-amd64:v1.10.1"
kube-system  21m        Normal    Pulled                    pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Successfully pulled image "k8s.gcr.io/kubernetes-dashboard-amd64:v1.10.1"
kube-system  21m        Normal    Created                  pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Created container kubernetes-dashboard
kube-system  21m        Normal    Started                  pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Started container kubernetes-dashboard
kube-system  21m        Normal    SuccessfulCreate          replicaset/dash-kubernetes-dashboard-dff4ccb96  Created pod: dash-kubernetes-dashboard-dff4ccb96-h5pbw
kube-system  21m        Normal    ScalingReplicaSet        deployment/dash-kubernetes-dashboard            Scaled up replica set dash-kubernetes-dashboard-dff4ccb96 to 1
kube-system  36m        Normal    Pulled                    pod/etcd-master                                  Container image "k8s.gcr.io/etcd:3.3.10" already present on machine
kube-system  36m        Normal    Created                  pod/etcd-master                                  Created container etcd
kube-system  36m        Normal    Started                  pod/etcd-master                                  Started container etcd
kube-system  18m        Warning  Unhealthy                pod/etcd-master                                  Liveness probe failed: Error: context deadline exceeded
kube-system  35m        Warning  FailedScheduling          pod/katacoda-cloud-provider-758cf7cf75-ktdmf    0/1 nodes are available: 1 node(s) had taints that the pod didn't tolerate.
kube-system  35m        Warning  FailedScheduling          pod/katacoda-cloud-provider-758cf7cf75-ktdmf    0/2 nodes are available: 2 node(s) had taints that the pod didn't tolerate.
kube-system  35m        Normal    Scheduled                pod/katacoda-cloud-provider-758cf7cf75-ktdmf    Successfully assigned kube-system/katacoda-cloud-provider-758cf7cf75-ktdmf to node01
kube-system  35m        Normal    Pulling                  pod/katacoda-cloud-provider-758cf7cf75-ktdmf    Pulling image "katacoda/katacoda-cloud-provider:0.0.1"
kube-system  35m        Normal    Pulled                    pod/katacoda-cloud-provider-758cf7cf75-ktdmf    Successfully pulled image "katacoda/katacoda-cloud-provider:0.0.1"
kube-system  33m        Normal    Created                  pod/katacoda-cloud-provider-758cf7cf75-ktdmf    Created container katacoda-cloud-provider
kube-system  33m        Normal    Started                  pod/katacoda-cloud-provider-758cf7cf75-ktdmf    Started container katacoda-cloud-provider
kube-system  5m20s      Warning  Unhealthy                pod/katacoda-cloud-provider-758cf7cf75-ktdmf    Liveness probe failed: Get http://127.0.0.1:10252/healthz: dial tcp 127.0.0.1:10252: connect: connection refused
kube-system  32m        Normal    Killing                  pod/katacoda-cloud-provider-758cf7cf75-ktdmf    Container katacoda-cloud-provider failed liveness probe, will be restarted
kube-system  33m        Normal    Pulled                    pod/katacoda-cloud-provider-758cf7cf75-ktdmf    Container image "katacoda/katacoda-cloud-provider:0.0.1" already present on machine
kube-system  23s        Warning  BackOff                  pod/katacoda-cloud-provider-758cf7cf75-ktdmf    Back-off restarting failed container
kube-system  35m        Normal    SuccessfulCreate          replicaset/katacoda-cloud-provider-758cf7cf75    Created pod: katacoda-cloud-provider-758cf7cf75-ktdmf
kube-system  35m        Normal    ScalingReplicaSet        deployment/katacoda-cloud-provider              Scaled up replica set katacoda-cloud-provider-758cf7cf75 to 1
kube-system  36m        Normal    Pulled                    pod/kube-apiserver-master                        Container image "k8s.gcr.io/kube-apiserver:v1.14.0" already present on machine
kube-system  36m        Normal    Created                  pod/kube-apiserver-master                        Created container kube-apiserver
kube-system  36m        Normal    Started                  pod/kube-apiserver-master                        Started container kube-apiserver
kube-system  35m        Warning  Unhealthy                pod/kube-apiserver-master                        Liveness probe failed: HTTP probe failed with statuscode: 403
kube-system  18m        Warning  Unhealthy                pod/kube-apiserver-master                        Liveness probe failed: HTTP probe failed with statuscode: 500
kube-system  36m        Normal    Pulled                    pod/kube-controller-manager-master              Container image "k8s.gcr.io/kube-controller-manager:v1.14.0" already present on machine
kube-system  36m        Normal    Created                  pod/kube-controller-manager-master              Created container kube-controller-manager
kube-system  36m        Normal    Started                  pod/kube-controller-manager-master              Started container kube-controller-manager
kube-system  35m        Normal    LeaderElection            endpoints/kube-controller-manager                master_fba4bedb-5207-11ea-bbaa-0242ac110043 became leader
kube-system  35m        Normal    Scheduled                pod/kube-keepalived-vip-xs784                    Successfully assigned kube-system/kube-keepalived-vip-xs784 tonode01
kube-system  35m        Normal    Pulling                  pod/kube-keepalived-vip-xs784                    Pulling image "gcr.io/google_containers/kube-keepalived-vip:0.9"
kube-system  35m        Normal    Pulled                    pod/kube-keepalived-vip-xs784                    Successfully pulled image "gcr.io/google_containers/kube-keepalived-vip:0.9"
kube-system  35m        Normal    Created                  pod/kube-keepalived-vip-xs784                    Created container kube-keepalived-vip
kube-system  35m        Normal    Started                  pod/kube-keepalived-vip-xs784                    Started container kube-keepalived-vip
kube-system  35m        Normal    SuccessfulCreate          daemonset/kube-keepalived-vip                    Created pod: kube-keepalived-vip-xs784
kube-system  35m        Normal    Scheduled                pod/kube-proxy-xwmrv                            Successfully assigned kube-system/kube-proxy-xwmrv to master
kube-system  35m        Normal    Pulled                    pod/kube-proxy-xwmrv                            Container image "k8s.gcr.io/kube-proxy:v1.14.0" already present on machine
kube-system  35m        Normal    Created                  pod/kube-proxy-xwmrv                            Created container kube-proxy
kube-system  35m        Normal    Started                  pod/kube-proxy-xwmrv                            Started container kube-proxy
kube-system  35m        Normal    Scheduled                pod/kube-proxy-zm2lb                            Successfully assigned kube-system/kube-proxy-zm2lb to node01
kube-system  35m        Normal    Pulling                  pod/kube-proxy-zm2lb                            Pulling image "k8s.gcr.io/kube-proxy:v1.14.0"
kube-system  35m        Normal    Pulled                    pod/kube-proxy-zm2lb                            Successfully pulled image "k8s.gcr.io/kube-proxy:v1.14.0"
kube-system  35m        Normal    Created                  pod/kube-proxy-zm2lb                            Created container kube-proxy
kube-system  35m        Normal    Started                  pod/kube-proxy-zm2lb                            Started container kube-proxy
kube-system  35m        Normal    SuccessfulCreate          daemonset/kube-proxy                            Created pod: kube-proxy-xwmrv
kube-system  35m        Normal    SuccessfulCreate          daemonset/kube-proxy                            Created pod: kube-proxy-zm2lb
kube-system  36m        Normal    Pulled                    pod/kube-scheduler-master                        Container image "k8s.gcr.io/kube-scheduler:v1.14.0" already present on machine
kube-system  36m        Normal    Created                  pod/kube-scheduler-master                        Created container kube-scheduler
kube-system  36m        Normal    Started                  pod/kube-scheduler-master                        Started container kube-scheduler
kube-system  35m        Normal    LeaderElection            endpoints/kube-scheduler                        master_f3ff19a8-5207-11ea-88c0-0242ac110043 became leader
kube-system  35m        Normal    Scheduled                pod/weave-net-2l2g8                              Successfully assigned kube-system/weave-net-2l2g8 to master
kube-system  35m        Normal    Pulled                    pod/weave-net-2l2g8                              Container image "weaveworks/weave-kube:2.5.1" already present on machine
kube-system  35m        Normal    Created                  pod/weave-net-2l2g8                              Created container weave
kube-system  35m        Normal    Started                  pod/weave-net-2l2g8                              Started container weave
kube-system  35m        Normal    Pulled                    pod/weave-net-2l2g8                              Container image "weaveworks/weave-npc:2.5.1" already present on machine
kube-system  35m        Normal    Created                  pod/weave-net-2l2g8                              Created container weave-npc
kube-system  35m        Normal    Started                  pod/weave-net-2l2g8                              Started container weave-npc
kube-system  35m        Warning  Unhealthy                pod/weave-net-2l2g8                              Readiness probe failed: Get http://127.0.0.1:6784/status: dialtcp 127.0.0.1:6784: connect: connection refused
kube-system  35m        Normal    Scheduled                pod/weave-net-xc54f                              Successfully assigned kube-system/weave-net-xc54f to node01
kube-system  35m        Normal    Pulled                    pod/weave-net-xc54f                              Container image "weaveworks/weave-kube:2.5.1" already present on machine
kube-system  35m        Normal    Created                  pod/weave-net-xc54f                              Created container weave
kube-system  35m        Normal    Started                  pod/weave-net-xc54f                              Started container weave
kube-system  35m        Normal    Pulled                    pod/weave-net-xc54f                              Container image "weaveworks/weave-npc:2.5.1" already present on machine
kube-system  35m        Normal    Created                  pod/weave-net-xc54f                              Created container weave-npc
kube-system  35m        Normal    Started                  pod/weave-net-xc54f                              Started container weave-npc
kube-system  35m        Warning  Unhealthy                pod/weave-net-xc54f                              Readiness probe failed: Get http://127.0.0.1:6784/status: dialtcp 127.0.0.1:6784: connect: connection refused
kube-system  35m        Normal    SuccessfulCreate          daemonset/weave-net                              Created pod: weave-net-2l2g8
kube-system  35m        Normal    SuccessfulCreate          daemonset/weave-net                              Created pod: weave-net-xc54f
</syntaxhighlight>
<syntaxhighlight lang='console'>
$ kubectl get event
LAST SEEN  TYPE      REASON            OBJECT                          MESSAGE
13s        Warning  Unhealthy          pod/app-prod-6bc4658c66-8hmlp  Liveness probe failed: Get http://10.20.1.8:80/: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
13s        Warning  Unhealthy          pod/app-prod-6bc4658c66-8npgc  Liveness probe failed: Get http://10.20.3.5:80/: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
6m5s        Warning  NodeSysctlChange  node/node01                    {"unmanaged": {"fs.aio-nr": "64"}}
90s        Warning  NodeSysctlChange  node/node02                    {"unmanaged": {"fs.aio-nr": "5450"}}
</syntaxhighlight>
<syntaxhighlight lang='console'>
$ kubectl describe pod/app-prod-6bc4658c66-8hmlp | grep Events: -A99
Events:
  Type    Reason    Age                    From              Message
  ----    ------    ----                  ----              -------
  Warning  Unhealthy  117s (x3706 over 16d)  kubelet, node01  Liveness probe failed: Get http://10.20.1.8:80/: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
</syntaxhighlight>
<syntaxhighlight lang='console'>
$ kubectl get event --field-selector=involvedObject.name=app-prod-6bc4658c66-8hmlp -owide
LAST SEEN  TYPE      REASON      OBJECT                          SUBOBJECT              SOURCE            MESSAGE                                                                                                                        FIRST SEEN  COUNT  NAME
22s        Warning  Unhealthy  pod/app-prod-6bc4658c66-8hmlp  spec.containers{app}  kubelet, node01  Liveness probe failed: Get http://10.20.1.8:80/: net/http: request canceled (Client.Timeout exceeded while awaiting headers)  16d          3712    app-prod-6bc4658c66-8hmlp.16125180cbb5a1e9
</syntaxhighlight>
<syntaxhighlight lang='console'>
$ kubectl describe node/node01 | grep Events: -A99
Events:
  Type    Reason            Age                From                    Message
  ----    ------            ----                ----                    -------
  Warning  NodeSysctlChange  19m (x66 over 16d)  sysctl-monitor, node01  {"unmanaged": {"fs.aio-nr": "64"}}
</syntaxhighlight>


==같이 보기==
==같이 보기==
* [[kubectl]]
* [[kubectl]]
* [[kubectl describe]]
* [[kubectl describe pod]]
* [[kubectl describe node]]
* [[kubectl get events -ojsonpath]]
* [[k8s 이벤트]]


==참고==
==참고==

2023년 10월 23일 (월) 17:27 기준 최신판

1 개요[ | ]

kubectl get events
kubectl get ev

2 예시 v1.25[ | ]

$ kubectl get ev
LAST SEEN   TYPE      REASON                      OBJECT                MESSAGE
5m5s        Normal    NodeHasSufficientMemory     node/cluster1-node1   Node cluster1-node1 status is now: NodeHasSufficientMemory
57m         Warning   FreeDiskSpaceFailed         node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1268858880 bytes, but freed 0 bytes
57m         Warning   ImageGCFailed               node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1268858880 bytes, but freed 0 bytes
52m         Warning   FreeDiskSpaceFailed         node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1270890496 bytes, but freed 0 bytes
52m         Warning   ImageGCFailed               node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1270890496 bytes, but freed 0 bytes
47m         Warning   FreeDiskSpaceFailed         node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1272733696 bytes, but freed 0 bytes
47m         Warning   ImageGCFailed               node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1272733696 bytes, but freed 0 bytes
32m         Warning   FreeDiskSpaceFailed         node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 930242560 bytes, but freed 0 bytes
27m         Warning   FreeDiskSpaceFailed         node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1151131648 bytes, but freed 0 bytes
27m         Warning   ImageGCFailed               node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1151131648 bytes, but freed 0 bytes
22m         Warning   FreeDiskSpaceFailed         node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1152888832 bytes, but freed 0 bytes
22m         Warning   ImageGCFailed               node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1152888832 bytes, but freed 0 bytes
17m         Warning   FreeDiskSpaceFailed         node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1154555904 bytes, but freed 0 bytes
17m         Warning   ImageGCFailed               node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1154555904 bytes, but freed 0 bytes
12m         Warning   FreeDiskSpaceFailed         node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1146241024 bytes, but freed 0 bytes
12m         Warning   ImageGCFailed               node/cluster1-node1   failed to garbage collect required amount of images. Wanted to free 1146241024 bytes, but freed 0 bytes
10m         Warning   EvictionThresholdMet        node/cluster1-node1   Attempting to reclaim memory
10m         Normal    NodeHasInsufficientMemory   node/cluster1-node1   Node cluster1-node1 status is now: NodeHasInsufficientMemory

3 기타[ | ]

$ kubectl get events
LASTSEEN   FIRSTSEEN   COUNT     NAME            KIND      SUBOBJECT                         TYPE      REASON
0s         0s          1         dapi-test-pod   Pod                                         Warning   InvalidEnvironmentVariableNames   kubelet, 127.0.0.1      Keys [1badkey, 2alsobad] from the EnvFrom secret default/mysecret were skipped since they are considered invalid environment variable names.
$ kubectl get events
LASTSEEN   FIRSTSEEN   COUNT     NAME                         KIND         SUBOBJECT                    TYPE      REASON                  SOURCE                  MESSAGE
3m         3m          1         guestbook-75786d799f-r6mxl   Pod                                       Normal    Scheduled               default-scheduler       Successfully assigned guestbook-75786d799f-r6mxl to 10.77.155.84
3m         3m          1         guestbook-75786d799f-r6mxl   Pod                                       Normal    SuccessfulMountVolume   kubelet, 10.77.155.84   MountVolume.SetUp succeeded for volume "default-token-5rlxc"
3m         3m          1         guestbook-75786d799f-r6mxl   Pod          spec.containers{guestbook}   Normal    Pulled                  kubelet, 10.77.155.84   Container image "ibmcom/guestbook:v1" already present on machine
3m         3m          1         guestbook-75786d799f-r6mxl   Pod          spec.containers{guestbook}   Normal    Created                 kubelet, 10.77.155.84   Created container
3m         3m          1         guestbook-75786d799f-r6mxl   Pod          spec.containers{guestbook}   Normal    Started                 kubelet, 10.77.155.84   Started container
3m         3m          1         guestbook-75786d799f-xvpvv   Pod          spec.containers{guestbook}   Normal    Killing                 kubelet, 10.77.155.84   Killing container with id docker://guestbook:Need to kill Pod
3m         3m          1         guestbook-75786d799f         ReplicaSet                                Normal    SuccessfulDelete        replicaset-controller   Deleted pod: guestbook-75786d799f-xvpvv
3m         3m          1         guestbook-75786d799f         ReplicaSet                                Normal    SuccessfulCreate        replicaset-controller   Created pod: guestbook-75786d799f-r6mxl
3m         3m          1         guestbook                    Deployment                                Normal    ScalingReplicaSet       deployment-controller   Scaled down replica set guestbook-75786d799f to 0
3m         3m          1         guestbook                    Deployment                                Normal    ScalingReplicaSet       deployment-controller   Scaled up replica set guestbook-75786d799f to 1
$ kubectl get events
LASTSEEN   FIRSTSEEN   COUNT     NAME                         KIND         SUBOBJECT                    TYPE      REASON                  SOURCE                  MESSAGE
3m         3m          1         guestbook-75786d799f-r6mxl   Pod                                       Normal    Scheduled               default-scheduler       Successfully assigned guestbook-75786d799f-r6mxl to 10.77.155.84
3m         3m          1         guestbook-75786d799f-r6mxl   Pod                                       Normal    SuccessfulMountVolume   kubelet, 10.77.155.84   MountVolume.SetUp succeeded for volume "default-token-5rlxc"
3m         3m          1         guestbook-75786d799f-r6mxl   Pod          spec.containers{guestbook}   Normal    Pulled                  kubelet, 10.77.155.84   Container image "ibmcom/guestbook:v1" already present on machine
3m         3m          1         guestbook-75786d799f-r6mxl   Pod          spec.containers{guestbook}   Normal    Created                 kubelet, 10.77.155.84   Created container
3m         3m          1         guestbook-75786d799f-r6mxl   Pod          spec.containers{guestbook}   Normal    Started                 kubelet, 10.77.155.84   Started container
3m         3m          1         guestbook-75786d799f-xvpvv   Pod          spec.containers{guestbook}   Normal    Killing                 kubelet, 10.77.155.84   Killing container with id docker://guestbook:Need to kill Pod
3m         3m          1         guestbook-75786d799f         ReplicaSet                                Normal    SuccessfulDelete        replicaset-controller   Deleted pod: guestbook-75786d799f-xvpvv
3m         3m          1         guestbook-75786d799f         ReplicaSet                                Normal    SuccessfulCreate        replicaset-controller   Created pod: guestbook-75786d799f-r6mxl
3m         3m          1         guestbook                    Deployment                                Normal    ScalingReplicaSet       deployment-controller   Scaled down replica set guestbook-75786d799f to 0
3m         3m          1         guestbook                    Deployment                                Normal    ScalingReplicaSet       deployment-controller   Scaled up replica set guestbook-75786d799f to 1
$ kubectl get events --all-namespaces
NAMESPACE     LAST SEEN   TYPE      REASON                    OBJECT                                           MESSAGE
default       36m         Normal    NodeHasSufficientMemory   node/master                                      Node master status is now: NodeHasSufficientMemory
default       36m         Normal    NodeHasNoDiskPressure     node/master                                      Node master status is now: NodeHasNoDiskPressure
default       36m         Normal    NodeHasSufficientPID      node/master                                      Node master status is now: NodeHasSufficientPID
default       35m         Normal    RegisteredNode            node/master                                      Node master event: Registered Node master in Controller
default       35m         Normal    Starting                  node/master                                      Starting kube-proxy.
default       35m         Normal    RegisteredNode            node/node01                                      Node node01 event: Registered Node node01 in Controller
default       35m         Normal    Starting                  node/node01                                      Starting kubelet.
default       35m         Normal    NodeHasSufficientMemory   node/node01                                      Node node01 status is now: NodeHasSufficientMemory
default       35m         Normal    NodeHasNoDiskPressure     node/node01                                      Node node01 status is now: NodeHasNoDiskPressure
default       35m         Normal    NodeHasSufficientPID      node/node01                                      Node node01 status is now: NodeHasSufficientPID
default       35m         Normal    NodeAllocatableEnforced   node/node01                                      Updated Node Allocatable limit across pods
default       35m         Normal    Starting                  node/node01                                      Starting kube-proxy.
default       35m         Normal    NodeReady                 node/node01                                      Node node01 status is now: NodeReady
kube-system   35m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   33m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   32m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   30m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   29m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   27m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   26m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   21m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   15m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   13m         Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   7m8s        Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   5m40s       Normal    LeaderElection            endpoints/cloud-controller-manager               katacoda-cloud-provider-758cf7cf75-ktdmf-external-cloud-controller became leader
kube-system   35m         Warning   FailedScheduling          pod/coredns-fb8b8dccf-wcmr6                      0/1 nodes are available: 1 node(s) had taints that the pod didn't tolerate.
kube-system   35m         Warning   FailedScheduling          pod/coredns-fb8b8dccf-wcmr6                      0/2 nodes are available: 2 node(s) had taints that the pod didn't tolerate.
kube-system   35m         Normal    Scheduled                 pod/coredns-fb8b8dccf-wcmr6                      Successfully assigned kube-system/coredns-fb8b8dccf-wcmr6 to node01
kube-system   35m         Normal    Pulled                    pod/coredns-fb8b8dccf-wcmr6                      Container image "k8s.gcr.io/coredns:1.3.1" already present on machine
kube-system   35m         Normal    Created                   pod/coredns-fb8b8dccf-wcmr6                      Created container coredns
kube-system   35m         Normal    Started                   pod/coredns-fb8b8dccf-wcmr6                      Started container coredns
kube-system   35m         Warning   FailedScheduling          pod/coredns-fb8b8dccf-xs6lx                      0/1 nodes are available: 1 node(s) had taints that the pod didn't tolerate.
kube-system   35m         Warning   FailedScheduling          pod/coredns-fb8b8dccf-xs6lx                      0/2 nodes are available: 2 node(s) had taints that the pod didn't tolerate.
kube-system   35m         Normal    Scheduled                 pod/coredns-fb8b8dccf-xs6lx                      Successfully assigned kube-system/coredns-fb8b8dccf-xs6lx to node01
kube-system   35m         Normal    Pulled                    pod/coredns-fb8b8dccf-xs6lx                      Container image "k8s.gcr.io/coredns:1.3.1" already present on machine
kube-system   35m         Normal    Created                   pod/coredns-fb8b8dccf-xs6lx                      Created container coredns
kube-system   35m         Normal    Started                   pod/coredns-fb8b8dccf-xs6lx                      Started container coredns
kube-system   35m         Normal    SuccessfulCreate          replicaset/coredns-fb8b8dccf                     Created pod: coredns-fb8b8dccf-xs6lx
kube-system   35m         Normal    SuccessfulCreate          replicaset/coredns-fb8b8dccf                     Created pod: coredns-fb8b8dccf-wcmr6
kube-system   35m         Normal    ScalingReplicaSet         deployment/coredns                               Scaled up replica set coredns-fb8b8dccf to 2
kube-system   21m         Normal    Scheduled                 pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Successfully assigned kube-system/dash-kubernetes-dashboard-dff4ccb96-h5pbw to node01
kube-system   21m         Normal    Pulling                   pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Pulling image "k8s.gcr.io/kubernetes-dashboard-amd64:v1.10.1"
kube-system   21m         Normal    Pulled                    pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Successfully pulled image "k8s.gcr.io/kubernetes-dashboard-amd64:v1.10.1"
kube-system   21m         Normal    Created                   pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Created container kubernetes-dashboard
kube-system   21m         Normal    Started                   pod/dash-kubernetes-dashboard-dff4ccb96-h5pbw    Started container kubernetes-dashboard
kube-system   21m         Normal    SuccessfulCreate          replicaset/dash-kubernetes-dashboard-dff4ccb96   Created pod: dash-kubernetes-dashboard-dff4ccb96-h5pbw
kube-system   21m         Normal    ScalingReplicaSet         deployment/dash-kubernetes-dashboard             Scaled up replica set dash-kubernetes-dashboard-dff4ccb96 to 1
kube-system   36m         Normal    Pulled                    pod/etcd-master                                  Container image "k8s.gcr.io/etcd:3.3.10" already present on machine
kube-system   36m         Normal    Created                   pod/etcd-master                                  Created container etcd
kube-system   36m         Normal    Started                   pod/etcd-master                                  Started container etcd
kube-system   18m         Warning   Unhealthy                 pod/etcd-master                                  Liveness probe failed: Error: context deadline exceeded
kube-system   35m         Warning   FailedScheduling          pod/katacoda-cloud-provider-758cf7cf75-ktdmf     0/1 nodes are available: 1 node(s) had taints that the pod didn't tolerate.
kube-system   35m         Warning   FailedScheduling          pod/katacoda-cloud-provider-758cf7cf75-ktdmf     0/2 nodes are available: 2 node(s) had taints that the pod didn't tolerate.
kube-system   35m         Normal    Scheduled                 pod/katacoda-cloud-provider-758cf7cf75-ktdmf     Successfully assigned kube-system/katacoda-cloud-provider-758cf7cf75-ktdmf to node01
kube-system   35m         Normal    Pulling                   pod/katacoda-cloud-provider-758cf7cf75-ktdmf     Pulling image "katacoda/katacoda-cloud-provider:0.0.1"
kube-system   35m         Normal    Pulled                    pod/katacoda-cloud-provider-758cf7cf75-ktdmf     Successfully pulled image "katacoda/katacoda-cloud-provider:0.0.1"
kube-system   33m         Normal    Created                   pod/katacoda-cloud-provider-758cf7cf75-ktdmf     Created container katacoda-cloud-provider
kube-system   33m         Normal    Started                   pod/katacoda-cloud-provider-758cf7cf75-ktdmf     Started container katacoda-cloud-provider
kube-system   5m20s       Warning   Unhealthy                 pod/katacoda-cloud-provider-758cf7cf75-ktdmf     Liveness probe failed: Get http://127.0.0.1:10252/healthz: dial tcp 127.0.0.1:10252: connect: connection refused
kube-system   32m         Normal    Killing                   pod/katacoda-cloud-provider-758cf7cf75-ktdmf     Container katacoda-cloud-provider failed liveness probe, will be restarted
kube-system   33m         Normal    Pulled                    pod/katacoda-cloud-provider-758cf7cf75-ktdmf     Container image "katacoda/katacoda-cloud-provider:0.0.1" already present on machine
kube-system   23s         Warning   BackOff                   pod/katacoda-cloud-provider-758cf7cf75-ktdmf     Back-off restarting failed container
kube-system   35m         Normal    SuccessfulCreate          replicaset/katacoda-cloud-provider-758cf7cf75    Created pod: katacoda-cloud-provider-758cf7cf75-ktdmf
kube-system   35m         Normal    ScalingReplicaSet         deployment/katacoda-cloud-provider               Scaled up replica set katacoda-cloud-provider-758cf7cf75 to 1
kube-system   36m         Normal    Pulled                    pod/kube-apiserver-master                        Container image "k8s.gcr.io/kube-apiserver:v1.14.0" already present on machine
kube-system   36m         Normal    Created                   pod/kube-apiserver-master                        Created container kube-apiserver
kube-system   36m         Normal    Started                   pod/kube-apiserver-master                        Started container kube-apiserver
kube-system   35m         Warning   Unhealthy                 pod/kube-apiserver-master                        Liveness probe failed: HTTP probe failed with statuscode: 403
kube-system   18m         Warning   Unhealthy                 pod/kube-apiserver-master                        Liveness probe failed: HTTP probe failed with statuscode: 500
kube-system   36m         Normal    Pulled                    pod/kube-controller-manager-master               Container image "k8s.gcr.io/kube-controller-manager:v1.14.0" already present on machine
kube-system   36m         Normal    Created                   pod/kube-controller-manager-master               Created container kube-controller-manager
kube-system   36m         Normal    Started                   pod/kube-controller-manager-master               Started container kube-controller-manager
kube-system   35m         Normal    LeaderElection            endpoints/kube-controller-manager                master_fba4bedb-5207-11ea-bbaa-0242ac110043 became leader
kube-system   35m         Normal    Scheduled                 pod/kube-keepalived-vip-xs784                    Successfully assigned kube-system/kube-keepalived-vip-xs784 tonode01
kube-system   35m         Normal    Pulling                   pod/kube-keepalived-vip-xs784                    Pulling image "gcr.io/google_containers/kube-keepalived-vip:0.9"
kube-system   35m         Normal    Pulled                    pod/kube-keepalived-vip-xs784                    Successfully pulled image "gcr.io/google_containers/kube-keepalived-vip:0.9"
kube-system   35m         Normal    Created                   pod/kube-keepalived-vip-xs784                    Created container kube-keepalived-vip
kube-system   35m         Normal    Started                   pod/kube-keepalived-vip-xs784                    Started container kube-keepalived-vip
kube-system   35m         Normal    SuccessfulCreate          daemonset/kube-keepalived-vip                    Created pod: kube-keepalived-vip-xs784
kube-system   35m         Normal    Scheduled                 pod/kube-proxy-xwmrv                             Successfully assigned kube-system/kube-proxy-xwmrv to master
kube-system   35m         Normal    Pulled                    pod/kube-proxy-xwmrv                             Container image "k8s.gcr.io/kube-proxy:v1.14.0" already present on machine
kube-system   35m         Normal    Created                   pod/kube-proxy-xwmrv                             Created container kube-proxy
kube-system   35m         Normal    Started                   pod/kube-proxy-xwmrv                             Started container kube-proxy
kube-system   35m         Normal    Scheduled                 pod/kube-proxy-zm2lb                             Successfully assigned kube-system/kube-proxy-zm2lb to node01
kube-system   35m         Normal    Pulling                   pod/kube-proxy-zm2lb                             Pulling image "k8s.gcr.io/kube-proxy:v1.14.0"
kube-system   35m         Normal    Pulled                    pod/kube-proxy-zm2lb                             Successfully pulled image "k8s.gcr.io/kube-proxy:v1.14.0"
kube-system   35m         Normal    Created                   pod/kube-proxy-zm2lb                             Created container kube-proxy
kube-system   35m         Normal    Started                   pod/kube-proxy-zm2lb                             Started container kube-proxy
kube-system   35m         Normal    SuccessfulCreate          daemonset/kube-proxy                             Created pod: kube-proxy-xwmrv
kube-system   35m         Normal    SuccessfulCreate          daemonset/kube-proxy                             Created pod: kube-proxy-zm2lb
kube-system   36m         Normal    Pulled                    pod/kube-scheduler-master                        Container image "k8s.gcr.io/kube-scheduler:v1.14.0" already present on machine
kube-system   36m         Normal    Created                   pod/kube-scheduler-master                        Created container kube-scheduler
kube-system   36m         Normal    Started                   pod/kube-scheduler-master                        Started container kube-scheduler
kube-system   35m         Normal    LeaderElection            endpoints/kube-scheduler                         master_f3ff19a8-5207-11ea-88c0-0242ac110043 became leader
kube-system   35m         Normal    Scheduled                 pod/weave-net-2l2g8                              Successfully assigned kube-system/weave-net-2l2g8 to master
kube-system   35m         Normal    Pulled                    pod/weave-net-2l2g8                              Container image "weaveworks/weave-kube:2.5.1" already present on machine
kube-system   35m         Normal    Created                   pod/weave-net-2l2g8                              Created container weave
kube-system   35m         Normal    Started                   pod/weave-net-2l2g8                              Started container weave
kube-system   35m         Normal    Pulled                    pod/weave-net-2l2g8                              Container image "weaveworks/weave-npc:2.5.1" already present on machine
kube-system   35m         Normal    Created                   pod/weave-net-2l2g8                              Created container weave-npc
kube-system   35m         Normal    Started                   pod/weave-net-2l2g8                              Started container weave-npc
kube-system   35m         Warning   Unhealthy                 pod/weave-net-2l2g8                              Readiness probe failed: Get http://127.0.0.1:6784/status: dialtcp 127.0.0.1:6784: connect: connection refused
kube-system   35m         Normal    Scheduled                 pod/weave-net-xc54f                              Successfully assigned kube-system/weave-net-xc54f to node01
kube-system   35m         Normal    Pulled                    pod/weave-net-xc54f                              Container image "weaveworks/weave-kube:2.5.1" already present on machine
kube-system   35m         Normal    Created                   pod/weave-net-xc54f                              Created container weave
kube-system   35m         Normal    Started                   pod/weave-net-xc54f                              Started container weave
kube-system   35m         Normal    Pulled                    pod/weave-net-xc54f                              Container image "weaveworks/weave-npc:2.5.1" already present on machine
kube-system   35m         Normal    Created                   pod/weave-net-xc54f                              Created container weave-npc
kube-system   35m         Normal    Started                   pod/weave-net-xc54f                              Started container weave-npc
kube-system   35m         Warning   Unhealthy                 pod/weave-net-xc54f                              Readiness probe failed: Get http://127.0.0.1:6784/status: dialtcp 127.0.0.1:6784: connect: connection refused
kube-system   35m         Normal    SuccessfulCreate          daemonset/weave-net                              Created pod: weave-net-2l2g8
kube-system   35m         Normal    SuccessfulCreate          daemonset/weave-net                              Created pod: weave-net-xc54f
$ kubectl get event
LAST SEEN   TYPE      REASON             OBJECT                          MESSAGE
13s         Warning   Unhealthy          pod/app-prod-6bc4658c66-8hmlp   Liveness probe failed: Get http://10.20.1.8:80/: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
13s         Warning   Unhealthy          pod/app-prod-6bc4658c66-8npgc   Liveness probe failed: Get http://10.20.3.5:80/: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
6m5s        Warning   NodeSysctlChange   node/node01                     {"unmanaged": {"fs.aio-nr": "64"}}
90s         Warning   NodeSysctlChange   node/node02                     {"unmanaged": {"fs.aio-nr": "5450"}}
$ kubectl describe pod/app-prod-6bc4658c66-8hmlp | grep Events: -A99
Events:
  Type     Reason     Age                    From              Message
  ----     ------     ----                   ----              -------
  Warning  Unhealthy  117s (x3706 over 16d)  kubelet, node01   Liveness probe failed: Get http://10.20.1.8:80/: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
$ kubectl get event --field-selector=involvedObject.name=app-prod-6bc4658c66-8hmlp -owide
LAST SEEN   TYPE      REASON      OBJECT                          SUBOBJECT              SOURCE            MESSAGE                                                                                                                        FIRST SEEN   COUNT   NAME
22s         Warning   Unhealthy   pod/app-prod-6bc4658c66-8hmlp   spec.containers{app}   kubelet, node01   Liveness probe failed: Get http://10.20.1.8:80/: net/http: request canceled (Client.Timeout exceeded while awaiting headers)   16d          3712    app-prod-6bc4658c66-8hmlp.16125180cbb5a1e9
$ kubectl describe node/node01 | grep Events: -A99
Events:
  Type     Reason            Age                 From                    Message
  ----     ------            ----                ----                    -------
  Warning  NodeSysctlChange  19m (x66 over 16d)  sysctl-monitor, node01  {"unmanaged": {"fs.aio-nr": "64"}}

4 같이 보기[ | ]

5 참고[ | ]

문서 댓글 ({{ doc_comments.length }})
{{ comment.name }} {{ comment.created | snstime }}