Difference between revisions of "FailedScheduling"
Jump to navigation
Jump to search
Line 1: | Line 1: | ||
− | |||
− | |||
[[kubectl get events -A]] | [[kubectl get events -A]] | ||
your-namespace 22s Warning [[FailedScheduling]] pod/your_elasticsearch-master-1 0/8 nodes are available: 1 [[Insufficient cpu]], 1 node(s) had taint {risk: etl}, that the pod didn't tolerate, 1 node(s) were unschedulable, 3 node(s) had [[taint]] {NAME: run-your-command}, that the pod didn't tolerate, 7 node(s) had [[volume node affinity conflict]]. | your-namespace 22s Warning [[FailedScheduling]] pod/your_elasticsearch-master-1 0/8 nodes are available: 1 [[Insufficient cpu]], 1 node(s) had taint {risk: etl}, that the pod didn't tolerate, 1 node(s) were unschedulable, 3 node(s) had [[taint]] {NAME: run-your-command}, that the pod didn't tolerate, 7 node(s) had [[volume node affinity conflict]]. |
Revision as of 14:50, 6 November 2022
kubectl get events -A your-namespace 22s Warning FailedScheduling pod/your_elasticsearch-master-1 0/8 nodes are available: 1 Insufficient cpu, 1 node(s) had taint {risk: etl}, that the pod didn't tolerate, 1 node(s) were unschedulable, 3 node(s) had taint {NAME: run-your-command}, that the pod didn't tolerate, 7 node(s) had volume node affinity conflict.
your-namespace 2m16s Warning FailedScheduling pod/your-elasticsearch-master-1 0/13 nodes are available: 1 Insufficient cpu, 1 node(s) had taint {risk: etl}, that the pod didn't tolerate, 3 node(s) had taint {Name: run-your-command}, that the pod didn't tolerate, 4 node(s) had volume node affinity conflict, 4 node(s) were unschedulable.
default 8m51s Warning FailedScheduling pod/myprometheus-alertmanager-5967d4ff85-5glkh running PreBind plugin "VolumeBinding": binding volumes: timed out waiting for the condition default 4m58s Normal ExternalProvisioning persistentvolumeclaim/myprometheus-alertmanager waiting for a volume to be created, either by external provisioner "ebs.csi.aws.com" or manually created by system administrator
default 32s Warning FailedScheduling pod/elasticsearch-master-0 0/2 nodes are available: 2 Insufficient cpu, 2 Insufficient memory.
default 4m7s Warning FailedScheduling pod/kibana-kibana-654ccb45bd-pbp4r 0/2 nodes are available: 2 Insufficient cpu.
kubectl get events -o yaml
kubectl get events -o yaml | grep message message: '0/2 nodes are available: 2 Insufficient cpu.' message: 'pod didn''t trigger scale-up: '
apiVersion: v1 items: - apiVersion: v1 count: 83 eventTime: null firstTimestamp: "2022-11-01T06:21:01Z" involvedObject: apiVersion: v1 kind: Pod name: kibana-kibana-654ccb45bd-pbp4r namespace: default resourceVersion: "800158" uid: b10009ea-51df-47b8-be45-3e616f192f1a kind: Event lastTimestamp: "2022-11-01T07:46:28Z" message: '0/2 nodes are available: 2 Insufficient cpu.' metadata: creationTimestamp: "2022-11-01T06:21:01Z" name: kibana-kibana-654ccb45bd-pbp4r.172361a82959aa95 namespace: default resourceVersion: "2825" uid: 13937a60-c494-47fa-aaa0-5380a6e1412f reason: FailedScheduling reportingComponent: "" reportingInstance: "" source: component: default-scheduler type: Warning - apiVersion: v1 count: 511 eventTime: null firstTimestamp: "2022-11-01T06:21:04Z" involvedObject: apiVersion: v1 kind: Pod name: kibana-kibana-654ccb45bd-pbp4r namespace: default resourceVersion: "800163" uid: b10009ea-51df-47b8-be45-3e616f192f1a kind: Event lastTimestamp: "2022-11-01T07:46:18Z" message: 'pod didnt trigger scale-up: ' metadata: creationTimestamp: "2022-11-01T06:21:04Z" name: kibana-kibana-654ccb45bd-pbp4r.172361a8a65f467d namespace: default resourceVersion: "2824" uid: 38724dae-0dde-42e3-848b-fe24bbbccda3 reason: NotTriggerScaleUp reportingComponent: "" reportingInstance: "" source: component: cluster-autoscaler type: Normal kind: List metadata: resourceVersion: ""
Related
See also
- Kubernetes events, Kubernetes node events:
[ Normal | Warning ]
,kubectl events, kubectl get events
:BackOff
,FailedMount, FailedAttachVolume, TaintManagerEviction, InvalidDiskCapacity
,FailedScheduling, kubectl describe events, Unhealthy
,conditions:, UpdatedLoadBalancer, BackoffLimitExceeded
kubectl get
: [all
|componentstatus | pods | apiservices | events
|event
|nodes
|replicaset
|services
|svc
|namespaces
|hpa
|secret | deployments | serviceaccounts | endpoints | daemonset | networkpolicy
],kubectl get -h
Advertising: