Difference between revisions of "FailedScheduling"
Jump to navigation
Jump to search
Line 88: | Line 88: | ||
* [[Killing]] | * [[Killing]] | ||
* <code>[[kubectl get events -A]]</code> | * <code>[[kubectl get events -A]]</code> | ||
+ | [[default-scheduler]] | ||
== See also == | == See also == |
Revision as of 16:56, 13 December 2022
kubectl get events -A your-namespace 22s Warning FailedScheduling pod/your_elasticsearch-master-1 0/8 nodes are available: 1 Insufficient cpu, 1 node(s) had taint {risk: etl}, that the pod didn't tolerate, 1 node(s) were unschedulable, 3 node(s) had taint {NAME: run-your-command}, that the pod didn't tolerate, 7 node(s) had volume node affinity conflict.
0/1 nodes are available: 1 node(s) had taint {node.kubernetes.io/not-ready: }, that the pod didn't tolerate.
your-namespace 2m16s Warning FailedScheduling pod/your-elasticsearch-master-1 0/13 nodes are available: 1 Insufficient cpu, 1 node(s) had taint {risk: etl}, that the pod didn't tolerate, 3 node(s) had taint {Name: run-your-command}, that the pod didn't tolerate, 4 node(s) had volume node affinity conflict, 4 node(s) were unschedulable.
default 8m51s Warning FailedScheduling pod/myprometheus-alertmanager-5967d4ff85-5glkh running PreBind plugin "VolumeBinding": binding volumes: timed out waiting for the condition default 4m58s Normal ExternalProvisioning persistentvolumeclaim/myprometheus-alertmanager waiting for a volume to be created, either by external provisioner "ebs.csi.aws.com" or manually created by system administrator
default 32s Warning FailedScheduling pod/elasticsearch-master-0 0/2 nodes are available: 2 Insufficient cpu, 2 Insufficient memory.
default 4m7s Warning FailedScheduling pod/kibana-kibana-654ccb45bd-pbp4r 0/2 nodes are available: 2 Insufficient cpu.
default 18m Warning FailedScheduling pod/sentry-web-79b8bbccdf-r2qr4 0/4 nodes are available: 4 pod has unbound immediate PersistentVolumeClaims.
kubectl get events -o yaml
kubectl get events -o yaml | grep message message: '0/2 nodes are available: 2 Insufficient cpu.' message: 'pod didn''t trigger scale-up: '
apiVersion: v1 items: - apiVersion: v1 count: 83 eventTime: null firstTimestamp: "2022-11-01T06:21:01Z" involvedObject: apiVersion: v1 kind: Pod name: kibana-kibana-654ccb45bd-pbp4r namespace: default resourceVersion: "800158" uid: b10009ea-51df-47b8-be45-3e616f192f1a kind: Event lastTimestamp: "2022-11-01T07:46:28Z" message: '0/2 nodes are available: 2 Insufficient cpu.' metadata: creationTimestamp: "2022-11-01T06:21:01Z" name: kibana-kibana-654ccb45bd-pbp4r.172361a82959aa95 namespace: default resourceVersion: "2825" uid: 13937a60-c494-47fa-aaa0-5380a6e1412f reason: FailedScheduling reportingComponent: "" reportingInstance: "" source: component: default-scheduler type: Warning - apiVersion: v1 count: 511 eventTime: null firstTimestamp: "2022-11-01T06:21:04Z" involvedObject: apiVersion: v1 kind: Pod name: kibana-kibana-654ccb45bd-pbp4r namespace: default resourceVersion: "800163" uid: b10009ea-51df-47b8-be45-3e616f192f1a kind: Event lastTimestamp: "2022-11-01T07:46:18Z" message: 'pod didnt trigger scale-up: ' metadata: creationTimestamp: "2022-11-01T06:21:04Z" name: kibana-kibana-654ccb45bd-pbp4r.172361a8a65f467d namespace: default resourceVersion: "2824" uid: 38724dae-0dde-42e3-848b-fe24bbbccda3 reason: NotTriggerScaleUp reportingComponent: "" reportingInstance: "" source: component: cluster-autoscaler type: Normal kind: List metadata: resourceVersion: ""
Related
default-scheduler
See also
- Kubernetes events, Kubernetes node events:
[ Normal | Warning ]
,kubectl events, kubectl get events
:BackOff
,FailedMount, FailedAttachVolume, TaintManagerEviction, InvalidDiskCapacity
,FailedScheduling, kubectl describe events, Unhealthy
,conditions:, UpdatedLoadBalancer, BackoffLimitExceeded
- K8s troubleshooting:
kubectl logs, kubectl top, kubectl get events -A, kubectl describe pod
, Liveness, Readiness,Kubernetes events
, Pulling image, OOMKilled, ProbeWarning, Reason,FailedScheduling
,errImagePull, ImagePullBackOff
, Kubelet conditions:MemoryPressure, DiskPressure, KubeletHasSufficientPID, KubeletReady, kubectl [ debug | attach | exec ] kubectl cluster-info dump, SimKube, KWOK
Advertising: