-
Notifications
You must be signed in to change notification settings - Fork 121
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
chaoskube is restarting with status as Running and crashLoopBackoff #202
Comments
Please let me know your output of |
Hi , root@kmaster:~# helm version --short root@kmaster:~# helm install chaoskube stable/chaoskube -n chaoskube --set dryRun=false --debug install.go:159: [debug] Original chart version: "" client.go:108: [debug] creating 1 resource(s) COMPUTED VALUES: HOOKS:
|
Please make sure you have permission to read the logs. Currently you get: $ kubectl -n chaoskube logs -f $POD
Error from server (InternalError): Internal error occurred: Authorization error (user=kube-apiserver-kubelet-client, verb=get, resource=nodes, subresource=proxy) Please set the helm install chaoskube stable/chaoskube -n chaoskube --set dryRun=false,debug=true |
I am getting below issues Error: cannot re-use a name that is still in use |
If i set debug =true also not getting any logs You can follow the logs to see what chaoskube does:
root@kmaster: |
Can i try in minikube server with chaoskube |
I restarted kubectl and daemon services and tried still getting this issues unable to get pod detsila chaoskube is restarting not killing pods namespace/chaoskube created You can follow the logs to see what chaoskube does:
root@kmaster: |
I ran with chaoskube01 & below are the details client.go:108: [debug] creating 1 resource(s) COMPUTED VALUES: HOOKS:
|
Please capture the logs of chaoskube so we can see what's going wrong:
Just use |
Getting timeout issues NAMESPACE NAME READY STATUS RESTARTS AGE |
Hi Martin, I have deleted chaoskube pod and started running & observing forbidden. Appreciate your inputs on this. Regards, |
Hi Martin, Chaoskube is not killing random pods. Appreciate your help on this.. Regards, |
It doesn't have permission to connect to the Kubernetes API. Install your Helm chart with |
Hi Martin, After updating rbac.create =true You can follow the logs to see what chaoskube does:
You are running in dry-run mode. No pod is actually terminated. |
Hi Martin, I am getting only the messages for terminating but it is not actually killing the pods time="2020-05-12T15:22:54Z" level=info msg="terminating pod" name=kube-proxy-2qjcz namespace=kube-system I can see the nginx is still running. Regards, |
Because you re-enabled dry-run mode:
Make sure to use |
Every 2.0s: kubectl get deployments,pods --all-namespaces kmaster: Mon Apr 27 17:58:26 2020
NAMESPACE NAME READY UP-TO-DATE AVAILABLE AGE
default deployment.apps/apache 0/1 1 0 2d7h
default deployment.apps/chaoskube-1587990114 0/1 1 0 6m28s
default deployment.apps/chaoskube-1587990210 0/1 1 0 4m55s
default deployment.apps/nginx 2/2 2 2 2d6h
kube-system deployment.apps/coredns 2/2 2 2 2d7h
more-apps deployment.apps/chaoskube 0/1 1 0 2d
more-apps deployment.apps/ghost 2/2 2 2 2d6h
NAMESPACE NAME READY STATUS RESTARTS AGE
default pod/apache-8454694d99-xgvrl 0/1 ImagePullBackOff 0 2d7h
default pod/chaoskube-1587990114-79c757f5cb-q8b8j 0/1 CrashLoopBackOff 5 6m28s
default pod/chaoskube-1587990210-795f6d7848-bs8dg 0/1 CrashLoopBackOff 4 4m55s
default pod/nginx-5ccf85b585-cv8m5 1/1 Running 0 2d6h
default pod/nginx-5ccf85b585-ws4xj 1/1 Running 0 2d6h
kube-system pod/coredns-66bff467f8-jnx5b 1/1 Running 0 2d7h
kube-system pod/coredns-66bff467f8-s9qp5 1/1 Running 0 2d7h
kube-system pod/etcd-kmaster 1/1 Running 0 2d7h
kube-system pod/kube-apiserver-kmaster 1/1 Running 0 2d7h
kube-system pod/kube-controller-manager-kmaster 1/1 Running 0 2d7h
kube-system pod/kube-flannel-ds-amd64-6sd7v 1/1 Running 0 2d7h
kube-system pod/kube-flannel-ds-amd64-9gg8w 1/1 Running 2 2d7h
kube-system pod/kube-flannel-ds-amd64-bqvbg 1/1 Running 0 2d7h
kube-system pod/kube-proxy-58zlr 1/1 Running 0 2d7h
kube-system pod/kube-proxy-dp6xb 1/1 Running 0 2d7h
kube-system pod/kube-proxy-vm75r 1/1 Running 0 2d7h
kube-system pod/kube-scheduler-kmaster 1/1 Running 0 2d7h
more-apps pod/chaoskube-56998c669c-kp5rx 0/1 CrashLoopBackOff 16 2d
more-apps pod/ghost-588cb7bd9f-746bg 1/1 Running 0 2d6h
more-apps pod/ghost-588cb7bd9f-9qtm6 1/1 Running 0 2d6h
Details of deployment file details
root@kmaster:~# helm install stable/chaoskube --generate-name
NAME: chaoskube-1587990114
LAST DEPLOYED: Mon Apr 27 17:51:58 2020
NAMESPACE: default
STATUS: deployed
REVISION: 1
TEST SUITE: None
NOTES:
chaoskube is running and will kill arbitrary pods every 10m.
You can follow the logs to see what chaoskube does:
You are running in dry-run mode. No pod is actually terminated.
root@kmaster:~# helm install stable/chaoskube --generate-name --debug --set dryRun=false
install.go:159: [debug] Original chart version: ""
install.go:176: [debug] CHART PATH: /root/.cache/helm/repository/chaoskube-3.1.4.tgz
client.go:108: [debug] creating 1 resource(s)
NAME: chaoskube-1587990210
LAST DEPLOYED: Mon Apr 27 17:53:31 2020
NAMESPACE: default
STATUS: deployed
REVISION: 1
TEST SUITE: None
USER-SUPPLIED VALUES:
dryRun: false
COMPUTED VALUES:
affinity: {}
annotations: null
debug: false
dryRun: false
excludedDaysOfYear: null
excludedPodNames: null
excludedTimesOfDay: null
excludedWeekdays: null
gracePeriod: -1s
image: quay.io/linki/chaoskube
imageTag: v0.14.0
includedPodNames: null
interval: 10m
labels: null
logFormat: null
metrics:
enabled: false
port: 8080
service:
port: 8080
type: ClusterIP
serviceMonitor:
additionalLabels: {}
enabled: false
minimumAge: 0s
name: chaoskube
namespaces: null
nodeSelector: {}
podAnnotations: {}
podLabels: {}
priorityClassName: ""
rbac:
create: false
serviceAccountName: default
replicas: 1
resources: {}
timezone: UTC
tolerations: []
HOOKS:
MANIFEST:
Source: chaoskube/templates/deployment.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
name: chaoskube-1587990210
labels:
app.kubernetes.io/name: chaoskube
app.kubernetes.io/managed-by: "Helm"
app.kubernetes.io/instance: "chaoskube-1587990210"
helm.sh/chart: chaoskube-3.1.4
spec:
replicas: 1
selector:
matchLabels:
app.kubernetes.io/name: chaoskube
app.kubernetes.io/instance: chaoskube-1587990210
template:
metadata:
labels:
app.kubernetes.io/name: chaoskube
app.kubernetes.io/managed-by: "Helm"
app.kubernetes.io/instance: "chaoskube-1587990210"
helm.sh/chart: chaoskube-3.1.4
spec:
containers:
- name: chaoskube
image: quay.io/linki/chaoskube:v0.14.0
args:
- --interval=10m
- --labels=
- --annotations=
- --namespaces=
- --no-dry-run
- --excluded-weekdays=
- --excluded-times-of-day=
- --excluded-days-of-year=
- --timezone=UTC
- --minimum-age=0s
- --grace-period=-1s
- --metrics-address=
resources:
{}
securityContext:
runAsNonRoot: true
runAsUser: 65534
readOnlyRootFilesystem: true
capabilities:
drop: ["ALL"]
serviceAccountName: "default"
NOTES:
chaoskube is running and will kill arbitrary pods every 10m.
You can follow the logs to see what chaoskube does:
root@kmaster:
# POD=$(kubectl -n default get pods -l='app.kubernetes.io/instance=chaoskube-1587990210' --output=jsonpath='{.items[0].metadata.name}')#root@kmaster:
root@kmaster:
# kubectl -n default logs -f $POD# kubectl describe pods chaoskube-1587990210-795f6d7848-bs8dg -n defaultError from server (NotFound): the server could not find the requested resource ( pods/log chaoskube-1587990210-795f6d7848-bs8dg)
root@kmaster:
Name: chaoskube-1587990210-795f6d7848-bs8dg
Namespace: default
Priority: 0
Node: knode/10.0.3.15
Start Time: Mon, 27 Apr 2020 17:53:31 +0530
Labels: app.kubernetes.io/instance=chaoskube-1587990210
app.kubernetes.io/managed-by=Helm
app.kubernetes.io/name=chaoskube
helm.sh/chart=chaoskube-3.1.4
pod-template-hash=795f6d7848
Annotations:
Status: Running
IP: 192.168.1.16
IPs:
IP: 192.168.1.16
Controlled By: ReplicaSet/chaoskube-1587990210-795f6d7848
Containers:
chaoskube:
Container ID: docker://189430f9e1c73b8d8f91fa83202c39a5c6b090cfd5847684b06cc1bad1a9fc8c
Image: quay.io/linki/chaoskube:v0.14.0
Image ID: docker-pullable://quay.io/linki/chaoskube@sha256:74a8314513d94da26d407f29d2dba621ec9e607f5b2abbe07c3f6a521e00c7a4
Port:
Host Port:
Args:
--interval=10m
--labels=
--annotations=
--namespaces=
--no-dry-run
--excluded-weekdays=
--excluded-times-of-day=
--excluded-days-of-year=
--timezone=UTC
--minimum-age=0s
--grace-period=-1s
--metrics-address=
State: Running
Started: Mon, 27 Apr 2020 17:54:46 +0530
Last State: Terminated
Reason: Error
Exit Code: 1
Started: Mon, 27 Apr 2020 17:54:04 +0530
Finished: Mon, 27 Apr 2020 17:54:34 +0530
Ready: True
Restart Count: 2
Environment:
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from default-token-gdsnt (ro)
Conditions:
Type Status
Initialized True
Ready True
ContainersReady True
PodScheduled True
Volumes:
default-token-gdsnt:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-gdsnt
Optional: false
QoS Class: BestEffort
Node-Selectors:
Tolerations: node.kubernetes.io/not-ready:NoExecute for 300s
node.kubernetes.io/unreachable:NoExecute for 300s
Events:
Type Reason Age From Message
Normal Scheduled 80s default-scheduler Successfully assigned default/chaoskube-1587990210-795f6d7848-bs8dg to knode
Warning BackOff 16s kubelet, knode Back-off restarting failed container
Normal Pulled 5s (x3 over 79s) kubelet, knode Container image "quay.io/linki/chaoskube:v0.14.0" already present on machine
Normal Created 5s (x3 over 78s) kubelet, knode Created container chaoskube
Normal Started 5s (x3 over 78s) kubelet, knode Started container chaoskube
root@kmaster:~# kubectl describe pods chaoskube-1587990210-795f6d7848-bs8dg -n default
Name: chaoskube-1587990210-795f6d7848-bs8dg
Namespace: default
Priority: 0
Node: knode/10.0.3.15
Start Time: Mon, 27 Apr 2020 17:53:31 +0530
Labels: app.kubernetes.io/instance=chaoskube-1587990210
app.kubernetes.io/managed-by=Helm
app.kubernetes.io/name=chaoskube
helm.sh/chart=chaoskube-3.1.4
pod-template-hash=795f6d7848
Annotations:
Status: Running
IP: 192.168.1.16
IPs:
IP: 192.168.1.16
Controlled By: ReplicaSet/chaoskube-1587990210-795f6d7848
Containers:
chaoskube:
Container ID: docker://189430f9e1c73b8d8f91fa83202c39a5c6b090cfd5847684b06cc1bad1a9fc8c
Image: quay.io/linki/chaoskube:v0.14.0
Image ID: docker-pullable://quay.io/linki/chaoskube@sha256:74a8314513d94da26d407f29d2dba621ec9e607f5b2abbe07c3f6a521e00c7a4
Port:
Host Port:
Args:
--interval=10m
--labels=
--annotations=
--namespaces=
--no-dry-run
--excluded-weekdays=
--excluded-times-of-day=
--excluded-days-of-year=
--timezone=UTC
--minimum-age=0s
--grace-period=-1s
--metrics-address=
State: Running
Started: Mon, 27 Apr 2020 17:54:46 +0530
Last State: Terminated
Reason: Error
Exit Code: 1
Started: Mon, 27 Apr 2020 17:54:04 +0530
Finished: Mon, 27 Apr 2020 17:54:34 +0530
Ready: True
Restart Count: 2
Environment:
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from default-token-gdsnt (ro)
Conditions:
Type Status
Initialized True
Ready True
ContainersReady True
PodScheduled True
Volumes:
default-token-gdsnt:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-gdsnt
Optional: false
QoS Class: BestEffort
Node-Selectors:
Tolerations: node.kubernetes.io/not-ready:NoExecute for 300s
node.kubernetes.io/unreachable:NoExecute for 300s
Events:
Type Reason Age From Message
Normal Scheduled 85s default-scheduler Successfully assigned default/chaoskube-1587990210-795f6d7848-bs8dg to knode
Warning BackOff 21s kubelet, knode Back-off restarting failed container
Normal Pulled 10s (x3 over 84s) kubelet, knode Container image "quay.io/linki/chaoskube:v0.14.0" already present on machine
Normal Created 10s (x3 over 83s) kubelet, knode Created container chaoskube
Normal Started 10s (x3 over 83s) kubelet, knode Started container chaoskube
Appreciate your support on this issue...
The text was updated successfully, but these errors were encountered: