Hey guys,
I’ve seen this happen before, too. It happened again today. After leaving my cluster alone for a few days, the two helloworld pods that I have are now in some weird state. They show as this:
NAME READY STATUS RESTARTS AGE
helloworld-v1.0-54655d76f-c2fbp 0/2 Init:CrashLoopBackOff 46 4d21h
helloworld-v1.0-54655d76f-fskpw 0/2 Init:CrashLoopBackOff 46 4d21h
Notice how it has restarted 46 times already.
However, I can still hit the helloworld service with curl and get a proper response. And the logs do show those requests with a 200 status code:
127.0.0.1 - - [18/Feb/2020 22:13:43] "GET /hello HTTP/1.1" 200 -
127.0.0.1 - - [18/Feb/2020 22:22:48] "GET /hello HTTP/1.1" 200 -
The log for istio-init does not show anything that jumps out besides “Chain already exists”, which, from what I’ve read, is to be expected.
The log for istio-proxy has a bunch of this line:
[Envoy (Epoch 0)] [2020-02-18 23:54:16.184][20][warning][config] [bazel-out/k8-opt/bin/external/envoy/source/common/config/_virtual_includes/grpc_stream_lib/common/config/grpc_stream.h:91] gRPC config stream closed: 13,
The log for helloworld does not have anything besides those GET ones I pasted above.
When I describe the pod, I see this for istio-init:
State: Waiting Reason: CrashLoopBackOff Last State: Terminated Reason: Error Exit Code: 1 Started: Tue, 18 Feb 2020 15:56:01 -0800 Finished: Tue, 18 Feb 2020 15:56:01 -0800 Ready: False Restart Count: 46
helloworld and istio-proxy both look fine.
helloworld:
State: Running Started: Thu, 06 Feb 2020 12:33:31 -0800 Ready: True
istio-proxy:
State: Running Started: Thu, 06 Feb 2020 12:33:31 -0800 Ready: True
Notice how helloworld and istio-proxy’s start dates are 2/6, while istio-init’s is 2/18, which leads me to believe that istio-init is really the only one that keeps dying.
Any idea what’s going on here? I don’t think I’ve seen this before, when Citadel was still using self-signed certs, so it’s gotta be the external CA key/cert that’s causing the issue somehow.
Thanks for any inputs.