After making the upgrade from 1.2.* to 1.3* (I’ve gone from 1.3.0 to 1.3.3 with the same results). I’m seeing exponential CPU utilization in the istio-ingressgateway pods. While in 1.2.5 the pod took less than 50m CPU and immediately after upgrading to 1.3.0 and 1.3.3 the CPU spikes to 500-800m CPU and scales out to the maximum value of instances that the autoscaler allows.
I’ve tested this against low and high traffic clusters (We’re running about 10 small ones) and the result is the same from one with around 300RPS to 0 RPS.
When exec-ing into the pod itself and running a top/htop, the envoy process spikes to +100% about every 30 seconds. Is there information on what went into the last versions of proxyv2 that would cause this? I can make adjustments to the clusters to accommodate for the time being but I’m a little concerned if I see spikes in traffic and the system can’t handle it.