- Sidecar Injection Problems
- The result of sidecar injection was not what I expected
- Pods cannot be created at all
- x509 certificate related errors
- no such hosts or no endpoints available errors in deployment status
- Automatic sidecar injection fails if the Kubernetes API server has proxy settings
- Limitations for using Tcpdump in pods
- Cluster is not scaled down automatically
Sidecar Injection Problems
The result of sidecar injection was not what I expected
This includes an injected sidecar when it wasn’t expected and a lackof injected sidecar when it was.
Ensure your pod is not in the
kube-system
orkube-public
namespace.Automatic sidecar injection will be ignored for pods in these namespaces.Ensure your pod does not have
hostNetwork: true
in its pod spec.Automatic sidecar injection will be ignored for pods that are on the host network.
The sidecar model assumes that the iptables changes required for Envoy to intercepttraffic are within the pod. For pods on the host network this assumption is violated,and this can lead to routing failures at the host level.
- Check the webhook’s
namespaceSelector
to determine whether thewebhook is scoped to opt-in or opt-out for the target namespace.
The namespaceSelector
for opt-in will look like the following:
$ kubectl get mutatingwebhookconfiguration istio-sidecar-injector -o yaml | grep "namespaceSelector:" -A5
namespaceSelector:
matchLabels:
istio-injection: enabled
rules:
- apiGroups:
- ""
The injection webhook will be invoked for pods createdin namespaces with the istio-injection=enabled
label.
$ kubectl get namespace -L istio-injection
NAME STATUS AGE ISTIO-INJECTION
default Active 18d enabled
istio-system Active 3d
kube-public Active 18d
kube-system Active 18d
The namespaceSelector
for opt-out will look like the following:
$ kubectl get mutatingwebhookconfiguration istio-sidecar-injector -o yaml | grep "namespaceSelector:" -A5
namespaceSelector:
matchExpressions:
- key: istio-injection
operator: NotIn
values:
- disabled
rules:
- apiGroups:
- ""
The injection webhook will be invoked for pods created in namespaceswithout the istio-injection=disabled
label.
$ kubectl get namespace -L istio-injection
NAME STATUS AGE ISTIO-INJECTION
default Active 18d
istio-system Active 3d disabled
kube-public Active 18d disabled
kube-system Active 18d disabled
Verify the application pod’s namespace is labeled properly and (re) label accordingly, e.g.
$ kubectl label namespace istio-system istio-injection=disabled --overwrite
(repeat for all namespaces in which the injection webhook should be invoked for new pods)
$ kubectl label namespace default istio-injection=enabled --overwrite
- Check default policy
Check the default injection policy in the istio-sidecar-injector
configmap
.
$ kubectl -n istio-system get configmap istio-sidecar-injector -o jsonpath='{.data.config}' | grep policy:
policy: enabled
Allowed policy values are disabled
and enabled
. The default policyonly applies if the webhook’s namespaceSelector
matches the targetnamespace. Unrecognized policy causes injection to be disabled completely.
- Check the per-pod override annotation
The default policy can be overridden with thesidecar.istio.io/inject
annotation in the pod template spec’s metadata.The deployment’s metadata is ignored. Annotation valueof true
forces the sidecar to be injected while a value offalse
forces the sidecar to not be injected.
The following annotation overrides whatever the default policy
wasto force the sidecar to be injected:
$ kubectl get deployment sleep -o yaml | grep "sidecar.istio.io/inject:" -C3
template:
metadata:
annotations:
sidecar.istio.io/inject: "true"
labels:
app: sleep
Pods cannot be created at all
Run kubectl describe -n namespace deployment name
on the failingpod’s deployment. Failure to invoke the injection webhook willtypically be captured in the event log.
x509 certificate related errors
Warning FailedCreate 3m (x17 over 8m) replicaset-controller Error creating: Internal error occurred: \
failed calling admission webhook "sidecar-injector.istio.io": Post https://istio-sidecar-injector.istio-system.svc:443/inject: \
x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying \
to verify candidate authority certificate "Kubernetes.cluster.local")
x509: certificate signed by unknown authority
errors are typicallycaused by an empty caBundle
in the webhook configuration.
Verify the caBundle
in the mutatingwebhookconfiguration
matches theroot certificate mounted in the istio-sidecar-injector
pod.
$ kubectl get mutatingwebhookconfiguration istio-sidecar-injector -o yaml -o jsonpath='{.webhooks[0].clientConfig.caBundle}' | md5sum
4b95d2ba22ce8971c7c92084da31faf0 -
$ kubectl -n istio-system get secret istio.istio-sidecar-injector-service-account -o jsonpath='{.data.root-cert\.pem}' | md5sum
4b95d2ba22ce8971c7c92084da31faf0 -
The CA certificate should match. If they do not, restart thesidecar-injector pods.
$ kubectl -n istio-system patch deployment istio-sidecar-injector \
-p "{\"spec\":{\"template\":{\"metadata\":{\"labels\":{\"date\":\"`date +'%s'`\"}}}}}"
deployment.extensions "istio-sidecar-injector" patched
no such hosts or no endpoints available errors in deployment status
Injection is fail-close. If the istio-sidecar-injector
pod is not ready, podscannot be created. In such cases you’ll see an error about no endpoints available
.
Internal error occurred: failed calling admission webhook "istio-sidecar-injector.istio.io": \
Post https://istio-sidecar-injector.istio-system.svc:443/admitPilot?timeout=30s: \
no endpoints available for service "istio-sidecar-injector"
$ kubectl -n istio-system get pod -listio=sidecar-injector
NAME READY STATUS RESTARTS AGE
istio-sidecar-injector-5dbbbdb746-d676g 1/1 Running 0 2d
$ kubectl -n istio-system get endpoints istio-sidecar-injector
NAME ENDPOINTS AGE
istio-sidecar-injector 10.48.6.108:15014,10.48.6.108:443 3d
If the pods or endpoints aren’t ready, check the pod logs and statusfor any indication about why the webhook pod is failing to start andserve traffic.
$ for pod in $(kubectl -n istio-system get pod -listio=sidecar-injector -o jsonpath='{.items[*].metadata.name}'); do \
kubectl -n istio-system logs ${pod} \
done
$ for pod in $(kubectl -n istio-system get pod -listio=sidecar-injector -o name); do \
kubectl -n istio-system describe ${pod} \
done
Automatic sidecar injection fails if the Kubernetes API server has proxy settings
When the Kubernetes API server includes proxy settings such as:
env:
- name: http_proxy
value: http://proxy-wsa.esl.foo.com:80
- name: https_proxy
value: http://proxy-wsa.esl.foo.com:80
- name: no_proxy
value: 127.0.0.1,localhost,dockerhub.foo.com,devhub-docker.foo.com,10.84.100.125,10.84.100.126,10.84.100.127
With these settings, Sidecar injection fails. The only related failure log can be found in kube-apiserver
log:
W0227 21:51:03.156818 1 admission.go:257] Failed calling webhook, failing open sidecar-injector.istio.io: failed calling admission webhook "sidecar-injector.istio.io": Post https://istio-sidecar-injector.istio-system.svc:443/inject: Service Unavailable
Make sure both pod and service CIDRs are not proxied according to *_proxy
variables. Check the kube-apiserver
files and logs to verify the configuration and whether any requests are being proxied.
One workaround is to remove the proxy settings from the kube-apiserver
manifest, another workaround is to include istio-sidecar-injector.istio-system.svc
or .svc
in the no_proxy
value. Make sure that kube-apiserver
is restarted after each workaround.
An issue was filed with Kubernetes related to this and has since been closed.https://github.com/kubernetes/kubernetes/pull/58698#discussion_r163879443
Limitations for using Tcpdump in pods
Tcpdump doesn’t work in the sidecar pod - the container doesn’t run as root. However any other container in the same pod will see all the packets, since thenetwork namespace is shared. iptables
will also see the pod-wide configuration.
Communication between Envoy and the app happens on 127.0.0.1, and is not encrypted.
Cluster is not scaled down automatically
Due to the fact that the sidecar container mounts a local storage volume, thenode autoscaler is unable to evict nodes with the injected pods. This isa known issue. The workaround isto add a pod annotation "cluster-autoscaler.kubernetes.io/safe-to-evict":"true"
to the injected pods.