Profile applicability: Level 1
Allow Kubelet to manage iptables.
Kubelets can automatically manage the required changes to iptables based on how you
choose your networking options for the pods. It is recommended to let kubelets manage
the changes to iptables. This ensures that the iptables configuration remains in sync
with pods networking configuration. Manually configuring iptables with dynamic pod
network configuration changes might hamper the communication between pods/containers
and to the outside world. You might have iptables rules too restrictive or too open.
NoteSee the Azure AKS documentation for the default value.
|
Impact
Kubelet would manage the iptables on the system and keep it in sync. If you are using
any other iptables management solution, then there might be some conflicts.
Audit
Audit Method 1If using a Kubelet configuration file, check that there is an entry for
makeIPTablesUtilChains set to true.-
SSH to the relevant node.
-
Run the following command on each node to find the appropriate Kubelet config file:
ps -ef | grep kubelet
The output should return something similar to--config /etc/kubernetes/kubelet/kubelet-config.json, which is the location of the Kubelet config file. -
Open the Kubelet config file:
cat /etc/kubernetes/kubelet/kubelet-config.json
-
Verify that if the
makeIPTablesUtilChainsargument exists then it is set totrue.If the--make-iptables-util-chainsargument does not exist, and there is a Kubelet config file specified by--config, verify that the file does not setmakeIPTablesUtilChainstofalse.
If using the api configz endpoint, search for the status of
"makeIPTablesUtilChains":true by extracting the live configuration from the nodes running kubelet.Set the local proxy port and the following variables and provide proxy port number
and node name:
HOSTNAME_PORT="localhost-and-port-number" NODE_NAME="The- Name-Of-Node-To-Extract-Configuration"
from the output of "kubectl get nodes"kubectl proxy --port=8001 &
export HOSTNAME_PORT=localhost:8001
export NODE_NAME=ip-192.168.31.226.aks.internal
curl -sSL "http://${HOSTNAME_PORT}/api/v1/nodes/${NODE_NAME}/proxy/configz"
Remediation
Remediation Method 1If modifying the Kubelet config file, edit the
kubelet-config.json file /etc/kubernetes/kubelet/kubelet-config.json and set the below parameter:"makeIPTablesUtilChains": trueRemediation Method 2
If using executable arguments, edit the kubelet service file
/etc/systemd/system/kubelet.service.d/10-kubelet-args.conf on each worker node and add the below parameter at the end of the KUBELET_ARGS variable
string:--make-iptables-util-chains:trueRemediation Method 3
If using the api configz endpoint, search for the status of
"makeIPTablesUtilChains": true by extracting the live configuration from the nodes running kubelet. See detailed
step-by-step configmap procedures in Reconfigure a Node's Kubelet in a Live Cluster,
and then rerun the curl statement from audit process to check for kubelet configuration
changes: kubectl proxy --port=8001 &
export HOSTNAME_PORT=localhost:8001
export NODE_NAME=ip-192.168.31.226.aks.internal
curl -sSL "http://${HOSTNAME_PORT}/api/v1/nodes/${NODE_NAME}/proxy/configz"
For all three remediations: Based on your system, restart the kubelet service and check status:
systemctl daemon-reload systemctl restart kubelet.service systemctl status kubelet -l
