#!/bin/bash check_5() { logit "" local id="5" local desc="Container Runtime" checkHeader="$id - $desc" info "$checkHeader" startsectionjson "$id" "$desc" } check_running_containers() { # If containers is empty, there are no running containers if [ -z "$containers" ]; then info " * No containers running, skipping Section 5" return fi # Make the loop separator be a new-line in POSIX compliant fashion set -f; IFS=$' ' } check_5_1() { local id="5.1" local desc="Ensure swarm mode is not Enabled, if not needed (Automated)" local remediation="If swarm mode has been enabled on a system in error, you should run the command: docker swarm leave" local remediationImpact="Disabling swarm mode will impact the operation of Docker Enterprise components if these are in use." local check="$id - $desc" starttestjson "$id" "$desc" if docker info 2>/dev/null | grep -e "Swarm:*\sinactive\s*" >/dev/null 2>&1; then pass -s "$check" logcheckresult "PASS" return fi warn -s "$check" logcheckresult "WARN" } check_5_2() { if [ -z "$containers" ]; then return fi local id="5.2" local desc="Ensure that, if applicable, an AppArmor Profile is enabled (Automated)" local remediation="If AppArmor is applicable for your Linux OS, you should enable it. Alternatively, Docker's default AppArmor policy can be used." local remediationImpact="The container will have the security controls defined in the AppArmor profile. It should be noted that if the AppArmor profile is misconfigured, this may cause issues with the operation of the container." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 no_apparmor_containers="" for c in $containers; do policy=$(docker inspect --format 'AppArmorProfile={{ .AppArmorProfile }}' "$c") if [ "$policy" = "AppArmorProfile=" ] || [ "$policy" = "AppArmorProfile=[]" ] || [ "$policy" = "AppArmorProfile=" ] || [ "$policy" = "AppArmorProfile=unconfined" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * No AppArmorProfile Found: $c" no_apparmor_containers="$no_apparmor_containers $c" fail=1 continue fi warn " * No AppArmorProfile Found: $c" no_apparmor_containers="$no_apparmor_containers $c" fi done # We went through all the containers and found none without AppArmor if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers with no AppArmorProfile" "$no_apparmor_containers" } check_5_3() { if [ -z "$containers" ]; then return fi local id="5.3" local desc="Ensure that, if applicable, SELinux security options are set (Automated)" local remediation="Set the SELinux State. Set the SELinux Policy. Create or import a SELinux policy template for Docker containers. Start Docker in daemon mode with SELinux enabled. Start your Docker container using the security options." local remediationImpact="Any restrictions defined in the SELinux policy will be applied to your containers. It should be noted that if your SELinux policy is misconfigured, this may have an impact on the correct operation of the affected containers." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 no_securityoptions_containers="" for c in $containers; do policy=$(docker inspect --format 'SecurityOpt={{ .HostConfig.SecurityOpt }}' "$c") if [ "$policy" = "SecurityOpt=" ] || [ "$policy" = "SecurityOpt=[]" ] || [ "$policy" = "SecurityOpt=" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * No SecurityOptions Found: $c" no_securityoptions_containers="$no_securityoptions_containers $c" fail=1 continue fi warn " * No SecurityOptions Found: $c" no_securityoptions_containers="$no_securityoptions_containers $c" fi done # We went through all the containers and found none without SELinux if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers with no SecurityOptions" "$no_securityoptions_containers" } check_5_4() { if [ -z "$containers" ]; then return fi local id="5.4" local desc="Ensure that Linux kernel capabilities are restricted within containers (Automated)" local remediation="You could remove all the currently configured capabilities and then restore only the ones you specifically use: docker run --cap-drop=all --cap-add={,} " local remediationImpact="Restrictions on processes within a container are based on which Linux capabilities are in force. Removal of the NET_RAW capability prevents the container from creating raw sockets which is good security practice under most circumstances, but may affect some networking utilities." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 caps_containers="" for c in $containers; do container_caps=$(docker inspect --format 'CapAdd={{ .HostConfig.CapAdd }}' "$c") caps=$(echo "$container_caps" | tr "[:lower:]" "[:upper:]" | \ sed 's/CAPADD/CapAdd/' | \ sed -r "s/CAP_AUDIT_WRITE|CAP_CHOWN|CAP_DAC_OVERRIDE|CAP_FOWNER|CAP_FSETID|CAP_KILL|CAP_MKNOD|CAP_NET_BIND_SERVICE|CAP_NET_RAW|CAP_SETFCAP|CAP_SETGID|CAP_SETPCAP|CAP_SETUID|CAP_SYS_CHROOT|\s//g" | \ sed -r "s/AUDIT_WRITE|CHOWN|DAC_OVERRIDE|FOWNER|FSETID|KILL|MKNOD|NET_BIND_SERVICE|NET_RAW|SETFCAP|SETGID|SETPCAP|SETUID|SYS_CHROOT|\s//g") if [ "$caps" != 'CapAdd=' ] && [ "$caps" != 'CapAdd=[]' ] && [ "$caps" != 'CapAdd=' ] && [ "$caps" != 'CapAdd=' ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Capabilities added: $caps to $c" caps_containers="$caps_containers $c" fail=1 continue fi warn " * Capabilities added: $caps to $c" caps_containers="$caps_containers $c" fi done # We went through all the containers and found none with extra capabilities if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Capabilities added for containers" "$caps_containers" } check_5_5() { if [ -z "$containers" ]; then return fi local id="5.5" local desc="Ensure that privileged containers are not used (Automated)" local remediation="You should not run containers with the --privileged flag." local remediationImpact="If you start a container without the --privileged flag, it will not have excessive default capabilities." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 privileged_containers="" for c in $containers; do privileged=$(docker inspect --format '{{ .HostConfig.Privileged }}' "$c") if [ "$privileged" = "true" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Container running in Privileged mode: $c" privileged_containers="$privileged_containers $c" fail=1 continue fi warn " * Container running in Privileged mode: $c" privileged_containers="$privileged_containers $c" fi done # We went through all the containers and found no privileged containers if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers running in privileged mode" "$privileged_containers" } check_5_6() { if [ -z "$containers" ]; then return fi local id="5.6" local desc="Ensure sensitive host system directories are not mounted on containers (Automated)" local remediation="You should not mount directories which are security sensitive on the host within containers, especially in read-write mode." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" # List of sensitive directories to test for. Script uses new-lines as a separator. # Note the lack of identation. It needs it for the substring comparison. sensitive_dirs='/ /boot /dev /etc /lib /proc /sys /usr' fail=0 sensitive_mount_containers="" for c in $containers; do volumes=$(docker inspect --format '{{ .Mounts }}' "$c") if docker inspect --format '{{ .VolumesRW }}' "$c" 2>/dev/null 1>&2; then volumes=$(docker inspect --format '{{ .VolumesRW }}' "$c") fi # Go over each directory in sensitive dir and see if they exist in the volumes for v in $sensitive_dirs; do sensitive=0 if echo "$volumes" | grep -e "{.*\s$v\s.*true\s.*}" 2>/tmp/null 1>&2; then sensitive=1 fi if [ $sensitive -eq 1 ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Sensitive directory $v mounted in: $c" sensitive_mount_containers="$sensitive_mount_containers $c:$v" fail=1 continue fi warn " * Sensitive directory $v mounted in: $c" sensitive_mount_containers="$sensitive_mount_containers $c:$v" fi done done # We went through all the containers and found none with sensitive mounts if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers with sensitive directories mounted" "$sensitive_mount_containers" } check_5_7() { if [ -z "$containers" ]; then return fi local id="5.7" local desc="Ensure sshd is not run within containers (Automated)" local remediation="Uninstall the SSH daemon from the container and use docker exec to enter a container on the remote host." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 ssh_exec_containers="" printcheck=0 for c in $containers; do processes=$(docker exec "$c" ps -el 2>/dev/null | grep -c sshd | awk '{print $1}') if [ "$processes" -ge 1 ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Container running sshd: $c" ssh_exec_containers="$ssh_exec_containers $c" fail=1 printcheck=1 else warn " * Container running sshd: $c" ssh_exec_containers="$ssh_exec_containers $c" fi fi exec_check=$(docker exec "$c" ps -el 2>/dev/null) if [ $? -eq 265 ]; then if [ $printcheck -eq 0 ]; then warn -s "$check" printcheck=1 fi warn " * Docker exec fails: $c" ssh_exec_containers="$ssh_exec_containers $c" fail=1 fi done # We went through all the containers and found none with sshd if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers with sshd/docker exec failures" "$ssh_exec_containers" } check_5_8() { if [ -z "$containers" ]; then return fi local id="5.8" local desc="Ensure privileged ports are not mapped within containers (Automated)" local remediation="You should not map container ports to privileged host ports when starting a container. You should also, ensure that there is no such container to host privileged port mapping declarations in the Dockerfile." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 privileged_port_containers="" for c in $containers; do # Port format is private port -> ip: public port ports=$(docker port "$c" | awk '{print $0}' | cut -d ':' -f2) # iterate through port range (line delimited) for port in $ports; do if [ -n "$port" ] && [ "$port" -lt 1025 ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Privileged Port in use: $port in $c" privileged_port_containers="$privileged_port_containers $c:$port" fail=1 continue fi warn " * Privileged Port in use: $port in $c" privileged_port_containers="$privileged_port_containers $c:$port" fi done done # We went through all the containers and found no privileged ports if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers using privileged ports" "$privileged_port_containers" } check_5_9() { if [ -z "$containers" ]; then return fi local id="5.9" local desc="Ensure that only needed ports are open on the container (Manual)" local remediation="You should ensure that the Dockerfile for each container image only exposes needed ports." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 open_port_containers="" for c in $containers; do ports=$(docker port "$c" | awk '{print $0}' | cut -d ':' -f2) for port in $ports; do if [ -n "$port" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Port in use: $port in $c" open_port_containers="$open_port_containers $c:$port" fail=1 continue fi warn " * Port in use: $port in $c" open_port_containers="$open_port_containers $c:$port" fi done done # We went through all the containers and found none with open ports if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers with open ports" "$open_port_containers" } check_5_10() { if [ -z "$containers" ]; then return fi local id="5.10" local desc="Ensure that the host's network namespace is not shared (Automated)" local remediation="You should not pass the --net=host option when starting any container." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 net_host_containers="" for c in $containers; do mode=$(docker inspect --format 'NetworkMode={{ .HostConfig.NetworkMode }}' "$c") if [ "$mode" = "NetworkMode=host" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Container running with networking mode 'host': $c" net_host_containers="$net_host_containers $c" fail=1 continue fi warn " * Container running with networking mode 'host': $c" net_host_containers="$net_host_containers $c" fi done # We went through all the containers and found no Network Mode host if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers running with networking mode 'host'" "$net_host_containers" } check_5_11() { if [ -z "$containers" ]; then return fi local id="5.11" local desc="Ensure that the memory usage for containers is limited (Automated)" local remediation="You should run the container with only as much memory as it requires by using the --memory argument." local remediationImpact="If correct memory limits are not set on each container, one process can expand its usage and cause other containers to run out of resources." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 mem_unlimited_containers="" for c in $containers; do memory=$(docker inspect --format '{{ .HostConfig.Memory }}' "$c") if docker inspect --format '{{ .Config.Memory }}' "$c" 2> /dev/null 1>&2; then memory=$(docker inspect --format '{{ .Config.Memory }}' "$c") fi if [ "$memory" = "0" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Container running without memory restrictions: $c" mem_unlimited_containers="$mem_unlimited_containers $c" fail=1 continue fi warn " * Container running without memory restrictions: $c" mem_unlimited_containers="$mem_unlimited_containers $c" fi done # We went through all the containers and found no lack of Memory restrictions if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Container running without memory restrictions" "$mem_unlimited_containers" } check_5_12() { if [ -z "$containers" ]; then return fi local id="5.12" local desc="Ensure that CPU priority is set appropriately on containers (Automated)" local remediation="You should manage the CPU runtime between your containers dependent on their priority within your organization. To do so start the container using the --cpu-shares argument." local remediationImpact="If you do not correctly assign CPU thresholds, the container process may run out of resources and become unresponsive. If CPU resources on the host are not constrainted, CPU shares do not place any restrictions on individual resources." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 cpu_unlimited_containers="" for c in $containers; do cpushares=$(docker inspect --format '{{ .HostConfig.CpuShares }}' "$c") nanocpus=$(docker inspect --format '{{ .HostConfig.NanoCpus }}' "$c") if docker inspect --format '{{ .Config.CpuShares }}' "$c" 2> /dev/null 1>&2; then cpushares=$(docker inspect --format '{{ .Config.CpuShares }}' "$c") nanocpus=$(docker inspect --format '{{ .Config.NanoCpus }}' "$c") fi if [ "$cpushares" = "0" ] && [ "$nanocpus" = "0" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Container running without CPU restrictions: $c" cpu_unlimited_containers="$cpu_unlimited_containers $c" fail=1 continue fi warn " * Container running without CPU restrictions: $c" cpu_unlimited_containers="$cpu_unlimited_containers $c" fi done # We went through all the containers and found no lack of CPUShare restrictions if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers running without CPU restrictions" "$cpu_unlimited_containers" } check_5_13() { if [ -z "$containers" ]; then return fi local id="5.13" local desc="Ensure that the container's root filesystem is mounted as read only (Automated)" local remediation="You should add a --read-only flag at a container's runtime to enforce the container's root filesystem being mounted as read only." local remediationImpact="Enabling --read-only at container runtime may break some container OS packages if a data writing strategy is not defined. You should define what the container's data should and should not persist at runtime in order to decide which strategy to use." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 fsroot_mount_containers="" for c in $containers; do read_status=$(docker inspect --format '{{ .HostConfig.ReadonlyRootfs }}' "$c") if [ "$read_status" = "false" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Container running with root FS mounted R/W: $c" fsroot_mount_containers="$fsroot_mount_containers $c" fail=1 continue fi warn " * Container running with root FS mounted R/W: $c" fsroot_mount_containers="$fsroot_mount_containers $c" fi done # We went through all the containers and found no R/W FS mounts if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers running with root FS mounted R/W" "$fsroot_mount_containers" } check_5_14() { if [ -z "$containers" ]; then return fi local id="5.14" local desc="Ensure that incoming container traffic is bound to a specific host interface (Automated)" local remediation="You should bind the container port to a specific host interface on the desired host port. Example: docker run --detach --publish 10.2.3.4:49153:80 nginx In this example, the container port 80 is bound to the host port on 49153 and would accept incoming connection only from the 10.2.3.4 external interface." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 incoming_unbound_containers="" for c in $containers; do for ip in $(docker port "$c" | awk '{print $3}' | cut -d ':' -f1); do if [ "$ip" = "0.0.0.0" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Port being bound to wildcard IP: $ip in $c" incoming_unbound_containers="$incoming_unbound_containers $c:$ip" fail=1 continue fi warn " * Port being bound to wildcard IP: $ip in $c" incoming_unbound_containers="$incoming_unbound_containers $c:$ip" fi done done # We went through all the containers and found no ports bound to 0.0.0.0 if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers with port bound to wildcard IP" "$incoming_unbound_containers" } check_5_15() { if [ -z "$containers" ]; then return fi local id="5.15" local desc="Ensure that the 'on-failure' container restart policy is set to '5' (Automated)" local remediation="If you wish a container to be automatically restarted, a sample command is docker run --detach --restart=on-failure:5 nginx" local remediationImpact="If this option is set, a container will only attempt to restart itself 5 times." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 maxretry_unset_containers="" for c in $containers; do container_name=$(docker inspect "$c" --format '{{.Name}}') if [ "$(docker info --format '{{.Swarm.LocalNodeState}}')" = "active" ]; then for s in $(docker service ls --format '{{.Name}}'); do if echo $container_name | grep -q "$s"; then task_id=$(docker inspect "$c" --format '{{.Name}}' | awk -F '.' '{print $NF}') # a container name could arbitrary include a service one: it belongs to a service (created by Docker # as part of the service), if the container task ID matches one of the task IDs of the service. if docker service ps --no-trunc "$s" --format '{{.ID}}' | grep -q "$task_id"; then restart_policy=$(docker inspect --format '{{ .Spec.TaskTemplate.RestartPolicy.MaxAttempts }}' "$s") break fi fi done fi if docker inspect --format '{{ .HostConfig.RestartPolicy.MaximumRetryCount }}' "$c" &>/dev/null; then restart_policy=$(docker inspect --format '{{ .HostConfig.RestartPolicy.MaximumRetryCount }}' "$c") fi if [ "$restart_policy" -gt "5" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * MaximumRetryCount is not set to 5 or less: $c" maxretry_unset_containers="$maxretry_unset_containers $c" fail=1 continue fi warn " * MaximumRetryCount is not set to 5 or less: $c" maxretry_unset_containers="$maxretry_unset_containers $c" fi done # We went through all the containers and they all had MaximumRetryCount=5 if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers with MaximumRetryCount not set to 5 or less" "$maxretry_unset_containers" } check_5_16() { if [ -z "$containers" ]; then return fi local id="5.16" local desc="Ensure that the host's process namespace is not shared (Automated)" local remediation="You should not start a container with the --pid=host argument." local remediationImpact="Container processes cannot see processes on the host system." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 pidns_shared_containers="" for c in $containers; do mode=$(docker inspect --format 'PidMode={{.HostConfig.PidMode }}' "$c") if [ "$mode" = "PidMode=host" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Host PID namespace being shared with: $c" pidns_shared_containers="$pidns_shared_containers $c" fail=1 continue fi warn " * Host PID namespace being shared with: $c" pidns_shared_containers="$pidns_shared_containers $c" fi done # We went through all the containers and found none with PidMode as host if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers sharing host PID namespace" "$pidns_shared_containers" } check_5_17() { if [ -z "$containers" ]; then return fi local id="5.17" local desc="Ensure that the host's IPC namespace is not shared (Automated)" local remediation="You should not start a container with the --ipc=host argument." local remediationImpact="Shared memory segments are used in order to accelerate interprocess communications, commonly in high-performance applications. If this type of application is containerized into multiple containers, you might need to share the IPC namespace of the containers in order to achieve high performance. Under these circumstances, you should still only share container specific IPC namespaces and not the host IPC namespace." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 ipcns_shared_containers="" for c in $containers; do mode=$(docker inspect --format 'IpcMode={{.HostConfig.IpcMode }}' "$c") if [ "$mode" = "IpcMode=host" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Host IPC namespace being shared with: $c" ipcns_shared_containers="$ipcns_shared_containers $c" fail=1 continue fi warn " * Host IPC namespace being shared with: $c" ipcns_shared_containers="$ipcns_shared_containers $c" fi done # We went through all the containers and found none with IPCMode as host if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers sharing host IPC namespace" "$ipcns_shared_containers" } check_5_18() { if [ -z "$containers" ]; then return fi local id="5.18" local desc="Ensure that host devices are not directly exposed to containers (Manual)" local remediation="You should not directly expose host devices to containers. If you do need to expose host devices to containers, you should use granular permissions as appropriate to your organization." local remediationImpact="You would not be able to use host devices directly within containers." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 hostdev_exposed_containers="" for c in $containers; do devices=$(docker inspect --format 'Devices={{ .HostConfig.Devices }}' "$c") if [ "$devices" != "Devices=" ] && [ "$devices" != "Devices=[]" ] && [ "$devices" != "Devices=" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then info -c "$check" info " * Container has devices exposed directly: $c" hostdev_exposed_containers="$hostdev_exposed_containers $c" fail=1 continue fi info " * Container has devices exposed directly: $c" hostdev_exposed_containers="$hostdev_exposed_containers $c" fi done # We went through all the containers and found none with devices if [ $fail -eq 0 ]; then pass -c "$check" logcheckresult "PASS" return fi logcheckresult "INFO" "Containers with host devices exposed directly" "$hostdev_exposed_containers" } check_5_19() { if [ -z "$containers" ]; then return fi local id="5.19" local desc="Ensure that the default ulimit is overwritten at runtime if needed (Manual)" local remediation="You should only override the default ulimit settings if needed in a specific case." local remediationImpact="If ulimits are not set correctly, overutilization by individual containers could make the host system unusable." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 no_ulimit_containers="" for c in $containers; do ulimits=$(docker inspect --format 'Ulimits={{ .HostConfig.Ulimits }}' "$c") if [ "$ulimits" = "Ulimits=" ] || [ "$ulimits" = "Ulimits=[]" ] || [ "$ulimits" = "Ulimits=" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then info -c "$check" info " * Container no default ulimit override: $c" no_ulimit_containers="$no_ulimit_containers $c" fail=1 continue fi info " * Container no default ulimit override: $c" no_ulimit_containers="$no_ulimit_containers $c" fi done # We went through all the containers and found none without Ulimits if [ $fail -eq 0 ]; then pass -c "$check" logcheckresult "PASS" return fi logcheckresult "INFO" "Containers with no default ulimit override" "$no_ulimit_containers" } check_5_20() { if [ -z "$containers" ]; then return fi local id="5.20" local desc="Ensure mount propagation mode is not set to shared (Automated)" local remediation="Do not mount volumes in shared mode propagation." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 mountprop_shared_containers="" for c in $containers; do if docker inspect --format 'Propagation={{range $mnt := .Mounts}} {{json $mnt.Propagation}} {{end}}' "$c" | \ grep shared 2>/dev/null 1>&2; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Mount propagation mode is shared: $c" mountprop_shared_containers="$mountprop_shared_containers $c" fail=1 continue fi warn " * Mount propagation mode is shared: $c" mountprop_shared_containers="$mountprop_shared_containers $c" fi done # We went through all the containers and found none with shared propagation mode if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers with shared mount propagation" "$mountprop_shared_containers" } check_5_21() { if [ -z "$containers" ]; then return fi local id="5.21" local desc="Ensure that the host's UTS namespace is not shared (Automated)" local remediation="You should not start a container with the --uts=host argument." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 utcns_shared_containers="" for c in $containers; do mode=$(docker inspect --format 'UTSMode={{.HostConfig.UTSMode }}' "$c") if [ "$mode" = "UTSMode=host" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Host UTS namespace being shared with: $c" utcns_shared_containers="$utcns_shared_containers $c" fail=1 continue fi warn " * Host UTS namespace being shared with: $c" utcns_shared_containers="$utcns_shared_containers $c" fi done # We went through all the containers and found none with UTSMode as host if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers sharing host UTS namespace" "$utcns_shared_containers" } check_5_22() { if [ -z "$containers" ]; then return fi local id="5.22" local desc="Ensure the default seccomp profile is not Disabled (Automated)" local remediation="By default, seccomp profiles are enabled. You do not need to do anything unless you want to modify and use a modified seccomp profile." local remediationImpact="With Docker 1.10 and greater, the default seccomp profile blocks syscalls, regardless of -- cap-add passed to the container." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 seccomp_disabled_containers="" for c in $containers; do if docker inspect --format 'SecurityOpt={{.HostConfig.SecurityOpt }}' "$c" | \ grep -E 'seccomp:unconfined|seccomp=unconfined' 2>/dev/null 1>&2; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Default seccomp profile disabled: $c" seccomp_disabled_containers="$seccomp_disabled_containers $c" fail=1 else warn " * Default seccomp profile disabled: $c" seccomp_disabled_containers="$seccomp_disabled_containers $c" fi fi done # We went through all the containers and found none with default secomp profile disabled if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers with default seccomp profile disabled" "$seccomp_disabled_containers" } check_5_23() { if [ -z "$containers" ]; then return fi local id="5.23" local desc="Ensure that docker exec commands are not used with the privileged option (Automated)" local remediation="You should not use the --privileged option in docker exec commands." local remediationImpact="If you need enhanced capabilities within a container, then run it with all the permissions it requires. These should be specified individually." local check="$id - $desc" starttestjson "$id" "$desc" note -c "$check" logcheckresult "NOTE" } check_5_24() { if [ -z "$containers" ]; then return fi local id="5.24" local desc="Ensure that docker exec commands are not used with the user=root option (Manual)" local remediation="You should not use the --user=root option in docker exec commands." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" note -c "$check" logcheckresult "NOTE" } check_5_25() { if [ -z "$containers" ]; then return fi local id="5.25" local desc="Ensure that cgroup usage is confirmed (Automated)" local remediation="You should not use the --cgroup-parent option within the docker run command unless strictly required." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 unexpected_cgroup_containers="" for c in $containers; do mode=$(docker inspect --format 'CgroupParent={{.HostConfig.CgroupParent }}x' "$c") if [ "$mode" != "CgroupParent=x" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Confirm cgroup usage: $c" unexpected_cgroup_containers="$unexpected_cgroup_containers $c" fail=1 continue fi warn " * Confirm cgroup usage: $c" unexpected_cgroup_containers="$unexpected_cgroup_containers $c" fi done # We went through all the containers and found none with UTSMode as host if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers using unexpected cgroup" "$unexpected_cgroup_containers" } check_5_26() { if [ -z "$containers" ]; then return fi local id="5.26" local desc="Ensure that the container is restricted from acquiring additional privileges (Automated)" local remediation="You should start your container with the options: docker run --rm -it --security-opt=no-new-privileges ubuntu bash" local remediationImpact="The no_new_priv option prevents LSMs like SELinux from allowing processes to acquire new privileges." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 no_priv_config=0 addprivs_containers="" if get_docker_effective_command_line_args '--no-new-privileges' | grep "no-new-privileges" >/dev/null 2>&1; then no_priv_config=1 elif get_docker_configuration_file_args 'no-new-privileges' | grep true >/dev/null 2>&1; then no_priv_config=1 else for c in $containers; do if ! docker inspect --format 'SecurityOpt={{.HostConfig.SecurityOpt }}' "$c" | grep 'no-new-privileges' 2>/dev/null 1>&2; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Privileges not restricted: $c" addprivs_containers="$addprivs_containers $c" fail=1 continue fi warn " * Privileges not restricted: $c" addprivs_containers="$addprivs_containers $c" fi done fi # We went through all the containers and found none with capability to acquire additional privileges if [ $fail -eq 0 ] || [ $no_priv_config -eq 1 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers without restricted privileges" "$addprivs_containers" } check_5_27() { if [ -z "$containers" ]; then return fi local id="5.27" local desc="Ensure that container health is checked at runtime (Automated)" local remediation="You should run the container using the --health-cmd parameter." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 nohealthcheck_containers="" for c in $containers; do if ! docker inspect --format '{{ .Id }}: Health={{ .State.Health.Status }}' "$c" 2>/dev/null 1>&2; then if [ $fail -eq 0 ]; then warn -s "$check" warn " * Health check not set: $c" nohealthcheck_containers="$nohealthcheck_containers $c" fail=1 continue fi warn " * Health check not set: $c" nohealthcheck_containers="$nohealthcheck_containers $c" fi done if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers without health check" "$nohealthcheck_containers" } check_5_28() { if [ -z "$containers" ]; then return fi local id="5.28" local desc="Ensure that Docker commands always make use of the latest version of their image (Manual)" local remediation="You should use proper version pinning mechanisms (the tag which is assigned by default is still vulnerable to caching attacks) to avoid extracting cached older versions. Version pinning mechanisms should be used for base images, packages, and entire images. You can customize version pinning rules according to your requirements." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" info -c "$check" logcheckresult "INFO" } check_5_29() { if [ -z "$containers" ]; then return fi local id="5.29" local desc="Ensure that the PIDs cgroup limit is used (Automated)" local remediation="Use --pids-limit flag with an appropriate value when launching the container." local remediationImpact="Set the PIDs limit value as appropriate. Incorrect values might leave containers unusable." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 nopids_limit_containers="" for c in $containers; do pidslimit="$(docker inspect --format '{{.HostConfig.PidsLimit }}' "$c")" if [ "$pidslimit" = "0" ] || [ "$pidslimit" = "" ] || [ "$pidslimit" = "-1" ]; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * PIDs limit not set: $c" nopids_limit_containers="$nopids_limit_containers $c" fail=1 continue fi warn " * PIDs limit not set: $c" nopids_limit_containers="$nopids_limit_containers $c" fi done # We went through all the containers and found all with PIDs limit if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers without PIDs cgroup limit" "$nopids_limit_containers" } check_5_30() { if [ -z "$containers" ]; then return fi local id="5.30" local desc="Ensure that Docker's default bridge 'docker0' is not used (Manual)" local remediation="You should follow the Docker documentation and set up a user-defined network. All the containers should be run in this network." local remediationImpact="User-defined networks need to be configured and managed in line with organizational security policy." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 docker_network_containers="" networks=$(docker network ls -q 2>/dev/null) for net in $networks; do if docker network inspect --format '{{ .Options }}' "$net" 2>/dev/null | grep "com.docker.network.bridge.name:docker0" >/dev/null 2>&1; then docker0Containers=$(docker network inspect --format='{{ range $k, $v := .Containers }} {{ $k }} {{ end }}' "$net" | \ sed -e 's/^ //' -e 's/ /\n/g' 2>/dev/null) if [ -n "$docker0Containers" ]; then if [ $fail -eq 0 ]; then info -c "$check" fail=1 fi for c in $docker0Containers; do if [ -z "$exclude" ]; then cName=$(docker inspect --format '{{.Name}}' "$c" 2>/dev/null | sed 's/\///g') else pattern=$(echo "$exclude" | sed 's/,/|/g') cName=$(docker inspect --format '{{.Name}}' "$c" 2>/dev/null | sed 's/\///g' | grep -Ev "$pattern" ) fi if [ -n "$cName" ]; then info " * Container in docker0 network: $cName" docker_network_containers="$docker_network_containers $c:$cName" fi done fi fi done # We went through all the containers and found none in docker0 network if [ $fail -eq 0 ]; then pass -c "$check" logcheckresult "PASS" return fi logcheckresult "INFO" "Containers using docker0 network" "$docker_network_containers" } check_5_31() { if [ -z "$containers" ]; then return fi local id="5.31" local desc="Ensure that the host's user namespaces are not shared (Automated)" local remediation="You should not share user namespaces between host and containers." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 hostns_shared_containers="" for c in $containers; do if docker inspect --format '{{ .HostConfig.UsernsMode }}' "$c" 2>/dev/null | grep -i 'host' >/dev/null 2>&1; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Namespace shared: $c" hostns_shared_containers="$hostns_shared_containers $c" fail=1 continue fi warn " * Namespace shared: $c" hostns_shared_containers="$hostns_shared_containers $c" fi done # We went through all the containers and found none with host's user namespace shared if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers sharing host user namespace" "$hostns_shared_containers" } check_5_32() { if [ -z "$containers" ]; then return fi local id="5.32" local desc="Ensure that the Docker socket is not mounted inside any containers (Automated)" local remediation="You should ensure that no containers mount docker.sock as a volume." local remediationImpact="None." local check="$id - $desc" starttestjson "$id" "$desc" fail=0 docker_sock_containers="" for c in $containers; do if docker inspect --format '{{ .Mounts }}' "$c" 2>/dev/null | grep 'docker.sock' >/dev/null 2>&1; then # If it's the first container, fail the test if [ $fail -eq 0 ]; then warn -s "$check" warn " * Docker socket shared: $c" docker_sock_containers="$docker_sock_containers $c" fail=1 continue fi warn " * Docker socket shared: $c" docker_sock_containers="$docker_sock_containers $c" fi done # We went through all the containers and found none with docker.sock shared if [ $fail -eq 0 ]; then pass -s "$check" logcheckresult "PASS" return fi logcheckresult "WARN" "Containers sharing docker socket" "$docker_sock_containers" } check_5_end() { endsectionjson }