为什么Docker容器无法相互通信?

sta*_*wer 5 linux ip routing containers docker

我创建了一个小项目来测试Docker集群.基本上,cluster.sh脚本启动三个相同的容器,并使用管道bridge1在主机上配置bridge (eth1)并向每个容器添加NIC().

如果我登录其中一个容器,我可以使用arping其他容器:

# 172.17.99.1
root@d01eb56fce52:/# arping 172.17.99.2
ARPING 172.17.99.2
42 bytes from aa:b3:98:92:0b:08 (172.17.99.2): index=0 time=1.001 sec
42 bytes from aa:b3:98:92:0b:08 (172.17.99.2): index=1 time=1.001 sec
42 bytes from aa:b3:98:92:0b:08 (172.17.99.2): index=2 time=1.001 sec
42 bytes from aa:b3:98:92:0b:08 (172.17.99.2): index=3 time=1.001 sec
^C
--- 172.17.99.2 statistics ---
5 packets transmitted, 4 packets received,  20% unanswered (0 extra)
Run Code Online (Sandbox Code Playgroud)

所以似乎数据包可以通过bridge1.

但问题是我不能ping等容器,我也不能经由像任何工具发送的任何IP数据包telnetnetcat.

相反,桥docker0和NIC eth0在所有容器中都能正常工作.

这是我的路线表

# 172.17.99.1
root@d01eb56fce52:/# ip route
default via 172.17.42.1 dev eth0 
172.17.0.0/16 dev eth0  proto kernel  scope link  src 172.17.0.17 
172.17.99.0/24 dev eth1  proto kernel  scope link  src 172.17.99.1
Run Code Online (Sandbox Code Playgroud)

和桥配置

# host
$ brctl show
bridge name bridge id       STP enabled interfaces
bridge1     8000.8a6b21e27ae6   no      veth1pl25432
                                        veth1pl25587
                                        veth1pl25753
docker0     8000.56847afe9799   no      veth7c87801
                                        veth953a086
                                        vethe575fe2

# host
$ brctl showmacs bridge1
port no mac addr        is local?   ageing timer
  1 8a:6b:21:e2:7a:e6   yes        0.00
  2 8a:a3:b8:90:f3:52   yes        0.00
  3 f6:0c:c4:3d:f5:b2   yes        0.00

# host
$ ifconfig
bridge1   Link encap:Ethernet  HWaddr 8a:6b:21:e2:7a:e6  
          inet6 addr: fe80::48e9:e3ff:fedb:a1b6/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:163 errors:0 dropped:0 overruns:0 frame:0
          TX packets:68 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0 
          RX bytes:8844 (8.8 KB)  TX bytes:12833 (12.8 KB)

# I'm showing only one veth here for simplicity
veth1pl25432 Link encap:Ethernet  HWaddr 8a:6b:21:e2:7a:e6  
          inet6 addr: fe80::886b:21ff:fee2:7ae6/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:155 errors:0 dropped:0 overruns:0 frame:0
          TX packets:162 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:1000 
          RX bytes:12366 (12.3 KB)  TX bytes:23180 (23.1 KB)

...
Run Code Online (Sandbox Code Playgroud)

和IP FORWARD链

# host
$ sudo iptables -x -v --line-numbers -L FORWARD
Chain FORWARD (policy ACCEPT 10675 packets, 640500 bytes)
num      pkts      bytes target     prot opt in     out     source               destination         
1       15018 22400195 DOCKER     all  --  any    docker0  anywhere             anywhere            
2       15007 22399271 ACCEPT     all  --  any    docker0  anywhere             anywhere             ctstate RELATED,ESTABLISHED
3        8160   445331 ACCEPT     all  --  docker0 !docker0  anywhere             anywhere            
4          11      924 ACCEPT     all  --  docker0 docker0  anywhere             anywhere            
5          56     4704 ACCEPT     all  --  bridge1 bridge1  anywhere             anywhere            
Run Code Online (Sandbox Code Playgroud)

注意规则5的pkts cound不是0,这意味着ping已经正确路由(在路由右后执行FORWARD链?),但不知何故没有到达目的地.

我没有想法为什么docker0bridge1表现不同.有什么建议吗?

更新1

这是tcpdump目标容器从另一个容器中输出时的输出.

$ tcpdump -i eth1
tcpdump: verbose output suppressed, use -v or -vv for full protocol decode
listening on eth1, link-type EN10MB (Ethernet), capture size 65535 bytes
22:11:17.754261 IP 192.168.1.65 > 172.17.99.1: ICMP echo request, id 26443, seq 1, length 6
Run Code Online (Sandbox Code Playgroud)

请注意,源IP是192.168.1.65,这是eth0主机的,因此似乎有些SNAT桥上回事.

最后,打印出natIP表揭示了问题的原因:

$ sudo iptables -L -t nat
...
Chain POSTROUTING (policy ACCEPT)
target     prot opt source               destination         
MASQUERADE  all  --  172.17.0.0/16        anywhere
...
Run Code Online (Sandbox Code Playgroud)

因为我的容器的eth0IP已打开172.17.0.0/16,所以发送的数据包的源IP已更改.这就是为什么来自的回复ping不能回到源头.

结论

解决方案是将容器的eth0IP 更改为与默认网络不同的网络docker0.

sta*_*wer 1

复制自Update 1相关问题

以下是tcpdump从另一个容器 ping 时目标容器的输出。

$ tcpdump -i eth1
tcpdump: verbose output suppressed, use -v or -vv for full protocol decode
listening on eth1, link-type EN10MB (Ethernet), capture size 65535 bytes
22:11:17.754261 IP 192.168.1.65 > 172.17.99.1: ICMP echo request, id 26443, seq 1, length 6
Run Code Online (Sandbox Code Playgroud)

请注意,源 IP 是192.168.1.65,即eth0主机的 IP,因此网桥上似乎正在进行一些 SNAT。

最后打印出natIP表就知道了问题的原因:

$ sudo iptables -L -t nat
...
Chain POSTROUTING (policy ACCEPT)
target     prot opt source               destination         
MASQUERADE  all  --  172.17.0.0/16        anywhere
...
Run Code Online (Sandbox Code Playgroud)

因为我的容器eth0的IP是打开的172.17.0.0/16,所以发送的数据包的源IP发生了变化。这就是为什么来自的响应ping无法返回源的原因。

结论

解决方案是将容器的eth0IP 更改为与默认网络不同的网络docker0