mtu
Bernhard Schmidt
berni at birkenwald.de
Tue Feb 3 00:13:55 CET 2009
On Tue, Feb 03, 2009 at 08:13:04AM +1100, Geoff Huston wrote:
Hello Geoff,
> It seems that the most pragmatic advice is to use a lower-than-1500 MTU
> from the outset for IPv6 servers. More details of what I found are at
> http://ispcolumn.isoc.org/2009-01/mtu6.html
Great summary, thanks a lot for this excellent article!
However, I cannot agree with your conclusion. On the one hand we see a
lot of people, especially in the research networks lobbying for networks
that are transparent for 4470 byte or ~9000 byte frames to reduce
overhead and increase (host) throughput with >>GE rates, and on the
other hand we cannot even get 1500 byte reliably and propose to set the
MTU of servers (all servers?) to a lower value? That can't be right.
We need to rat out those issues before it's too late. We have been
running our webservers (http://www.lrz.de, granted it's not really high
profile) v6-enabled with MTU 1500 for three years now, same for MX and
other services. We have some thousand clients all running on MTU 1500
links (and thus sending out MSS 1440 and relying on IPv6 pMTU discovery)
and have not heard complaints. If you are repeatedly having issues chances
are high that the problem is close to your own network, which might make
debugging and contacting the appropriate parties a bit easier.
The most common issues are IPv6 tunnels on interprovider links. Most
implementations (including Cisco and Juniper) set the IPv6 MTU of the
tunnel to the IPv4 MTU (to the tunnel destination) minus 20/24 bytes of
overhead. Which is a good default when you have the standard 1500 byte
core links, but bad when your tunnelbox is connected with (for example)
POS (4470 bytes) or some jumbo-enabled ethernet link. IPv6 MTU is set to
4450 bytes, a native 1500 byte IPv6 packet comes in, gets encapsulated,
sent as 1520 byte IPv4 packet through the core and then dies at the 1500
byte IX fabric to the peer. Bam!
I've seen this issue multiple times now. These defaults are service
affecting. Even worse, two or three times when I told the engineering
folks of the affected network about the problem they did not fix the
tunnel immediately or turned it down, but kept it running. After all
they see traffic through it, so it can't be broken. But they broke a lot
of connections through that tunnel without even noticing it. So a lot
of education about pMTU and the effects of broken pMTU due to
misconfigured tunnels is still necessary. Your article, although a bit
lengthy for a quick slap, is very helpful in this.
If you are having issues and you have a Linux box around, try
tracepath6. It is a really great tool to find MTU issues on routers, and
notifying the affected party of this problem helps you and a lot of
other people.
OBtracerouteoftheday:
traceroute to www.rfc-editor.org (2001:1878:400:1:214:4fff:fe67:9351) from 2001:4ca0:0:f000:211:43ff:fe7e:3a76, port 80, from port 60366, 30 hops max, 60 byte packets
1 vl-23.csr2-2wr.lrz-muenchen.de (2001:4ca0:0:f000::2) 0.400 ms 0.302 ms 0.282 ms
2 vl-3051.csr1-2wr.lrz-muenchen.de (2001:4ca0:0:51::1) 0.530 ms 0.400 ms 0.336 ms
3 xr-gar1-te1-3-108.x-win.dfn.de (2001:638:c:a003::1) 0.543 ms 0.393 ms 0.362 ms
4 2001:638:c:c043::2 (2001:638:c:c043::2) 8.272 ms 8.133 ms 7.889 ms
5 dfn.rt1.fra.de.geant2.net (2001:798:14:10aa::1) 7.826 ms 7.808 ms 7.834 ms
6 abilene-wash-gw.rt1.fra.de.geant2.net (2001:798:14:10aa::12) 100.938 ms 119.152 ms 106.065 ms
7 2001:468:ff:209::2 (2001:468:ff:209::2) 117.410 ms 117.312 ms 117.330 ms
8 2001:468:ff:204:8000::2 (2001:468:ff:204:8000::2) 127.892 ms 127.959 ms 127.867 ms
9 2001:468:ff:407::2 (2001:468:ff:407::2) 174.721 ms 173.358 ms 173.695 ms
10 2001:468:ff:716::1 (2001:468:ff:716::1) 173.303 ms 185.809 ms 174.810 ms
11 kreonet-1-lo-jmb-706.sttlwa.pacificwave.net (2001:504:b:10::6) 169.214 ms 169.212 ms 169.147 ms
12 cenichpr-1-is-jmb-778.snvaca.pacificwave.net (2001:504:b:88::129) 184.891 ms 184.551 ms 184.509 ms
13 lax-hpr--svl-hpr-10ge.cenic.net (2001:468:e00:403::1) 184.587 ms 184.534 ms 184.519 ms
14 2607:f380::4:0:103 (2607:f380::4:0:103) 180.625 ms 180.597 ms 180.691 ms
15 2001:1878:8::2 (2001:1878:8::2) 180.895 ms 180.827 ms 180.773 ms
16 www.rfc-editor.org (2001:1878:400:1:214:4fff:fe67:9351) 181.112 ms [open] 181.000 ms 180.865 ms
Am I the only one shuddering when I see Kreonet in the middle of an
Europe-to-US trace? Fortunately traffic stays within .us and does not
take a detour, but that might just be my luck today.
Bernhard
More information about the ipv6-ops
mailing list