mtu

Bernhard Schmidt berni at birkenwald.de
Tue Feb 3 00:13:55 CET 2009


On Tue, Feb 03, 2009 at 08:13:04AM +1100, Geoff Huston wrote:

Hello Geoff,

> It seems that the most pragmatic advice is to use a lower-than-1500 MTU
> from the outset for IPv6 servers. More details of what I found are at
> http://ispcolumn.isoc.org/2009-01/mtu6.html

Great summary, thanks a lot for this excellent article!

However, I cannot agree with your conclusion. On the one hand we see a
lot of people, especially in the research networks lobbying for networks
that are transparent for 4470 byte or ~9000 byte frames to reduce
overhead and increase (host) throughput with >>GE rates, and on the
other hand we cannot even get 1500 byte reliably and propose to set the
MTU of servers (all servers?) to a lower value? That can't be right.

We need to rat out those issues before it's too late. We have been
running our webservers (http://www.lrz.de, granted it's not really high
profile) v6-enabled with MTU 1500 for three years now, same for MX and
other services. We have some thousand clients all running on MTU 1500
links (and thus sending out MSS 1440 and relying on IPv6 pMTU discovery)
and have not heard complaints. If you are repeatedly having issues chances
are high that the problem is close to your own network, which might make
debugging and contacting the appropriate parties a bit easier.

The most common issues are IPv6 tunnels on interprovider links. Most
implementations (including Cisco and Juniper) set the IPv6 MTU of the
tunnel to the IPv4 MTU (to the tunnel destination) minus 20/24 bytes of
overhead. Which is a good default when you have the standard 1500 byte
core links, but bad when your tunnelbox is connected with (for example)
POS (4470 bytes) or some jumbo-enabled ethernet link. IPv6 MTU is set to
4450 bytes, a native 1500 byte IPv6 packet comes in, gets encapsulated,
sent as 1520 byte IPv4 packet through the core and then dies at the 1500
byte IX fabric to the peer. Bam!

I've seen this issue multiple times now. These defaults are service
affecting. Even worse, two or three times when I told the engineering
folks of the affected network about the problem they did not fix the
tunnel immediately or turned it down, but kept it running. After all
they see traffic through it, so it can't be broken. But they broke a lot
of connections through that tunnel without even noticing it. So a lot
of education about pMTU and the effects of broken pMTU due to
misconfigured tunnels is still necessary. Your article, although a bit
lengthy for a quick slap, is very helpful in this.

If you are having issues and you have a Linux box around, try
tracepath6. It is a really great tool to find MTU issues on routers, and
notifying the affected party of this problem helps you and a lot of
other people. 

OBtracerouteoftheday:

traceroute to www.rfc-editor.org (2001:1878:400:1:214:4fff:fe67:9351) from 2001:4ca0:0:f000:211:43ff:fe7e:3a76, port 80, from port 60366, 30 hops max, 60 byte packets
 1  vl-23.csr2-2wr.lrz-muenchen.de (2001:4ca0:0:f000::2)  0.400 ms  0.302 ms  0.282 ms
 2  vl-3051.csr1-2wr.lrz-muenchen.de (2001:4ca0:0:51::1)  0.530 ms  0.400 ms  0.336 ms
 3  xr-gar1-te1-3-108.x-win.dfn.de (2001:638:c:a003::1)  0.543 ms  0.393 ms  0.362 ms
 4  2001:638:c:c043::2 (2001:638:c:c043::2)  8.272 ms  8.133 ms  7.889 ms
 5  dfn.rt1.fra.de.geant2.net (2001:798:14:10aa::1)  7.826 ms  7.808 ms  7.834 ms
 6  abilene-wash-gw.rt1.fra.de.geant2.net (2001:798:14:10aa::12)  100.938 ms  119.152 ms  106.065 ms
 7  2001:468:ff:209::2 (2001:468:ff:209::2)  117.410 ms  117.312 ms  117.330 ms
 8  2001:468:ff:204:8000::2 (2001:468:ff:204:8000::2)  127.892 ms  127.959 ms  127.867 ms
 9  2001:468:ff:407::2 (2001:468:ff:407::2)  174.721 ms  173.358 ms  173.695 ms
10  2001:468:ff:716::1 (2001:468:ff:716::1)  173.303 ms  185.809 ms  174.810 ms
11  kreonet-1-lo-jmb-706.sttlwa.pacificwave.net (2001:504:b:10::6)  169.214 ms  169.212 ms  169.147 ms
12  cenichpr-1-is-jmb-778.snvaca.pacificwave.net (2001:504:b:88::129)  184.891 ms  184.551 ms  184.509 ms
13  lax-hpr--svl-hpr-10ge.cenic.net (2001:468:e00:403::1)  184.587 ms  184.534 ms  184.519 ms
14  2607:f380::4:0:103 (2607:f380::4:0:103)  180.625 ms  180.597 ms  180.691 ms
15  2001:1878:8::2 (2001:1878:8::2)  180.895 ms  180.827 ms  180.773 ms
16  www.rfc-editor.org (2001:1878:400:1:214:4fff:fe67:9351)  181.112 ms [open]  181.000 ms  180.865 ms

Am I the only one shuddering when I see Kreonet in the middle of an 
Europe-to-US trace? Fortunately traffic stays within .us and does not
take a detour, but that might just be my luck today.

Bernhard


More information about the ipv6-ops mailing list