Re: WGLC for draft-ietf-bfd-large-packets

Gyan Mishra <> Wed, 16 October 2019 04:47 UTC

Return-Path: <>
Received: from localhost (localhost []) by (Postfix) with ESMTP id 50B5912006F for <>; Tue, 15 Oct 2019 21:47:00 -0700 (PDT)
X-Virus-Scanned: amavisd-new at
X-Spam-Flag: NO
X-Spam-Score: -1.998
X-Spam-Status: No, score=-1.998 tagged_above=-999 required=5 tests=[BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, FREEMAIL_FROM=0.001, RCVD_IN_DNSWL_NONE=-0.0001, SPF_HELO_NONE=0.001, SPF_PASS=-0.001, URIBL_BLOCKED=0.001] autolearn=ham autolearn_force=no
Authentication-Results: (amavisd-new); dkim=pass (2048-bit key)
Received: from ([]) by localhost ( []) (amavisd-new, port 10024) with ESMTP id 6HctnJacBelR for <>; Tue, 15 Oct 2019 21:46:58 -0700 (PDT)
Received: from ( [IPv6:2607:f8b0:4864:20::72b]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by (Postfix) with ESMTPS id C9872120090 for <>; Tue, 15 Oct 2019 21:46:57 -0700 (PDT)
Received: by with SMTP id w2so21560840qkf.2 for <>; Tue, 15 Oct 2019 21:46:57 -0700 (PDT)
DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed;; s=20161025; h=from:mime-version:subject:in-reply-to:date:cc :content-transfer-encoding:message-id:references:to; bh=z/Thy27Ka9fWGBXIscck/Ot8oOKxP1/ESYMWYP7TrWE=; b=FnHAehXETOwrPvM6PK4WxADQ5dNhlRnW84ke1C7fazBaAmi+y0ID41FvFUUr+yNfGE rvK4auFBBUYAaHQTH/Lf7a4aD/U957/oQbfxq4R/83ow08RyCPuIST3K094Bcp9z9E+y qU29q5LIquYINrRYglXhaYWFi0gwCGvNcm0K4NNJO5rf5NRTB2K+xgzEZT7Z+SeK2dOG liz1fQ1NGMTSl88ex9olINUtfDVtVEmjO3Uq+/Io72GuAoK3LbQ0ynsTLae2GVDGOhIW vDEqz2H51NHJjIOgigzH6T8XhwzCUCI1kGqqCUcesJNWNUZd6h4abBCmCebuxMBCwDOm +yiQ==
X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed;; s=20161025; h=x-gm-message-state:from:mime-version:subject:in-reply-to:date:cc :content-transfer-encoding:message-id:references:to; bh=z/Thy27Ka9fWGBXIscck/Ot8oOKxP1/ESYMWYP7TrWE=; b=NZY8l3+OFDYvygDTO9FnskRRxFoXiT0m+xQ1Dfm8beTLoh0kWyuu7dx8buS6mitbUf bav3u7RQHjrCAKGOI/MkYXwrDCf4c1Hy3HbChnwY5772w1HGNkofq9O1kzhMhvZvJ3TO s3vUrGAGVi9KmzdzXbIF5STPNsO0oGIFAJQigkMu91VHMEgxg0ClgkITLzr8lsyhJE27 9hG24J6nRvmrEbN6+7U/249O1eM5CpXSRmajRyvn0+UcEcbCkr16nw551Rs9vVEVCjbz QP7N9oRLLz9OLGKHWZVOGR+MZlIRpqGuPG3gLV/CkqNQUcvNuGH1HYo59scKPlF80JuZ CPpw==
X-Gm-Message-State: APjAAAW11vezwtzYWHcL/6hDUCcNlD4R1czUvyqDuiP5n16a3swawjMs lzmKJyQIQxdTcLhN1CYbKzw=
X-Google-Smtp-Source: APXvYqyPvmKREcQHVjoVTPIxyUQU3ZbrkNRbiAJzP+ONZ8Z1nz2awbx/8rvRKpS74KoIRSeb4a9Z0Q==
X-Received: by 2002:a37:98c2:: with SMTP id a185mr40095127qke.438.1571201216451; Tue, 15 Oct 2019 21:46:56 -0700 (PDT)
Received: from [] ( []) by with ESMTPSA id q47sm16886749qtq.95.2019. (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Tue, 15 Oct 2019 21:46:51 -0700 (PDT)
From: Gyan Mishra <>
X-Google-Original-From: Gyan Mishra <>
Content-Type: text/plain; charset=utf-8
Mime-Version: 1.0 (1.0)
Subject: Re: WGLC for draft-ietf-bfd-large-packets
X-Mailer: iPhone Mail (16G102)
In-Reply-To: <>
Date: Wed, 16 Oct 2019 00:46:51 -0400
Cc: Jeffrey Haas <>, "Reshad Rahman (rrahman)" <>, "" <>
Content-Transfer-Encoding: quoted-printable
Message-Id: <>
References: <> <> <> <> <> <> <> <> <> <> <> <>
To: "Les Ginsberg (ginsberg)" <>
Archived-At: <>
X-Mailman-Version: 2.1.29
Precedence: list
List-Id: "RTG Area: Bidirectional Forwarding Detection DT" <>
List-Unsubscribe: <>, <>
List-Archive: <>
List-Post: <>
List-Help: <>
List-Subscribe: <>, <>
X-List-Received-Date: Wed, 16 Oct 2019 04:47:00 -0000

In-line comment
Sent from my iPhone

> On Oct 3, 2019, at 8:02 PM, Les Ginsberg (ginsberg) <> wrote:
> Jeff -
> For some reason this is proving to be harder than I think it should be.
> I keep thinking I am being transparent - yet you keep reading "ulterior motives" into what I say.
> There are no ulterior motives.
> Let me try again...inline...
>> -----Original Message-----
>> From: Jeffrey Haas <>
>> Sent: Thursday, October 03, 2019 1:13 PM
>> To: Les Ginsberg (ginsberg) <>
>> Cc: Ketan Talaulikar (ketant) <>om>; Reshad Rahman
>> (rrahman) <>om>;
>> Subject: Re: WGLC for draft-ietf-bfd-large-packets
>> Les,
>> On Fri, Sep 27, 2019 at 09:14:08PM +0000, Les Ginsberg (ginsberg) wrote:
>>>> The primary reason this is a "may" in the non-RFC 2119 sense is that our
>>>> experience also suggests that when the scaling impacts are primarily pps
>>>> rather than bps that this feature will likely have no major impact on
>>>> implementations beyond your valid concerns about exercising bugs.
>>>> I suspect had this not been mentioned at all, you would have been
>> happier.
>>>> But you're not the target audience for this weak caveat.
>>> [Les:] I am not opposed to a discussion of potential issues in the draft -
>>> rather I am encouraging it. But the current text isn't really on the mark
>>> as far as potential issues - and we seem to agree on that. It also
>>> suggests lengthening detection time to compensate - which I think is not
>>> at all what you want to suggest as it diminishes the value of the
>>> extension. It also isn't likely to address a real problem.
>> I think what I'm seeing from you is roughly:
>> - Note that larger MTUs may have impact on some implementations for BFD
>>  throughput.
>> - And simply stop there.
> [Les:] What I would like to see discussed are points "a" and "b" below.
> This is a section on deployment issues - not a normative part of the spec.
>>> For me, the potential issues are:
>>> a)Some BFD implementations might not be able to handle MTU sized BFD
>>> packets - not because of performance - but because they did not expect
>> BFD
>>> packets to be full size and therefore might have issues passing a large
>>> packet through the local processing engine.
>> In such cases, the BFD session wouldn't be able to come up.  Are you
>> picturing a problem more dire than that?
> [Les:] No. Again, as this is a discussion of deployment considerations I see this as an aid to indicate what problems may be seen.
> I am not asking you to "fix" the extension to overcome this.
>>> b)Accepted MTU is impacted by encapsulations and what layer is being
>>> considered (L2 or L3). And oftentimes link MTUs do not match on both
>> ends
>>> ("shudder"), so you might end up with unidirectional connectivity.
>> Did you mean for BFD or more in the general sense?
> [Les:] It is a problem in the general sense, but it is relevant here because the extension proposes to send large packets. Absent that, MTU mismatches would be very unlikely to affect BFD since the BFD packet size is small.
>> For BFD, if you have one side testing for large MTU but not the other, we
>> can still have a Up BFD session with possible packet drop for large packets
>> on the opposite side.  But there's the chance in some paths that MTU may be
>> unidirectionally different - e.g. satellite down vs. land up.[1]
> In such cases, configuring BFD large on both sides would be the right
>> answer.  But it's also possible that large packets may only need to be
>> unidirectionally delivered.
> [Les:] I agree - and I think it is valid to use the extension unidirectionally in such cases.
>>> I
>>> appreciate that this is exactly the problem that the extensions are
>>> designed to detect. I am just asking that these issues be discussed more
>>> explicitly as an aid to the implementor. If that also makes Transports ADs
>>> happier that is a side benefit - but that's not my motivation.
>> We're happy to have that in the document.
> [Les:] Great!!
>>>>> What might be better?
>>>>> 1)Some statement that MTU isn't necessarily a consistent value for all
>>>>> systems connected to an interface - which can impact the results when
>> large
>>>>> BFD packets are used. Implementations might then want to consider
>>>>> supporting "bfd-mtu" configuration and/or iterating across a range of
>> packet
>>>>> sizes to determine what works and what doesn't.
>>>> I'm not clear what you intend by this statement.
>>>> Are you asking that we emphasize the use case in a different way?  The
>>>> Introduction currently states:
>>>>  "However,
>>>>   some applications may require that the Path MTU [RFC1191] between
>>>>   those two systems meets a certain minimum criteria.  When the Path
>>>>   MTU decreases below the minimum threshold, those applications may
>>>>   wish to consider the path unusable."
>>>> I'm also unclear what "Implementations" may refer to here.  BFD?  An
>>>> arbitrary user application?  If the latter, the application may not have
>>>> strict control over the generation of a given PDU size; e.g. TCP
>>>> applications.
>>> [Les:] I am talking about BFD implementations.
>>> I suppose one can imagine each BFD client requesting a certain MTU value -
>>> but that wouldn't be my choice.
>> BFD conversations happen between pairs of devices.  In the case that you
>> have multiple devices connected to a network segment, each conversation
>> could (and may intentionally) have different properties.
>> An easy example of this is two devices running an IGP may want fast failure
>> and two other devices running BGP may be happy with just under second-
>> level
>> failure.
>> So too could some device decide that it cares about bi-directional
>> path MTU while the others may not.
> [Les:] I agree. My point was BFD sessions are requested by clients (such as a routing protocol). That client may/may not care about MTU e.g., a routing protocol may not use MTU sized packets.
> But if the goal is to validate that MTU sized data traffic can successfully be sent then "someone" has to enable that. And I would argue that the most logical place to enable the feature is under BFD itself since a routing protocol (BGP/OSPF) won't necessarily care about MTU.
> Since you are speaking at the "device" level (not BFD client level) I think we are in agreement.
>> Given prior BFD documents' lack of discussion about such multi-access
>> network considerations, I'm not sure it's in character to have it just for
>> such a case, if that's what you're concerned with.
>>> I would think the value we want is really the maximum L3 payload that the
>>> link is intended to support - which should be independent of the BFD
>>> client. This might be larger than any client actually uses - but that
>>> seems like a good thing.
>> In this case we have actual existence proof of desired behavior.  The links
>> may be 9k but the user cares only about 1500 bytes end to end. If 1500 bytes
>> for BFD large works but 9k doesn't, we've not tested what the user actually
>> desired.
> [Les:] This is fine. This is consistent with my suggestion that an implementation supports a "bfd-mtu" knob. This value can be <= link_mtu.
>    Les

[Gyan] As far as routing protocols go OSPF does MTU check which has to match or the neighbor will not establish and that is during the initial exchange.  BGP does not have an MTU check however pmtud is generally enabled by default so the MSS should change from default 560 to 1460 on 1500 MTU interfaces and higher on jumbo enabled interfaces but still an MTU mismatch will not prevent the neighbor from coming up.  ISIS also check MTU which has to match as well or the neighbor will not establish. I think this draft is very useful in testing the large jumbo packets throughout bidirectional state with the interface MTU size used in the control packets instead of the default smaller packets which should detect MTU mismatch on point to point links and asynchronous mode or even echo mode won’t bring routing protocol up with the mismatch since now the larger jumbo configured interface MTU is used in the control packets.
>> -- Jeff