Re: [Lsr] LSR Flooding Reduction Drafts - Moving Forward

Tony Przygienda <tonysietf@gmail.com> Fri, 24 August 2018 21:26 UTC

Return-Path: <tonysietf@gmail.com>
X-Original-To: lsr@ietfa.amsl.com
Delivered-To: lsr@ietfa.amsl.com
Received: from localhost (localhost [127.0.0.1]) by ietfa.amsl.com (Postfix) with ESMTP id 08E64129385 for <lsr@ietfa.amsl.com>; Fri, 24 Aug 2018 14:26:03 -0700 (PDT)
X-Virus-Scanned: amavisd-new at amsl.com
X-Spam-Flag: NO
X-Spam-Score: -1.999
X-Spam-Level:
X-Spam-Status: No, score=-1.999 tagged_above=-999 required=5 tests=[BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, FREEMAIL_FROM=0.001, HTML_MESSAGE=0.001, RCVD_IN_DNSWL_NONE=-0.0001, SPF_PASS=-0.001] autolearn=unavailable autolearn_force=no
Authentication-Results: ietfa.amsl.com (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com
Received: from mail.ietf.org ([4.31.198.44]) by localhost (ietfa.amsl.com [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id Vv6v4jkkjc15 for <lsr@ietfa.amsl.com>; Fri, 24 Aug 2018 14:26:00 -0700 (PDT)
Received: from mail-ed1-x52b.google.com (mail-ed1-x52b.google.com [IPv6:2a00:1450:4864:20::52b]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by ietfa.amsl.com (Postfix) with ESMTPS id 541EB130DC0 for <lsr@ietf.org>; Fri, 24 Aug 2018 14:26:00 -0700 (PDT)
Received: by mail-ed1-x52b.google.com with SMTP id h4-v6so6683379edi.6 for <lsr@ietf.org>; Fri, 24 Aug 2018 14:26:00 -0700 (PDT)
DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=578u/6kVJWbNMfhF2nB4Aph/y3AcEoOedvahVwGaYu4=; b=T2SLHCEpfPXSjh87BOhxHXg6qGLEgodSAYznOvoHoFr40KJ2zrcKJxl6f7GqTtUTtl RdHS6pcSkcCHK/t6PeiJ6ZcigsPkDZTlgOID53yC4qPih37xx9rasdynIhoQuDdqz1Dv 95uAE4IAEm1q7c5FOEJ+7COrw/Sn7KdG3qHqYYgGpLGIu8Eh5wtdietS9wiapA8W0Y5S cwjMpf0Cui8TcT7PSSI9isEfa+0C0hFNBvo+YkRHsg7zLzHO77p2kfAiyff612zqn0tP 9ii4nFKs8iOhEx1kxfKgjgBgEc7XjwUgp/8Oh4L2tXpSfvVM68x+22VsbN+83chQsIUX tgDg==
X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=578u/6kVJWbNMfhF2nB4Aph/y3AcEoOedvahVwGaYu4=; b=f2V81O7JZxxLGA1h7O9xYSu77HFlSfV/PGMaOP8D0RKJLdqZOnV9bgxKCWXKAGI5rU Qx+DHnQXk4HTWC4e5fWVpSzEbtpWK615t5kiq3VY9JcZpdfArgZ6lTEQTjTPC0B46uQE Umlds0TzpxPo58Q39ad5GKWFdXcWEj5VVoI+W3OzUPu9DVuYnf1WyTWYALYUi1Hl4b6k mHL2BCG4YOsJJmWHiqDJ9VB8g78OM+r2JWQUXUXBT5t3UdyCc1ZJdih+o/PYKhhIrUGr zLkBuCjIIcZHqFZJaY0jwHDh9sjpS4bjfsaqp/opQ1gnkb0sgXrKLnK6NqCtfBKO6jJB QI6A==
X-Gm-Message-State: APzg51BnqYDcr3bEwTDpeMpDgwz6geBgWBdgW8n740JtAzsAhe/VmuC7 mwelKnE1gXVMG2pbfC+xDszA15915ErYsx2AzZA=
X-Google-Smtp-Source: ANB0VdZz7GRoxCy+T9uN8FY5YC1fAoyX5eQ+RkPOd/X+2uaX+ilJiLOJW6hJ7Qx3blFxca3cFwG5VwMXB0uEaHUnPcw=
X-Received: by 2002:aa7:d7c5:: with SMTP id e5-v6mr4567572eds.195.1535145958895; Fri, 24 Aug 2018 14:25:58 -0700 (PDT)
MIME-Version: 1.0
References: <8F5D2891-2DD1-4E51-9617-C30FF716E9FB@cisco.com> <C64E476F-1C00-435E-9C74-BEC3053377E8@gmail.com> <2F5FDB3F-ADCA-4DB4-83DA-D2BC3129D2F2@gmail.com> <5B7E78DD.90302@cisco.com> <172728E8-49E6-4F43-9356-815E1F4C22E7@gmail.com> <5B7FCAB3.6040600@cisco.com> <3D1DEC37-ACE7-4412-BB2E-4C441A4E7455@tony.li> <CCF220A3-8308-47B8-8CC6-1989705FF05C@cisco.com> <CA+wi2hNv8AVyR81LRmJ=Pd5_p5rS2djCOjY9YDgKxG=KEO_MkA@mail.gmail.com> <39509D13-4D2D-49A9-8738-C9D1F7C54223@tony.li> <CA+wi2hPsPQinwL2KL+xZnTYa1dwuw=thKjq_ddHnJQNnr19ugg@mail.gmail.com> <F0C191B8-5888-47FA-A4D6-D9328AA675C7@tony.li>
In-Reply-To: <F0C191B8-5888-47FA-A4D6-D9328AA675C7@tony.li>
From: Tony Przygienda <tonysietf@gmail.com>
Date: Fri, 24 Aug 2018 14:25:22 -0700
Message-ID: <CA+wi2hOMbX6C-zmHaLkYZ-zFXZM9Tsgusz4PnvwQPY1gq8+P3g@mail.gmail.com>
To: Tony Li <tony.li@tony.li>
Cc: acee=40cisco.com@dmarc.ietf.org, Peter Psenak <ppsenak@cisco.com>, lsr@ietf.org, Jeff Tantsura <jefftant.ietf@gmail.com>
Content-Type: multipart/alternative; boundary="0000000000008ef2e30574350508"
Archived-At: <https://mailarchive.ietf.org/arch/msg/lsr/KdLNc9iBZBZnIn7dM4otWHjphWY>
Subject: Re: [Lsr] LSR Flooding Reduction Drafts - Moving Forward
X-BeenThere: lsr@ietf.org
X-Mailman-Version: 2.1.27
Precedence: list
List-Id: Link State Routing Working Group <lsr.ietf.org>
List-Unsubscribe: <https://www.ietf.org/mailman/options/lsr>, <mailto:lsr-request@ietf.org?subject=unsubscribe>
List-Archive: <https://mailarchive.ietf.org/arch/browse/lsr/>
List-Post: <mailto:lsr@ietf.org>
List-Help: <mailto:lsr-request@ietf.org?subject=help>
List-Subscribe: <https://www.ietf.org/mailman/listinfo/lsr>, <mailto:lsr-request@ietf.org?subject=subscribe>
X-List-Received-Date: Fri, 24 Aug 2018 21:26:03 -0000

On Fri, Aug 24, 2018 at 1:36 PM <tony.li@tony.li> wrote:

>
> Tony,
>
> as to miscabling: yepp, the protocol has either to prevent adjacencies
> coming up or one has to deal with generic topologies. If you don't want to
> design miscabling prevention/topology ZTP into protocol (like ROLL or RIFT
> did) you have to deal with generic graph as you say. I think that if you
> have 99.99% of the time a predictable graph it's easier to restrict wrong
> cabling than deal with arbitrary topology when tackling this problem but
> that's my read.
>
>
>
> Well, you can take that approach, but you are at a risk of ignoring the
> one link that you needed to make the entire topology work better.  ;-)
>

yepp, no free lunch ... IME it's even more sublte. The more regular your
topology the more one can summarize & contain blast radius, the less, the
more topology information needs flooded to avoid blackholes, bow-ties to
the point where a generic topology (assuming prefix mobility) on the fabric
leaves one with host routes & scale limitation ...


> And if you don’t like the extra link problem, there’s also the missing
> link: what do you do when you have a leaf-spine topology, but there’s one
> link missing?  It’s not like you can ignore the missing link and flood on
> it anyway.  ;-)
>
>
that works just fine on -02 RIFT ...


>
> Another observation though would be that if you have a single mesh then
> centralized controller delay on failure becomes your delay bound how long
> flooding is disrupted possibly (unless your single covering graph has
> enough redundancy to deal with single llink failure, but then you're really
> having two as I suggest ;-). That could get ugly since you'll need
> make-before-break if installing a new mesh from controller me thinks with a
> round-trip from possibly a lot of nodes …
>
>
>
> Perhaps you didn’t understand the draft in detail.
>


>
> Even the loss of the area leader does not disrupt flooding.
>
> The flooding topology is in effect until a new area leader is elected and
> a new topology is distributed.  Yes, there is a hole in the flooding
> topology and you’re no longer bi-connected, but as long as it was still a
> single failure, you should still have a functioning flooding topology.  And
> because of that, it’s reasonable to assume that the area members can elect
> a new leader and switch to the new topology in an orderly fashion.
>
> It’s very true that there is a period where things are not bi-connected
> and a second failure will cause a flooding problem.  That period should be
> on the order of one failure detection, one flooding propagation, and one
> SPF computation.  If we expect failures to happen more frequently than
> that, then we need to call that out in our requirements. That type of
> scenario is perhaps reasonable under a MANET environment, but does not
> match any of my experience with typical data centers.
>
>
 Yes, I read the early versions, maybe missed some newer stuff, will read
newest and ponder more ... Though I think we're in sync ...



>
> >  iii) change in one of the vertex lifts
>>
>>
>> Sorry, I don’t understand point iii).
>>
>
> A mildly stuffed (or mathematically concise ;-) way to say that if you
> have one or two covering graphs (and vertex lift is the more precise word
> here since "covering graph" can be also an edge lift which is irrelevant
> here) and one of those subgraphs gets recomputed & distributed (due to
> failures, changes in some metrics, _whatever_) then this should not lead to
> disruption. Basically make-before-break as one possible design point,
> harder to achieve of course in distributed fashion …
>
>
>
> I think it would help the discussion if we phrased it less concisely. :-)
>

You know Aesop's fable on father, son & the donkey ? '-) I am normally
accused of excessive verbiage ;-) so that's somewhat of a compliment ;-)
But will take more care before introducing terms in the future.


>
>
> > moreover, I observe that IME ISIS is much more robust under such
>> optimizations since the CSNPs catch (@ a somehow ugly delay cost) any
>> corner cases whereas OSPF after IDBE will happily stay out of sync forever
>> if flooding skips something (that may actually become a reason to introduce
>> periodic stuff on OSPF to do CSNP equivalent albeit it won't be trivial in
>> backwards compatible way on my first thought, I was thinking a bit about
>> cut/snapshot consistency check [in practical terms OSPF hellos could carry
>> a checksum on the DB headers] but we never have that luxury on a link-state
>> routing protocol [i.e. we always operate under unbounded epsilon
>> consistency ;-) and in case of BGP stable oscialltions BTW not even that
>> ;^} ]).
>>
>> Emacs
>>
>>
> And that I cannot parse. Emacs? You want LISP code? But then, Dino may get
> offended ;-)
>
>
>
> Your comment seemed like just a poke in the perennial IS-IS vs. OSPF
> debate, which seems about as constructive as the Vi vs. Emacs debate.
>
>
ah, got you. Was not my intention, I dislike and like both protocol
equivalently, each has its warts [though if you look @ RIFT it's shameless
ISIS flooding rip-off ;-), that part IMO is somewhat better in ISIS] ;-)
But I thought I bring this angle into the thread since direct experience in
dealing with the problem of flood reduction taught me the ISIS
belts-suspender flooding can really save the bacon in such
algorithms/improvements. Unless we assume the algorithm will be perfect and
never loose anything, like never ever ... ;-)

e'one a good weekend ...

--- tony