Re: [trill] TRILL Resilient Distribution Trees, who have read it?

Mingui Zhang <zhangmingui@huawei.com> Wed, 19 December 2012 09:30 UTC

Return-Path: <zhangmingui@huawei.com>
X-Original-To: trill@ietfa.amsl.com
Delivered-To: trill@ietfa.amsl.com
Received: from localhost (localhost [127.0.0.1]) by ietfa.amsl.com (Postfix) with ESMTP id 8EBB921F8924 for <trill@ietfa.amsl.com>; Wed, 19 Dec 2012 01:30:40 -0800 (PST)
X-Virus-Scanned: amavisd-new at amsl.com
X-Spam-Flag: NO
X-Spam-Score: -4.458
X-Spam-Level:
X-Spam-Status: No, score=-4.458 tagged_above=-999 required=5 tests=[AWL=-1.488, BAYES_00=-2.599, FB_INCREASE_VOL=3.629, RCVD_IN_DNSWL_MED=-4]
Received: from mail.ietf.org ([64.170.98.30]) by localhost (ietfa.amsl.com [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id nCkB+L1GP7FI for <trill@ietfa.amsl.com>; Wed, 19 Dec 2012 01:30:39 -0800 (PST)
Received: from lhrrgout.huawei.com (lhrrgout.huawei.com [194.213.3.17]) by ietfa.amsl.com (Postfix) with ESMTP id EFFF921F8B0E for <trill@ietf.org>; Wed, 19 Dec 2012 01:30:38 -0800 (PST)
Received: from 172.18.7.190 (EHLO lhreml204-edg.china.huawei.com) ([172.18.7.190]) by lhrrg01-dlp.huawei.com (MOS 4.3.5-GA FastPath queued) with ESMTP id ANZ10834; Wed, 19 Dec 2012 09:30:37 +0000 (GMT)
Received: from LHREML402-HUB.china.huawei.com (10.201.5.241) by lhreml204-edg.china.huawei.com (172.18.7.223) with Microsoft SMTP Server (TLS) id 14.1.323.3; Wed, 19 Dec 2012 09:30:12 +0000
Received: from SZXEML445-HUB.china.huawei.com (10.82.67.183) by lhreml402-hub.china.huawei.com (10.201.5.241) with Microsoft SMTP Server (TLS) id 14.1.323.3; Wed, 19 Dec 2012 09:30:14 +0000
Received: from SZXEML507-MBS.china.huawei.com ([169.254.7.234]) by szxeml445-hub.china.huawei.com ([10.82.67.183]) with mapi id 14.01.0323.003; Wed, 19 Dec 2012 17:30:08 +0800
From: Mingui Zhang <zhangmingui@huawei.com>
To: Thomas Narten <narten@us.ibm.com>
Thread-Topic: [trill] TRILL Resilient Distribution Trees, who have read it?
Thread-Index: Ac3O0ux4S0JhVhpwS56Ahgx4dyNymwLBniSAAJL4eEAAM0G8gAAz5+dA
Date: Wed, 19 Dec 2012 09:30:07 +0000
Message-ID: <4552F0907735844E9204A62BBDD325E73213A81C@SZXEML507-MBS.china.huawei.com>
References: <4552F0907735844E9204A62BBDD325E732132977@SZXEML507-MBS.china.huawei.com> <201212141659.qBEGxxqU012075@cichlid.raleigh.ibm.com> <4552F0907735844E9204A62BBDD325E73213A467@SZXEML507-MBS.china.huawei.com> <201212181535.qBIFZqAs012411@cichlid.raleigh.ibm.com>
In-Reply-To: <201212181535.qBIFZqAs012411@cichlid.raleigh.ibm.com>
Accept-Language: en-US, zh-CN
Content-Language: zh-CN
X-MS-Has-Attach:
X-MS-TNEF-Correlator:
x-originating-ip: [10.111.102.185]
Content-Type: text/plain; charset="us-ascii"
Content-Transfer-Encoding: quoted-printable
MIME-Version: 1.0
X-CFilter-Loop: Reflected
Cc: "trill@ietf.org" <trill@ietf.org>
Subject: Re: [trill] TRILL Resilient Distribution Trees, who have read it?
X-BeenThere: trill@ietf.org
X-Mailman-Version: 2.1.12
Precedence: list
List-Id: "Developing a hybrid router/bridge." <trill.ietf.org>
List-Unsubscribe: <https://www.ietf.org/mailman/options/trill>, <mailto:trill-request@ietf.org?subject=unsubscribe>
List-Archive: <http://www.ietf.org/mail-archive/web/trill>
List-Post: <mailto:trill@ietf.org>
List-Help: <mailto:trill-request@ietf.org?subject=help>
List-Subscribe: <https://www.ietf.org/mailman/listinfo/trill>, <mailto:trill-request@ietf.org?subject=subscribe>
X-List-Received-Date: Wed, 19 Dec 2012 09:30:40 -0000

Hi Thomas,


>    They are seeing 34-100ms convergence times and think this can be
>    further reduced by an order of magnitude (recommendation 7)

When you say convergence time can be as short as 34 ms, it certainly does not include the failure detection time. Since failures are detected in ISIS by non-receipt of Hellos within three holding timers, which are "commonly set to a value of tens of seconds and, in any case, has a minimum expressible value of one second [rbBFD]". When the presenter mentioned the convergence time, I believe he meant "the time elapsed from a failure event realized by the control protocol until the forwarding has been recovered [CONVERGE]". However, the service disruption does include the failure detection time, during which the forwarding paths via the failure spot are already cut off. So, the service disruption time at least includes the following two parts: 1) Failure detection time; 2) Service resuming time (a.k.a. convergence time).

>Won't TRILL have similar convergence times? Are we talking about using
>failover to shave off a few tens of milliseconds? Or is there some
>sort of expectation that this will shave seconds or more off the
>recovery time?

We expect that "TRILL Resilient Distribution Trees" reduce the whole recovery time.

>Which applications are those? Are there particular deployment
>scenarios that are being targetted, or is this just a general
>argument?

Video distribution (e.g., IPTV) is such kind of application [RDT].

>And just how long does this period of "inconsistency" last? And just
>how quickly will a node know that a path that it is using has failed
>and that it needs to fallback to using the precomputed backup path?
>Won't it also take time for news of a failure to propagate through the
>network? In other words, won't the network already be reconfiguring
>itself even before a node realizes that a path has failed and it
>should switch to a backup?

The figure on page 4 of [CONVERGE] also tells us that inconsistence introduces a sky-high deal of looping traffic. This period can last seconds. RDT not only reduces the service disruption time but also decreases the looping traffic.

There are three kinds of protection described in the draft. Only the global 1:1 protection relies on the "news of a failure" to determine to switch to a backup.

We catch the issue in Section 5.4 of the draft. You will find there is no problem that network reconfiguration and the use of backup take place in parallel.

>So again, I'm really trying to understand just how much of an
>improvement the proposed approach is expected to produce.

Let me quantify the improvement we expected to produce:
"Service disruption time should be magnitude shorter than one second [RFC5714]". 
Exactly, we expect [RDT] meets "Realistic availability requirements (sub-second to sub-200msec)... [MOFRR]". 

References
[rbBFD] http://www.ietf.org/id/draft-ietf-trill-rbridge-bfd-07.txt
[CONVERGE] http://www.ieee802.org/1/files/public/docs2010/new-farkas-convergence-1110.pdf
[RDT] http://datatracker.ietf.org/doc/draft-zhang-trill-resilient-trees/?include_text=1
[RFC5714] http://tools.ietf.org/html/rfc5714
[MOFRR] http://tools.ietf.org/html/draft-karan-mofrr-02

Thanks,
Mingui






>-----Original Message-----
>From: Thomas Narten [mailto:narten@us.ibm.com]
>Sent: Tuesday, December 18, 2012 11:36 PM
>To: Mingui Zhang
>Cc: trill@ietf.org
>Subject: Re: [trill] TRILL Resilient Distribution Trees, who have read it?
>
>Hi.
>
>I understand at a high level what the reational is. What I'm really
>trying to get at is: does the cost/complexity justify the benefit? For
>this, some quantifiable metrics would help - not just generalities
>like "convergence will be faster", etc.
>
>Mingui Zhang <zhangmingui@huawei.com> writes:
>
>> Hi Thomas,
>
>> TRILL provides multicast service using ISIS link state routing. When
>>  there is a failure, RBridges are notified through the propagation
>>  of LSPs, which will trigger a campus-wide convergence on
>>  distribution trees. The convergence includes the following major
>>  processes: 1) the propagation of LSPs around the whole campus; 2)
>>  the recalculation of trees on affected RBridges; 3) the update of
>>  forwarding information on those RBridges. These processes make the
>>  convergence time-consuming.
>
>I take issue with this. The whole point of link-state flooding
>approaches is that they are fast - much faster than alternatives.
>
>And as a data point, back in Paris, Paul Unbehagen gave a presentation
>in the IS-IS WG on SPB deployment where (according to my notes):
>
>    They are seeing 34-100ms convergence times and think this can be
>    further reduced by an order of magnitude (recommendation 7)
>
>Won't TRILL have similar convergence times? Are we talking about using
>failover to shave off a few tens of milliseconds? Or is there some
>sort of expectation that this will shave seconds or more off the
>recovery time?
>
>That is, do we have any evidence (or targets) for TRILL convergence
>times (after a topology change) that are "too long"? Just "how fast"
>reconvergence is considered fast enough, and what is considered
>problematical?
>
>> As we know, most multicast traffic is
>>  generated by applications which are sensitive to interrupt
>>  latency. Therefore, the convergence will lead to the disruption of
>>  multicast service.
>
>Which applications are those? Are there particular deployment
>scenarios that are being targetted, or is this just a general
>argument?
>
>> Due to the propagation delay of LSPs, the distribution trees are
>>  calculated and installed on RBridges at different time, which
>>  brings _inconsistence_ of forwarding states among RBridges. Before
>>  the convergence finishes, multicast frames may be trapped by
>>  forwarding loops and finally dropped. This increases the volume of
>>  broadcast/multicast traffic.
>
>And just how long does this period of "inconsistency" last? And just
>how quickly will a node know that a path that it is using has failed
>and that it needs to fallback to using the precomputed backup path?
>Won't it also take time for news of a failure to propagate through the
>network? In other words, won't the network already be reconfiguring
>itself even before a node realizes that a path has failed and it
>should switch to a backup?
>
>So again, I'm really trying to understand just how much of an
>improvement the proposed approach is expected to produce.
>
>To be clear, I'm not necessarily opposed to this work. But I would
>like a better understanding of exactly how much benefit we can really
>expect, so I can understand how to weigh the tradeoffs between
>increased complexity vs. actual benefit.
>
>Thomas