[Pce] Re: Comments on draft-lee-pce-global-concurrent-optimization-03.txt

"Adrian Farrel" <adrian@olddog.co.uk> Fri, 11 May 2007 09:15 UTC

Return-path: <pce-bounces@lists.ietf.org>
Received: from [127.0.0.1] (helo=stiedprmman1.va.neustar.com) by megatron.ietf.org with esmtp (Exim 4.43) id 1HmRDk-0002d7-GT; Fri, 11 May 2007 05:15:28 -0400
Received: from [10.91.34.44] (helo=ietf-mx.ietf.org) by megatron.ietf.org with esmtp (Exim 4.43) id 1HmRDj-0002d2-CG for pce@ietf.org; Fri, 11 May 2007 05:15:27 -0400
Received: from rutherford.zen.co.uk ([212.23.3.142]) by ietf-mx.ietf.org with esmtp (Exim 4.43) id 1HmRDi-0007D9-Fx for pce@ietf.org; Fri, 11 May 2007 05:15:27 -0400
Received: from [88.96.235.142] (helo=cortex.aria-networks.com) by rutherford.zen.co.uk with esmtp (Exim 4.50) id 1HmRDg-0000Rx-3o for pce@ietf.org; Fri, 11 May 2007 09:15:24 +0000
Received: from your029b8cecfe ([217.158.132.35] RDNS failed) by cortex.aria-networks.com with Microsoft SMTPSVC(6.0.3790.3959); Fri, 11 May 2007 10:15:21 +0100
Message-ID: <00a201c793ac$dcfe9bc0$4802010a@your029b8cecfe>
From: Adrian Farrel <adrian@olddog.co.uk>
To: JP Vasseur <jvasseur@cisco.com>, Young Lee <ylee@huawei.com>
References: <001b01c7933f$605f27a0$520c7c0a@china.huawei.com> <1C05F8AA-1A9E-4130-8A25-1FE88ADE8576@cisco.com>
Date: Fri, 11 May 2007 10:13:45 +0100
Organization: Old Dog Consulting
MIME-Version: 1.0
Content-Type: text/plain; format="flowed"; charset="iso-8859-1"; reply-type="response"
Content-Transfer-Encoding: 7bit
X-Priority: 3
X-MSMail-Priority: Normal
X-Mailer: Microsoft Outlook Express 6.00.2900.2180
X-MimeOLE: Produced By Microsoft MimeOLE V6.00.2900.2180
X-OriginalArrivalTime: 11 May 2007 09:15:21.0724 (UTC) FILETIME=[E6AF13C0:01C793AC]
X-Originating-Rutherford-IP: [88.96.235.142]
X-Spam-Score: 0.2 (/)
X-Scan-Signature: 8cb9b411340046bf4080a729180a0672
Cc: pce@ietf.org
Subject: [Pce] Re: Comments on draft-lee-pce-global-concurrent-optimization-03.txt
X-BeenThere: pce@lists.ietf.org
X-Mailman-Version: 2.1.5
Precedence: list
Reply-To: Adrian Farrel <adrian@olddog.co.uk>
List-Id: Path Computation Element <pce.lists.ietf.org>
List-Unsubscribe: <https://www1.ietf.org/mailman/listinfo/pce>, <mailto:pce-request@lists.ietf.org?subject=unsubscribe>
List-Archive: <http://www1.ietf.org/pipermail/pce>
List-Post: <mailto:pce@lists.ietf.org>
List-Help: <mailto:pce-request@lists.ietf.org?subject=help>
List-Subscribe: <https://www1.ietf.org/mailman/listinfo/pce>, <mailto:pce-request@lists.ietf.org?subject=subscribe>
Errors-To: pce-bounces@lists.ietf.org

[Tidied up from Digest response thread]

<chair hat firmly OFF>


>>> The authors requested the WG to adopt draft-lee-pce-global-
>>> concurrent-optimization-03.txt as a PCE WG document but before
>>> pooling the list, >>> I'd like to make a few comments/requests:
>>>
>>> This solution is indeed compliant with RFC4655 and as pointed out in the 
>>> ID, PCEP already supports synchronized path computation requests through 
>>> the use of the SVEC object.
>>>
>>> 1) The PCEP extensions defined in this document are quite
>>> reasonable and do not substantially overload the protocol
>>> itself. That being said, the exchange of a substantially large
>>> amount of data will unavoidably stress the machinery in a significant 
>>> way. Scalability of solutions trying to achieve
>>> global optimization have been discussed in length so I won't propose to 
>>> re-open a fairly old debate but it is well-understood
>>> that such solutions do not scale well and the major bottleneck is not 
>>> just the path computation itself but the bulk of data
>>> that must be exchanged, synchronization issues, failures during
>>> reoptimization and so on.

Indeed, the path computation bottle-neck is clearly an algorithmic
issue that is out of scope here, and I know some would argue that
solutions exist that remove that bottle-neck.

I think that the other items listed would valuably be the subject of
health warnings in an applicability statement as JP suggests, but it
should be noted that this does not limit the applicability based on
the network size or the set of LSPs being reoptimized. Rather, it
limits the applicability to how the reoptimization is being used and
coordinated by a management plane.

The difference between a global reoptimization and a bulk re-route
are subtle. It may appear that for bulk re-route all existing
services have failed and so the worst case is that connectivity is
not restored, but in packet networks FRR may be in use, so bulk
re-route failures are significant. The applicability in *that* case
includes make-before-break to ensure no disruption to traffic.

But it is also commonly accepted that restoration or re-route after
failure has two problems. The first is system congestion which is
at least as bad as global concurrent reoptimization because each
LSP-to-be restored is the subject of a PCReq and a PCRep as well as
signaling. The second is that the sequential nature of the path
computations will almost inevitably lead to blocking.

So, while I agree that the gco draft should have a very thorough
applicability section that should highlight the potential problems,
I don't think that there should be an assumption of non-applicability
to specific networks or scenarios.


>>> Thus I'd suggest to add some applicability
>>> section to this ID that would discuss the context in which such
>>> solution would apply (e.g. network with thousands of packet LSPs
>>> (hopefully not!), optical LSPs with a few hundreds of LSPs with multi-
>>> constraints optimization problems where bandwidth fragmentation is a
>>> real issue because of a limited number of discrete bandwidth values).
>>>
>> We can add applicability section to elaborate this request. We can
>> indicate that this mechanism applies specifically to GMPLS optical 
>> networks with a few hundreds of LSPs.

I would be very concerned about that limitation.
It is great to hear that you have specific intention to apply this work
to optical networks, but I do not see anything in the I-D or what JP said
that forces us to constrain the applicability in this way.
By all means, cite optical networks with a limited number of nodes as a
good example of where you might use this technique, but otherwise, you
should structure the applicability section with regard to the potential
problems, not the dataplane technology.


>>> 2)
>>>
>>>     It is also envisioned that network operators might
>>>     require a global concurrent path computation in the event of
>>>     catastrophic network failures, where a set of TE LSPs need to
>>>     be optimally rerouted in real-time.
>>>
>>> I do not think that such model could be used for "real-time"  rerouting.
>>>
>>  I agree with you. We can remove this statement.

Define "real-time".

The applicability to the problem space is clear. A well-known set of
LSP has been impacted by a network failure. Rather than compute new
paths on demand from the head-end LSPs with the consequent resource-
contention blocking problems, there may be considerable benefit in
computing the paths of the LSPs as a set. (Of course, this assumes
the availability of certain information, but an NMS might reasonably
have access to this and make the bulk computation request.)

Thus, if there is any modification to make, I suggest only deleting
"in real-time".

>>> 3)
>>>
>>>     The main focus of this document is to highlight the PCC-PCE
>>>     communication needs in support of a concurrent path computation 
>>> application and to define protocol extensions to
>>>     meet those needs.
>>>
>>> You may want to stress the fact that in your ID the PCC is an NMS
>>> system and this is key. Indeed, one can define models where the PCCs are 
>>> LSRs and the PCE is used to provide globally optimal
>>> solutions ... Such models suffers from drastic scalability and
>>> robustness issues.
>>
>> The PCE GCO is primarily an NMS based solution. In section 3.3
>> (Application of PCE architecture) of the current draft clearly  spells 
>> out that GCO is NMS based solution.  With GCO, a PCC has
>> to know all LSP requests, hence this cannot be a LSR.
>
> Well that could be done with PCC=LSR thanks to complex  synchronization 
> (which I'm NOT advocating of course)
> hence my comment. Could you restate in the abstract/Introduction that in 
> your proposal the PCC is indeed an NMS.

Why would the procedures not be applicable to a head-end LSR
requesting bulk reoptimization of the set of LSPs for which it acts
as ingress? (I.e. a subset of all LSPs in the system.)

We already accept that such an LSR should be allowed to issue
individual LSP reoptimization requests. And we already accept that
a PCReq may contain multiple computation requests that are 'linked'.

So, if JP's request is that an LSR should not issue an optimization
request for an LSP for which it is not the head-end, I can see the
point. Although I might want to rephrase this because the important
question is what use is made of the computation response - that is,
if the PCC is unable to make use of the computed path, there is no
value in performing the computation.

Clearly, NMS is a primary application of this work, but it is not
the only application. Another key use would be the VNT Manager
component discussed in the multi-layer PCE drafts.

>>> 4) Green field: not sure to buy this argument since as soon as
>>> the TE LSPs are set up, the network is no longer in this green
>>> field state
>>>
>>  OK.  The main use of GCO application is re-optimization of an  existing 
>> network.

Well, I would be very worried if you deleted the green-field
applicability. But I do agree that it should be put lower down the
list (although I guess it comes first in the list because it comes
first in time). And it should be rephrased, because it is not
"reoptimization" since (as JP points out) you can't reoptimize LSPs
that don't exist, and once the LSPs do exist it is not a green field.


But noting that the draft is titled "PCE Global Concurrent
Optimization" not "PCE Global Concurrent Re-optimization" I have zero
problem with the existence of this section, and I believe that
network planners will want to make use of computation servers to plan
the LSPs that they will provision in their network. This might arise
either in the green field condition or when rolling out a new
customer on top of an existing network.

>>> 5)
>>>
>>>     Note that sequential re-
>>>     optimization of such TE LSPs is unlikely to produce substantial 
>>> improvements in overall network optimization
>>>     except in very sparsely utilized networks.
>>>
>>> Well, that DEPENDS ! I could show you distributed algorithms where
>>> sequential reoptimization allows for a significant improvements. I
>>> would suggest to remove that statement.

Actually, I think that the sequential algorithms that JP refers to
are actually *global* reoptimization. That is, they sequence through
the set of LSPs making optimization improvements. The fact that the
algorithms are distributed is neither here nor there.

> Yes, it actually depends on the topology, the traffic matrix, the  online 
> algorithm used, etc. We will delete this statement.

So rather than deleting the statement, I suggest qualifying it.
Of course, the potential for network-wide gains from reoptimization
of LSPs one-by-one is dependent on the network usage and size of the
LSPs being reoptimized. But the key point remains: if you only
compute the reoptimized path of one LSP at a time, and if you give
no consideration to the other LSPs in the network when you do it,
you run the risk of significant devaluation of the process.

This may be far more visible in networks with a low ratio of potential
LSPs per link (such as in an optical network), and far less visible in
packet networks with micro-flow LSPs.

[SNIP]

>>> 7) A word of cautious here
>>>
>>>           During a reoptimization it may be required to move a LSP
>>>           several times so as to avoid traffic disruption.  The 
>>> response message must allow indicating the path sequence
>>>           for each request.
>>
>> We all know that in some cases, traffic disruption may be avoided
>> thanks to a multi-step rerouting approach where some TE LSP may be
>> rerouted N times. This is another example where such model may have
>> significant impact on the network and even when traffic disruption
>> can be avoided, there is still an impact in term of control plane,
>> traffic shift (=> jitter) although this can be another constraint
>> taken in to account when computing the various rerouting steps. For
>> example, would you want to add a paragraph listing the drawbacks of
>> such approach (e.g. trade-off between optimization gain and network
>> impact, ....) ?
>>
>> We can add a paragraph to indicate the potential impact by this  feature.
>> By the way the trade-off is not optimization vs. network impact. It
>> is traffic disruption vs. network impact. If we want to avoid traffic 
>> disruption, we need this multiple rerouting, which is the
>> price to pay at the expense of network impact.
>
> OK let me restate my comment here: the point I was trying to make is  the 
> following: even if the set of TE LSPs can be reoptimized with no
> traffic loss for example, the need for multiple reroutes has a
> cost in term of traffic impact (jitter, ...) + control plane stress.  It 
> may be worth mentioning that aspect also, this is why I mentioned
> a trade-off between optimization versus network impact. Indeed, if you can 
> reoptimize the set of TE LSPs in order to reduce the max link  utilization 
> by 5% but this requires to reroute two hundreds TE LSPs with on the 
> average 4 reroutes per TE LSP, this may not be a good idea.

Well, that would depend, wouldn't it? If the network is unable to place
any more LSPs, one might think that retrieving 5% was pretty cool :-)

But I *do* think that a paragraph warning about the risks of re-routing
LSPs is valuable. It is often assumed that make-before-break is hitless,
and it is not. As JP points out, even in packet networks where the re-
route can occur between packets, there may be a momentary jitter impact
at the point of switch-over. This applies to *all* reoptimization and
re-route activities including FRR.

So obviously there is a trade-off to be highlighted and put under
policy control. Maybe only some LSPs are subject to re-routing - others
need to preserve their QoS.

And clearly, multiple "shuffling" of LSPs to make space is going to
increase the impact on the network. This means both that there should
be cost-benefit analysis of the reoptimization, and that such
reoptimization is unlikely to be a continuous process.

>>> 8) Objective functions should be moved to PCEP,
>
> I typed too fast: I indeed meant to refer to the objective function 
> draft, as agreed with Jerry since I was pushing myself for not adding
> more to PCEP.

[SNIP]

JP, are you saying that all objective function definitions should go
into the one objective function I-D? Or should that I-D define the
functions that we have in hand, create the code point registry, and
leave the definition of new objective functions for future I-Ds?

But still, the requirements for the objective functions need to stay
in the gco draft. So it is just a question of where the objective
functions are described in detail.

[SNIP]

Thanks,
Adrian




_______________________________________________
Pce mailing list
Pce@lists.ietf.org
https://www1.ietf.org/mailman/listinfo/pce