Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt
Jinoo Joung <jjoung@smu.ac.kr> Fri, 22 September 2023 02:50 UTC
Return-Path: <jjoung@smu.ac.kr>
X-Original-To: detnet@ietfa.amsl.com
Delivered-To: detnet@ietfa.amsl.com
Received: from localhost (localhost [127.0.0.1]) by ietfa.amsl.com (Postfix) with ESMTP id 08D52C152577 for <detnet@ietfa.amsl.com>; Thu, 21 Sep 2023 19:50:13 -0700 (PDT)
X-Virus-Scanned: amavisd-new at amsl.com
X-Spam-Flag: NO
X-Spam-Score: 3.097
X-Spam-Level: ***
X-Spam-Status: No, score=3.097 tagged_above=-999 required=5 tests=[BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, GB_SUMOF=5, HTML_MESSAGE=0.001, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_ZEN_BLOCKED_OPENDNS=0.001, SPF_HELO_NONE=0.001, SPF_NONE=0.001, T_SCC_BODY_TEXT_LINE=-0.01, URIBL_BLOCKED=0.001, URIBL_DBL_BLOCKED_OPENDNS=0.001, URIBL_ZEN_BLOCKED_OPENDNS=0.001] autolearn=no autolearn_force=no
Authentication-Results: ietfa.amsl.com (amavisd-new); dkim=pass (2048-bit key) header.d=smu-ac-kr.20230601.gappssmtp.com
Received: from mail.ietf.org ([50.223.129.194]) by localhost (ietfa.amsl.com [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id mCebDnLPv5y7 for <detnet@ietfa.amsl.com>; Thu, 21 Sep 2023 19:50:08 -0700 (PDT)
Received: from mail-ua1-x936.google.com (mail-ua1-x936.google.com [IPv6:2607:f8b0:4864:20::936]) (using TLSv1.3 with cipher TLS_AES_128_GCM_SHA256 (128/128 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits) server-digest SHA256) (No client certificate requested) by ietfa.amsl.com (Postfix) with ESMTPS id AE9C1C1524DC for <detnet@ietf.org>; Thu, 21 Sep 2023 19:50:08 -0700 (PDT)
Received: by mail-ua1-x936.google.com with SMTP id a1e0cc1a2514c-7a8b839fc0aso578543241.3 for <detnet@ietf.org>; Thu, 21 Sep 2023 19:50:08 -0700 (PDT)
DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=smu-ac-kr.20230601.gappssmtp.com; s=20230601; t=1695351007; x=1695955807; darn=ietf.org; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:from:to:cc:subject:date:message-id:reply-to; bh=vWTUcZkaqhjJJOmqkg8Fv9vEl52P9ghu+uENB1gR0P0=; b=UvJUAMC7WKTtJkw/4c+he1RvXELVt67ielJuhHr8CWhuNqrImRWpnpBk0v8GTcImV5 U/Xi73lBU5+itOFw5TU1rtK0XTWmGsSuBgbO+ec1fyPN144eSQ3vpT0CIdyGQVquj/x6 10JkIRtyBKXYHXsbR/LG6JUdm+n/9HxVJ85ZNo9xZXi6KvTJXlvUAIjAyG7k0p9HQEhr VeieRDXI2mXU4nJvYtyUshKoWccSJ/Yf0BdTvsEJBE13E9VXrVqoFj3TbNGsPQxdC5Lm il7E1XInBrAtlo/aXTk7T7qDQ2EUc7xiLKKMK0ztnZAZWSWxJvlJH3qbx6dkj5AFTCv6 tMaA==
X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1695351007; x=1695955807; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=vWTUcZkaqhjJJOmqkg8Fv9vEl52P9ghu+uENB1gR0P0=; b=RICECTPrw34H+tXVedYUDn+f5BKNa9WFgRu8TCYYVH51qC8TwUeiWmfPjlC9ztJ3oU fjo+cQX1gZHFe8vrTFOS2zoGJJp+P47Nbp1VQnbtUB1HBxaukUuQv/yNZvE5d6lZ6o1B MqDAw0ZdB74KqrX0hfqXhwnovb94d6z770/bPFrqOiWb7a3j/mK0YXbhzwXk/8UXLN3J ov9tE6BFhubKwaA43Ez4uPG6imwVRwgOQOxKLc0Vlz+e8rqY7aOjgzW2OfzDU2GFR80q fOvUCFnpmzh782PCV14K2+I3nJsZo/MUPAY++OMRVpBt3AGZMh/OXW4bzTyIRN5wAKCr 8QmQ==
X-Gm-Message-State: AOJu0Yw0TQBF0KhgtSpcN9P6WlAqiIo1Ze8dod9EHvcrYzKLLvr60ZkI IbyKo5Qvh5yCLL7VnnQ5geUlEQVqRP+3W7RTE521gA==
X-Google-Smtp-Source: AGHT+IG2fxpmxYifRyIPn0RPVzMrRVJ/ZLy4dVIf+5JIuGcKrWFDwuA5BWShwW1w/T1SFeaT7GJUxjy6g3b3OVFzZlQ=
X-Received: by 2002:a67:fc16:0:b0:44d:5435:a3e with SMTP id o22-20020a67fc16000000b0044d54350a3emr6507295vsq.9.1695351006772; Thu, 21 Sep 2023 19:50:06 -0700 (PDT)
MIME-Version: 1.0
References: <CA+8ZkcRTMgSifH6LOyTC6C04UcFn_zsOwrSuH-YxGbaqLcDWeg@mail.gmail.com> <202309211202010690585@zte.com.cn> <CA+8ZkcQA=ETPpjDH0piyv_cOp8jHXBV+EymAyjyr025yxjcy5w@mail.gmail.com> <ZQx4MwcoJBs+42OM@faui48e.informatik.uni-erlangen.de>
In-Reply-To: <ZQx4MwcoJBs+42OM@faui48e.informatik.uni-erlangen.de>
From: Jinoo Joung <jjoung@smu.ac.kr>
Date: Fri, 22 Sep 2023 11:49:58 +0900
Message-ID: <CA+8ZkcTbd2XCS9avwQeefpVr0h+qLgKZfB1f0Rhe+GgN-AD10Q@mail.gmail.com>
To: Toerless Eckert <tte@cs.fau.de>
Cc: peng.shaofu@zte.com.cn, detnet@ietf.org, draft-eckert-detnet-glbf@ietf.org
Content-Type: multipart/alternative; boundary="00000000000087556d0605e9aa7e"
Archived-At: <https://mailarchive.ietf.org/arch/msg/detnet/3ud4yQHTFAYyLVyINF_PWVDsaR8>
Subject: Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt
X-BeenThere: detnet@ietf.org
X-Mailman-Version: 2.1.39
Precedence: list
List-Id: Discussions on Deterministic Networking BoF and Proposed WG <detnet.ietf.org>
List-Unsubscribe: <https://www.ietf.org/mailman/options/detnet>, <mailto:detnet-request@ietf.org?subject=unsubscribe>
List-Archive: <https://mailarchive.ietf.org/arch/browse/detnet/>
List-Post: <mailto:detnet@ietf.org>
List-Help: <mailto:detnet-request@ietf.org?subject=help>
List-Subscribe: <https://www.ietf.org/mailman/listinfo/detnet>, <mailto:detnet-request@ietf.org?subject=subscribe>
X-List-Received-Date: Fri, 22 Sep 2023 02:50:13 -0000
Toerless, I will repeat what I have stated before: 1) Yes, burst accumulates over hops (or TSpec distortion, as you said); for Fair Queuing, C-SCORE, DRR, or FIFO. But the amount of accumulations are very different. 2) The E2E latency bound of FIFO is affected by these accumulated bursts. 3) However, the E2E latency bound of FQ, C-SCORE, or DRR is NOT affected. Let me show you again the E2E latency bound of C-SCORE. (b-M)/R + H*(M/R + Lmax/LC), b the initial burst, M the Maximum packet length, R the service rate of the flow, and Lmax the max packet length of the link, LC the Link Capacity. Why is it not affected? It has a similar reason as a flow in a TDM scheduler is not affected by the other flows' burst sizes. Why is a TDM not affected by others' bursts? If you still don't get it, or have doubts on RFC2212 or the expression above, I strongly recommend you read and understand the related papers from the 90s. Best, Jinoo On Fri, Sep 22, 2023 at 2:07 AM Toerless Eckert <tte@cs.fau.de> wrote: > Inline > > On Thu, Sep 21, 2023 at 07:23:24PM +0900, Jinoo Joung wrote: > > Hello Shaofu and Toerless, > > > > I think we are almost there, getting a consensus. > > Your comments are correct, and very close to each other. > > Your comments are about accumulation of burst, which can even explode in > > the FIFO cases. > > And i think my point is that the problem equally exists whatever the > cheduler is, because the worst cases can equally be created with > other schedulers as i tried to exemplify. > > > It is more prone to explode when there are loops in the topology. > > Such as rings in aggregastion networks, yes. But overall, the issue is that > we need a simple per-hop accumulative calculus for large scale networks > where > cahnges in allocation/use of one hop do not impact the othrer. > > > One thing to clarify with Toerless is: TSpec is given as rt+b. > > No matter how large b is, if the flow conforms, then the > > sustainable arrival rate is r. > > So your following statement is somewhat misleading. > > > > "So its easy to imagine that on each > > interfce we will receive a packet at the same time, and this creates an > > instantaneous burst of packets whose rate is larger than the sum of the > > admitted rates of those flows." > > ;-) Yes, precision in statements is important. so we have i=1..N | a_i(t) > <= r_i*t+b_i. > And my point was that sum(i=1...N|a_i(t))) >> sum(r_i*t) because the > period t we're observing may include all the b_i. > > > I have shown you in the previous mail the following ROUGH E2E latency > bound > > expression for a FIFO network. > > > > (b-M)/R + H*(M/R + Sum of bursts/Link Capacity). > > > > To be more precise, it is more like > > > > (b-M)/R + (M/R + Sum of bursts from other flows at node 1/LC of node 1) > > + (M/R + Sum of bursts from other flows at node 2/LC of node 2) + and so > on. > > > > As you two have mentioned, the flows' bursts are accumulating and getting > > larger and larger. > > This is the critical shortcoming of FIFO networks. > > The point in my last email was not even looking at many hops, but just the > fact that for a single hop we do get TSpec distortion of single flows, > whatever > the scheduler is, whenever for example we have 2 or more input interfaces > and same interface speed (or in general an aggregate higher arrival rate > than sending rate). And without shaper on the following hop, these > distortions will lead to an inability to have a simple per-hoip calculus, > because some flows will get a higher TSpec. > > Please review/comment if you think this can not happen for C-SCORE (or > rfc2212): > > 4 sets of 100 flows. Each flow r_i = 1 Mbps, b = 1500*8 bit. > > 4 Gbps links. each link carries 100 such flows. Each link > can habe a worst-case unfortunate timing that the 100 * 1500 byte packets > of the 100 flows it carries are back-to-back for packet p_i-1. and assume > all flows > where silent in before. And for the one flow whose packet ends up last in > the > outgoing interface queue, p_i follows at the regular interval, and all > other > flows are silent for their p_i. > > router has those 4 links on input, and 1 Gbps on output. So we end up > having to send 400 packets arriving with an aggregate rate of 4Gbps out > to the one Gbps link. in result, that last packet in the burst p_i-1 will > be delayed quite a bit, and therefore be closer to p_i and hence i has now > a higher TSpec on the following hop which (in my book) needs to be > corrected > on the next-hop with a shaper (or damper). > > > And we also know how much the burst increases in a node. > > Please take a look at the seminal work of Stiliadis > > https://ieeexplore.ieee.org/document/731196. > > Sure. > > > We can calculate whether the FIFO network will explode. > > If it explodes, the latency bound does not exist, and PBOO becomes > > meaningless, of course. > > Lets stick to a single hop an whether the TSpec for rfc2212 and C-SCORE > on the next-hop will not be potentially increased such as in my example. > > > So shaping at every node is effective, especially in the FIFO networks. > > In ATS, if it's just the IR + FIFO, E2E latency bound is > > > > (b-M)/R + H*(M/R + Sum of INITIAL bursts of other flows/LC), > > > > given that every node's link capacity is identical. > > And PBOO holds. > > Again. For my example, how would rfc2212 or C-SCORE be any different for > the increase in TSpec after a contention burst on the prior hop. > > Cheers > Toerless > > > Best, > > Jinoo > > > > > > > > > > > > > > On Thu, Sep 21, 2023 at 1:03 PM <peng.shaofu@zte.com.cn> wrote: > > > > > > > > Hi Jinoo, > > > > > > > > > Thanks for your explanation. > > > > > > > > > I agree with you that for a single observed flow, especialy with > > > ideal fluid model, > > > > > > each scheduler in the network can provide a guaranteed service for it, > > > even in the case of > > > > > > flow aggregation. > > > > > > > > > For example, multiple flows, each with arrival curve A_i(t) = b_i + > r_i * > > > t, may belongs to the > > > > > > same traffic class and consume the resources (burst and bandwidth) of > the > > > same > > > > > > out-interface on the intermediate node. Suppose that the scheduler > provide > > > a rate-latency > > > > > > service curve R*(t - T) for that traffic class, where R >= sum(r_i). > Then, > > > if each flow arrived > > > > > > idealy, i.e., complied its own arrival curve, for the oberserved flow, > it > > > will be ensured to get a > > > > > > guaranteed service rate R' from the total service rate R. > > > > > > > > > Suppose that on each node the guaranteed service curve for the observed > > > flow (e.g, flow 1) > > > > > > is R'*(t - T') , then, according to "Pay Bursts Only Once" rule, the > E2E > > > latency may be: > > > > > > T' * hops + b_1/R'. It seems that E2E latency just consider the burst > of > > > flow 1 (i.e., b_1) and > > > > > > only once, and never consider other flows.. > > > > > > > > > However, the truth is hidden in T'. > > > > > > > > > According to section 6 "Aggregate Scheduling" discussion in netcal > book, the > > > guaranteed > > > > > > service curve for flow 1, i.e., R'*(t - T'), may be deduced by: > > > > > > R*(t - T) - sum(b_2 + r_2*t, b_3 + r_3*t, ..., b_n + r_n*t) > > > > > > = (R - r_2 - r_3 -...- r_n) * (t - ((b_2 + b_3 + ... + b_n + R*T) / > (R - > > > r_2 - r_3 -...- r_n))) > > > > > > > > > Thus, R' equals to (R - r_2 - r_3 -...- r_n), > > > > > > and T' equals to ((b_2 + b_3 + ... + b_n + R*T) / (R - r_2 - r_3 -...- > > > r_n)). > > > > > > It can be seen that all bursts of other flows contribute the result of > T', > > > on each node, > > > > > > again and again. > > > > > > > > > If we take each b_i as 1/n of B, r_i as 1/n of R, T as 0, to simply > > > compare the above PBOO > > > > > > based latency estiamation and traffic class based latency estimation, > we > > > may find that > > > > > > the former is b_1/r_1 + hops*(n-1)*b_1/r_1, while the later is > b_1/r_1. It > > > is amazing that > > > > > > the former is about n times the latter . > > > > > > Please correct me if I misunderstand. > > > > > > > > > Regards, > > > > > > PSF > > > > > > > > > > > > Original > > > *From: *JinooJoung <jjoung@smu.ac.kr> > > > *To: *彭少富10053815; > > > *Cc: *Toerless Eckert <tte@cs.fau.de>;DetNet WG <detnet@ietf.org>; > > > draft-eckert-detnet-glbf@ietf.org <draft-eckert-detnet-glbf@ietf.org>; > > > *Date: *2023年09月19日 16:13 > > > *Subject: **Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt* > > > > > > Hello Shaofu. > > > > > > Thanks for the most relevant question, > > > which, as you have said, is related to Toerless' last question > regarding > > > "reshaping on merge points". > > > > > > If I may repeat Toerless' question below: > > > > > > "If you are relying on the same math as what rfc2212 claims, just > replacing > > > stateful WFQ with stateless, then it seems to me that you would equally > > > need > > > the shapers demanded by rfc2212 on merge-points. I do not see them in > > > C-SCORE." > > > > > > And your question is: > > > > > > " "Pay Bursts Only Once" may be only applied in the case that the > network > > > provides > > > a dedicated service rate to a flow. Our network naturally aggregates > flows > > > at every node, > > > therefore does not dedicate a service rate to a flow, and PBOO does not > > > apply." > > > My answer is: > > > > > > "Pay Bursts Only Once" is applied in the case that the network provides > > > a GUARANTEED service rate to a flow. > > > Fair queuing, C-SCORE, or even FIFO scheduler can guarantee a service > rate > > > to a flow. > > > As long as a flow, as a single entity, is guaranteed a service rate, > it is > > > not considered aggregated or merged. > > > Therefore reshaping is not necessary, and PBOO holds. > > > > > > Below is my long answer. Take a look if you'd like. > > > > > > Fair queuing, Deficit round robin, or even FIFO scheduler guarantees a > > > service rate to a flow, > > > if the total arrival rate is less than the link capacity. > > > The only caveat is that FIFO can only guarantee the service rate that > is > > > equal to the arrival rate, > > > while FQ and DRR can adjust the service rate to be larger than the > arrival > > > rate. > > > If such rate-guaranteeing schedulers are placed in a network, > > > then a flow is guaranteed to be served with a certain service rate, > > > and not considered to be "aggregated", in the intermediate nodes. > > > > > > RFC2212, page 9~10, in subsection "Policing", states the followings: > > > > > > "Reshaping is done at all heterogeneous source branch points and at all > > > source merge points." > > > > > > "Reshaping need only be done if ..." > > > > > > "A heterogeneous source branch point is a spot where the > > > multicast distribution tree from a source branches to multiple > > > distinct paths, and the TSpec’s of the reservations on the various > > > outgoing links are not all the same." > > > > > > "A source merge point is where the distribution paths or trees from two > > > different sources > > > (sharing the same reservation) merge." > > > > > > In short, RFC2212 states that reshaping CAN be necessary at the flow > > > aggregation and deaggregation points. > > > > > > Flow aggregation and deaggregation are something happening usually at > the > > > network boundary, between networks, etc, > > > with careful planning. Flow multiplexing into a FIFO is not considered > an > > > aggregation. > > > > > > Best, > > > Jinoo > > > > > > > > > On Tue, Sep 19, 2023 at 10:57 AM <peng.shaofu@zte.com.cn> wrote: > > > > > >> > > >> Hi Jinoo, Toerless, > > >> > > >> > > >> Sorry to interrupt your discussion. > > >> > > >> > > >> According to NetCal book ( > > >> https://leboudec.github.io/netcal/latex/netCalBook.pdf), > > >> > > >> "Pay Bursts Only Once" may be only applied in the case that the > network > > >> provide > > >> > > >> a dedicate service rate (may be protected by a dedicate queue, or > even a > > >> dedicate > > >> > > >> sub-link) for the observed flow, such as the guarantee service > defined in > > >> RFC2212. > > >> > > >> In brief, there is no other flows sharing service rate with the > observed > > >> flow. That is, > > >> > > >> there is no fan-in, no so-called " competition flows belongs to the > same > > >> traffic class > > >> > > >> at the intermediate node". > > >> > > >> > > >> Traffic class is often used to identify flow aggregation, which is not > > >> the case that > > >> > > >> "Pay Bursts Only Once" may be applied. It seems that in IP/MPLS > network, > > >> flow > > >> > > >> aggregation is naturely. The picture Toerless showed in the previous > mail > > >> is exactly > > >> > > >> related with flow aggregation, i.e, the observed flow may be > interfered > > >> by some > > >> > > >> competition flows belongs to the same traffic class at node A and > again > > >> interfered by > > >> > > >> other competition flows belongs to the same traffic class at node B > > >> separately. > > >> > > >> > > >> Please correct me if I misunderstood. > > >> > > >> > > >> Regards, > > >> > > >> PSF > > >> > > >> > > >> > > >> _______________________________________________ > > >> detnet mailing list > > >> detnet@ietf.org > > >> https://www.ietf.org/mailman/listinfo/detnet > > >> > > >> Toerless, > > >> It seems that you argue two things in the last email. > > >> > > >> 1) Your first argument: The E2E latency is the sum of per-hop > latencies. > > >> > > >> You are half-right. > > >> According to RFC2212, page 3, the generic expression for the E2E > latency > > >> bound of a flow is: > > >> > > >> [(b-M)/R*(p-R)/(p-r)] + (M+Ctot)/R + Dtot. (1) > > >> > > >> Let's call this expression (1). > > >> > > >> Here, b is the max burst, M the max packet length, R the service > rate, p > > >> the peak rate, r the arrival rate; > > >> and Ctot and Dtot are the sums of C and D, which are so called "error > > >> terms", over the hops. > > >> Thus, the E2E latency bound can be a linear function of the hop count, > > >> since Ctot and Dtot are functions of hop count. > > >> However, the first term [(b-M)/R*(p-R)/(p-r)], which includes b, is > not. > > >> So you can see the E2E latency is NOT just the sum of per-hop > latencies. > > >> > > >> 2) Your second argument: C-SCORE cannot be free from burst > accumulation > > >> and other flows' burst. > > >> My short answer: It is free from burst accumulation or other flows' > burst. > > >> > > >> Imagine an ideal system, in which your flow is completely isolated. > > >> It is ALONE in the network, whose link has the rate R in every hop. > > >> No other flows at all. > > >> > > >> Assume the flow's arrival process is indeed rt+b. At time 0, b arrives > > >> instantly. > > >> (This is the worst arrival, such that it makes the worst latency.) > > >> Then your flow experiences the worst latency b/R at the first node, > > >> and M/R (transmission delay of a packet) at the subsequent nodes. > > >> > > >> I.e. (b-M)/R + H*M/R, where H is the hop count. > > >> > > >> This is a special case of (1), where R is the same with r, C is M, > and D > > >> is zero. > > >> > > >> Fair queuing and C-SCORE are the best approximations of such an ideal > > >> system, with D equals Lmax/LC, > > >> where Lmax and LC are the max packet length in the link and the > capacity > > >> of the link, respectively. > > >> Therefore the E2E latency bound of C-SCORE is > > >> > > >> (b-M)/R + H*(M/R + Lmax/LC), > > >> > > >> which is again another special case of (1). > > >> Note that b is divided by R, not LC. > > >> > > >> It is well known that a FIFO scheduler's D (error term) is a function > of > > >> the sum of other flows' bursts. > > >> Their E2E latency bounds are approximately > > >> (b-M)/R + H*[M/R + (Sum of Bursts)/LC]. > > >> > > >> ATS or UBS also relies on FIFO, but the bursts are suppressed as the > > >> initial values, > > >> therefore enjoys a much better E2E latency expression. > > >> > > >> Hope this helps. > > >> > > >> Best, > > >> Jinoo > > >> > > >> > > >> > > >> On Sun, Sep 17, 2023 at 1:42 AM Toerless Eckert <tte@cs.fau.de> > wrote: > > >> > > >>> On Thu, Sep 14, 2023 at 08:08:24PM +0900, Jinoo Joung wrote: > > >>> > Toerless, thanks for the reply. > > >>> > In short: > > >>> > > > >>> > C-SCORE's E2E latency bound is NOT affected by the sum of bursts, > but > > >>> only > > >>> > by the flow's own burst. > > >>> > > >>> +----+ > > >>> If1 --->| | > > >>> ... | R |x---> > > >>> if100 --->| | > > >>> +----+ > > >>> > > >>> If you have a router with for example 100 input interfaces, all > sending > > >>> packets > > >>> to the same ooutput interfaces, they all will have uncorrelated flows > > >>> arriving > > >>> from the different interfaces, and at least each interface can have a > > >>> packet > > >>> arriving t the same time in R. > > >>> > > >>> The basic caluculus of UBS is simply the most simple and hence > > >>> conservative, > > >>> assuming all flows packet can arrive from different interface without > > >>> rate limits. But of course you can do the latency calculations in a > > >>> tighter > > >>> fashion for UBS. WOuld be interesting to learn if IEEE for TSN-ATS > (Qcr) > > >>> was > > >>> looking into any of this. E.g.: applying line shaping for the > aggregate > > >>> of > > >>> flows from the same input interface and incorporating the service > curve > > >>> of > > >>> the node. > > >>> > > >>> In any case, whether the tighter latency calculus is you do for Qcr, > it > > >>> is equally applicable to gLBF. > > >>> > > >>> > It is bounded by (B-L)/r + H(L/r + Lmax/R), where B is the flow's > max > > >>> burst > > >>> > size. > > >>> > > >>> See above picture. How could the packet in question not also suffer > the > > >>> latency introduced by the other 99 packets randomnly arriving at > almos > > >>> the same time, just a tad earlier. > > >>> > > >>> > You can also see that B appears only once, and not multiplied by > the > > >>> hop > > >>> > count H. So the burst is paid only once. > > >>> > > >>> C-SCORE is a stateless fair queuing (FQ) mechanism, but making FQ > state, > > >>> does not change the fact that FQ itself does not eliminate the > > >>> burst accumulation on merge points such as shon in the picture. > rfc2212 > > >>> which for example also recommends FQ independently mandates the use > of > > >>> reshaping. > > >>> > > >>> > Please see inline marked with JJ. > > >>> > > >>> yes, more inline. > > >>> > > >>> > On Thu, Sep 14, 2023 at 3:34 AM Toerless Eckert <tte@cs.fau.de> > wrote: > > >>> [cutting off dead leafs] > > >>> > > >>> > > I was only talking about the E2E latency bound guaranteeable by > > >>> DetNet > > >>> > > Admission Control (AC). > > >>> > > > >>> > JJ: Admission control has two aspects. > > >>> > First, guaranteeing the total arrival rate (or allocated service > rate) > > >>> does > > >>> > not exceed the link capacity. > > >>> > Second, assuring a service level agreement (SLA), or a requested > > >>> latency > > >>> > bound (RSpec) to a flow. > > >>> > By "E2E latency bound guaranteeable by DetNet Admission Control > (AC)", > > >>> > I think you mean that negotiating the SLA first, and then, based > on the > > >>> > negotiated latency, allocating per-hop latency to a flow. > > >>> > I suggest this per-hop latency allocation, and enforcing that > value in > > >>> a > > >>> > node, is not a good idea, > > >>> > since it can harm the advantage of "pay burst only once (PBOO)" > > >>> property. > > >>> > > >>> Its the only way AFAIK to achieve scalability in number of hops and > > >>> number of > > >>> flows by being able to calculate latency as a linear composition of > > >>> per-hop > > >>> latencies. This is what rfc2212 does, this is what Qcr does, *CQF, > gLBF > > >>> and so on. > > >>> > > >>> > PBOO can be interpreted roughly as: If your burst is resolved at a > > >>> point in > > >>> > a network, then it is resolved and does not bother you anymore. > > >>> > However, in the process of resolving, the delay is inevitable. > > >>> > > >>> Remember tht the problemtic burstyness is an increase in the > intra-flow > > >>> burstyness resulting from merge-point based unexpected latency > > >>> to p-1 of the flow followed by no such delay for packet p of the same > > >>> flow, > > >>> and hence clustering p-1 and p closer together, making it exceed its > > >>> reserved burst size. > > >>> > > >>> This can not be compensated for in a work-conserving way by simply > > >>> capturing > > >>> the per-hop latency of each packet p-1 and p alone, but it requires a > > >>> shaper > > >>> (or UR) to get rid off. > > >>> > > >>> > JJ: Because you don't know exactly where the burst is resolved, > > >>> > when you calculate the per-node latency bound, the latency due to > the > > >>> burst > > >>> > has to be added as a portion of the latency bound. > > >>> > Thus the sum of per-node bounds is much larger than the E2E latency > > >>> bound > > >>> > calculated with seeing the network as a whole. > > >>> > > >>> Consider a flow passing through 10 hops. On every hop, you have > > >>> potentially > > >>> merge-point with new traffic flows and new burst collision. All that > we > > >>> do > > >>> in the simple UBS/Qcr calculus is to take the worst case into > account, > > >>> where worst case may not even be admitted now, but which could be > > >>> admitted > > >>> in the future, and at that point in time you do not want to go back > and > > >>> change the latency guarantee for your already admitted flow. > > >>> > > >>> Src2 Src3 Src4 Src5 Src6 Src7 Src8 Src9 Src10 > > >>> | | | | | | | | | > > >>> v v v v v v v v v > > >>> Src1 -> R1 -> R2 -> R3 -> R4 -> R5 -> R6 -> R7 -> R8 -> R9 -> R10 -> > > >>> Rcv1 > > >>> | | | | | | | | | > > >>> v v v v v v v v v > > >>> Rcv2 Rcv3 Rcv4 Rcv5 Rcv6 Rcv7 Rcv8 Rcv9 Rcv10 > > >>> > > >>> Above is example, where Flow 1 from Src1 to Rcv1 will experience such > > >>> merge-point burst accumulation issue on every hop - worst case. And > as > > >>> mentioned > > >>> before, yes, when you use the simple calculus, you're also > > >>> overcalculating > > >>> the per-hop latency for flows that e.g.: all run in parallel to > Src1, but > > >>> that is just a matter of using stricter network calculus. And because > > >>> Network Calculus is complex, and i didn't want to start becoming an > > >>> expert on > > >>> it, i simply built the stateless solution in a way where i can reuse > a > > >>> pre-existing, proven and used-in-standards (Qcr) queuing-model, > calculus. > > >>> > > >>> > > >>> > JJ: If you enforce resolving it at the network entrance by a strict > > >>> > regulation, then you may end up resolving while not actually > needed to. > > >>> > However, this approach is feasible. I will think more about it. > > >>> > > >>> I think the flow-interleaving is a much more fundamental issue of > higher > > >>> utilization with large number of flows all with low bitrates. Look at > > >>> the examples of the draft-eckert-detnet-flow-interleaving, and tell > me > > >>> how else but time-division-multiplexing one would be able to solve > this. > > >>> Forget the complex option where flows from diffeent ingres routers to > > >>> different egres routers are involved. Just the most simple problem of > > >>> one ingres router PE1, maybe 10 hops through thre network to a PE2, > and > > >>> 10,000 > > >>> acyclic flows going to thre same egress router PE2. > > >>> > > >>> Seems to me quite obvious tht you can as well resolve the burst on > > >>> ingress > > >>> PE1 instead of hoping, and getting complex math by trying to do this > > >>> on further hops along the path. > > >>> > > >>> Cheers > > >>> Toerless > > >>> > > >>> > > > > >>> > > > It can be a function of many factors, such as number of flows, > > >>> their > > >>> > > > service rates, their max bursts, etc. > > >>> > > > The sum of service rates is a deciding factor of utilization. > > >>> > > > > >>> > > Given how queuing latency always occurs from collision of > packets in > > >>> > > buffers, > > >>> > > the sum of burst sizes is the much bigger problem for DetNet than > > >>> service > > >>> > > rates. But this is a side discuss. > > >>> > > > > >>> > > > >>> > JJ: That is a good point. So we should avoid queuing schemes whose > > >>> latency > > >>> > bounds are affected by the sum of bursts. > > >>> > Fortunately, C-SCORE's E2E latency bound is NOT affected by the > sum of > > >>> > bursts, but only by the flow's own burst. > > >>> > It is bounded by (B-L)/r + H(L/r + Lmax/R), where B is the flow's > max > > >>> burst > > >>> > size. > > >>> > You can also see that B appears only once, and not multiplied by > the > > >>> hop > > >>> > count H. > > >>> > So the burst is paid only once. > > >>> > > > >>> > > > >>> > > > So, based on an E2E latency bound expression, you can guess the > > >>> bound at > > >>> > > > 100% utilization. > > >>> > > > But you can always fill the link with flows of any burst sizes, > > >>> therefore > > >>> > > > the guess can be wrong. > > >>> > > > > >>> > > "guess" is not a good work for DetNet. > > >>> > > > > >>> > > A DetNet bounded latency mechanisms needs a latency bound > expression > > >>> > > (calculus) > > >>> > > to be a guaranteeable (over)estimate of the bounded latency > > >>> independent of > > >>> > > what other competing traffic there may be in the future. Not a > > >>> "guess". > > >>> > > > > >>> > > > >>> > JJ: Right. We should not guess. We should be able to provide an > exact > > >>> > mathematical expression for latency bound. > > >>> > Because you argued in the previous mail that the latency bound > should > > >>> be > > >>> > obtained based on 100% utilization, > > >>> > I was trying to explain why that should not be done. > > >>> > > > >>> > > > >>> > > > Admission control, on the other hand, can be based on > assumption > > >>> of high > > >>> > > > utilization level, but not necessarily 100%. > > >>> > > > > > >>> > > > You do not assume 100% utilization when you slot-schedule, > don't > > >>> you? > > >>> > > > > >>> > > I don't understand what you mean with slot-schedule, can you > please > > >>> > > explain ? > > >>> > > > > >>> > > > >>> > JJ: Slot-scheduling, which is a common term in the research > community, > > >>> > is a mapping of flow into a slot (or cycle) in a slot (or cycle) > based > > >>> > queuing methods, such as CQF. > > >>> > When we say a schedulability, it usually means whether we can > allocate > > >>> > requesting flows into slots (cycles) with a preconfigured cycle's > > >>> length > > >>> > and number. > > >>> > > > >>> > > > >>> > > > So "incremental scheduling" is now popular. > > >>> > > > > >>> > > Not sure what you mean with that term. > > >>> > > > >>> > > > >>> > JJ: The incremental scheduling means, when a new flow wants to > join the > > >>> > network, then the network examines the schedulability of the flow, > > >>> > without altering the existing flows' schedule. > > >>> > > > >>> > > > >>> > > > > >>> > > > > >>> > > I am only thinking about "admitting" when it comes to bounded > > >>> end-to-end > > >>> > > latency, aka: action by the AC of the DetNet controller-plane, > and > > >>> > > yes, that needs to support "on-demand" (incremental?), aka: > whenever > > >>> > > a new flow wants to be admitted. > > >>> > > > > >>> > > > 2) In the example I gave, the two flows travel the same path, > thus > > >>> the > > >>> > > > second link's occupancy is identical to the first one. > > >>> > > > Thus the competency levels in two links are the same, contrary > to > > >>> your > > >>> > > > argument. > > >>> > > > > >>> > > I guss we started from different assumptions about details not > > >>> explicitly > > >>> > > mentioned. For example, i guess we both assume that the sources > > >>> connect > > >>> > > to the > > >>> > > first router with arbitrary high interfaces so that we could ever > > >>> get close > > >>> > > to 2B/R on the first interface. > > >>> > > > > >>> > > But then we differed in a core detail. here is my assumption for > > >>> > > the topology / admission control: > > >>> > > > > >>> > > Src3 to R4 > > >>> > > +----+ \ +----+ / +----+ > > >>> > > Src1 ---| | \-| |-/ | | > > >>> > > Src2 ---| R1 |x-----| R2 |x----| R3 | > > >>> > > | |. | |. | | > > >>> > > +----+. +----+. +----+ > > >>> > > | | > > >>> > > 2B buffer 2B buffer > > >>> > > R srv rate R srv rate > > >>> > > > > >>> > > Aka: in my case, i was assuming that there could be case where > > >>> > > the interface from R2 to R3 could have a 2B/R queue (and not > > >>> > > assuming further optimizations in calculus). E.g.: in some > > >>> > > other possible scenario, Src2 sends to R2, and Src3 and Src1 to > > >>> > > R3 for example. > > >>> > > > > >>> > > > >>> > JJ: You can come up with an example that your scheme works well. > > >>> > But that does not negate the counterexample I gave. > > >>> > > > >>> > JJ: Again, there are only two flows. > > >>> > And, B is not the buffer size. B is the max burst size of a flow. > > >>> > R is the link capacity. > > >>> > Please review carefully the example I gave. > > >>> > > > >>> > > > >>> > > > > >>> > > You must have assumed that the totality of the DetNet admission > > >>> control > > >>> > > relevant topology is this: > > >>> > > > > >>> > > +----+ +----+ +----+ > > >>> > > Src1 ---| | | | | | > > >>> > > Src2 ---| R1 |x-----| R2 |x----| R3 | > > >>> > > | |. | |. | | > > >>> > > +----+. +----+. +----+ > > >>> > > | | > > >>> > > 2B buffer 2B buffer > > >>> > > R srv rate R srv rate > > >>> > > > > >>> > > Aka: DetNet admission control would have to be able to predict > that > > >>> > > under no permitted admission scenario, R2 would build a DetNet > queue, > > >>> > > so even when Src1 shows up as the first and only flow, the > admission > > >>> > > control could permit a latency to R3 of 2B/R - only for the > maximum > > >>> > > delay through R1 queue and 0 for R2 queue. > > >>> > > > > >>> > > But if this is the whole network and the admission control logic > > >>> > > can come to this conclusion, then of course it could equally do > the > > >>> > > optimization and not enable gLBF Dampening on R2 output > > >>> > > interface, or e.g.: set MAX=0 or the like. An e'voila, gLBF > > >>> > > would also give 2B/R - but as said, i think it's not a deployment > > >>> > > relevant example. > > >>> > > > >>> > > > >>> > JJ: If you can revise and advance the gLBF, that would be great. > > >>> > I am willing to join that effort, if you would like to. > > >>> > > > >>> > > > >>> > > > > >>> > > > > >>> > > Cheers > > >>> > > Toerless > > >>> > > > > >>> > > > Please see inline with JJ. > > >>> > > > > > >>> > > > Best, > > >>> > > > Jinoo > > >>> > > > > > >>> > > > On Wed, Sep 13, 2023 at 9:06 AM Toerless Eckert <tte@cs.fau.de > > > > >>> wrote: > > >>> > > > > > >>> > > > > On Fri, Jul 21, 2023 at 08:47:18PM +0900, Jinoo Joung wrote: > > >>> > > > > > Shaofu, thanks for the reply. > > >>> > > > > > It is my pleasure to discuss issues like this with you. > > >>> > > > > > > > >>> > > > > > The example network I gave is a simple one, but the > scenario > > >>> is the > > >>> > > worst > > >>> > > > > > that can happen. > > >>> > > > > > The E2E latency bounds are thus, > > >>> > > > > > > > >>> > > > > > for Case 1: ~ 2B/R > > >>> > > > > > for Case 2: ~ 2 * (2B/R) > > >>> > > > > > > >>> > > > > This is a bit terse, let me try to expand: > > >>> > > > > > > >>> > > > > Case 1 is FIFO or UBS/ATS, right / Case 2 is gLBF ? > > >>> > > > > > > >>> > > > > > >>> > > > JJ: Correct. > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > Assuming i am interpreting it right, then this is > inconsistent > > >>> with > > >>> > > your > > >>> > > > > setup: You said all links are the same so both hops do have > the > > >>> same > > >>> > > > > buffer and rates, so the admission controller also expect to > > >>> > > > > have to put as many flows on second link/queue that it fills > up > > >>> 2B. > > >>> > > > > > > >>> > > > > > >>> > > > JJ: Yes, two links are identical. However, as I have mentioned, > > >>> > > > an E2E latency bound is calculated based on a given network > > >>> environment. > > >>> > > > We don't always consider a filled up link capacity. > > >>> > > > BTW, B is the max burst size of the flow. > > >>> > > > > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > You just then made an example, where there was never such an > > >>> amount > > >>> > > > > of competing traffic on the second hop. But tht does not mean > > >>> that > > >>> > > > > the admission controller could guarantee in UBS/ATS would > have > > >>> > > > > less per-hop latency than 2B/R. > > >>> > > > > > >>> > > > > > >>> > > > JJ: Again, two links are identical and two flows travel both > links. > > >>> > > > The difference between Case 1 and Case 2 is not because of the > > >>> different > > >>> > > > competition level (they are identical.) > > >>> > > > but because of the non-work conserving behaviour of the second > > >>> link in > > >>> > > Case > > >>> > > > 2. > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > If the admission controller knew there would never be a > queue on > > >>> the > > >>> > > > > second hop, then gLBF likewise would not need to do a Damper > on > > >>> the > > >>> > > > > second hop. Hence as i said previously, the per-hop and > > >>> end-to-end > > >>> > > > > bounded latency guarantee is the same between UBS and gLBF. > > >>> > > > > > > >>> > > > > > And again, these are the WORST E2E latencies that a packet > can > > >>> > > experience > > >>> > > > > > in the two-hop network in the scenario. > > >>> > > > > > > >>> > > > > Its not the worst case latency for the UBS case. you just did > > >>> not have > > >>> > > > > an example to create the worst case amount of competing > traffic. > > >>> Or you > > >>> > > > > overestimed the amount of buffering and hence per-hop latency > > >>> for the > > >>> > > > > UBS/ATS casee. > > >>> > > > > > > >>> > > > > > In any network that is more complex, the E2E latency > bounds of > > >>> two > > >>> > > > > schemes > > >>> > > > > > are very different. > > >>> > > > > > > >>> > > > > Counterexample: > > >>> > > > > > > >>> > > > > You have a network with TSN-ATS. You have an admission > > >>> controller. > > >>> > > > > You only have one priority for simplicity of example. > > >>> > > > > > > >>> > > > > You do not want to dymamically signal changed end-to-end > > >>> latencies > > >>> > > > > to applications... because its difficult. So you need to plan > > >>> > > > > for worst-case bounded latencies under maximum amount of > traffic > > >>> > > > > load. In a simple case this means you give each interface > > >>> > > > > a queue size B(i)/r = 10usec. Whenever a new flow needs to be > > >>> > > > > added to the network, you find a path where all the buffers > > >>> > > > > have enough space for your new flows burst and you signal > > >>> > > > > to the application that the end-t-end guaranteed latency is > > >>> > > > > P(path)+N*10usec, where P is the physical propagation latecy > of > > >>> > > > > the path and N is the number of hops it has. > > >>> > > > > > > >>> > > > > And in result, all packets from the flow will arrive with > > >>> > > > > a latency between P(path)...P(path)+N*10usec - depending > > >>> > > > > on network load/weather. > > >>> > > > > > > >>> > > > > Now we replace UBS in the routers with gLBF. What changes ? > > >>> > > > > > > >>> > > > > 1) With UBS the controller still had to signal every new and > > >>> > > > > to-be-deleted flow to every router along it path to set up > the > > >>> > > > > IR for the flow. This goes away (big win). > > >>> > > > > > > >>> > > > > 2) The forwarding is in our opinion cheaper/faster to > implement > > >>> > > > > (because of lack of memory read/write cycle of IR). > > >>> > > > > > > >>> > > > > 3) The application now sees all packets arrive at fixed > latency > > >>> > > > > of P(path)+N*10usec. Which arguably to the application that > > >>> > > > > MUST have bounded latency is from all examples i know > > >>> > > > > seen rather as a benefit than as a downside. > > >>> > > > > > > >>> > > > > Cheers > > >>> > > > > Toerless > > >>> > > > > > > >>> > > > > > > >>> > > > > > > > >>> > > > > > Best, > > >>> > > > > > Jinoo > > >>> > > > > > > > >>> > > > > > On Fri, Jul 21, 2023 at 8:31 PM <peng.shaofu@zte.com.cn> > > >>> wrote: > > >>> > > > > > > > >>> > > > > > > > > >>> > > > > > > Hi Jinoo, > > >>> > > > > > > > > >>> > > > > > > > > >>> > > > > > > I tried to reply briefly. If Toerless have free time, can > > >>> confirm > > >>> > > it. > > >>> > > > > > > > > >>> > > > > > > > > >>> > > > > > > Here, when we said latency bound formula, it refers to > > >>> worst-case > > >>> > > > > latency. > > >>> > > > > > > > > >>> > > > > > > > > >>> > > > > > > Intuitively, the worst-case latency for gLBF (damper + > > >>> shaper + > > >>> > > > > scheduler) > > >>> > > > > > > > > >>> > > > > > > is that: > > >>> > > > > > > > > >>> > > > > > > damping delay per hop is always 0. (because > scheduling > > >>> delay = > > >>> > > MAX) > > >>> > > > > > > > > >>> > > > > > > shaping delay is always 0. (because all are > eligibility > > >>> > > arrivals) > > >>> > > > > > > > > >>> > > > > > > scheduling delay is always MAX (i.e., concurent full > > >>> burst > > >>> > > from all > > >>> > > > > > > eligibility arrivals on each hop) > > >>> > > > > > > > > >>> > > > > > > > > >>> > > > > > > Similarly, the worst-case latency for UBS (shaper + > > >>> scheduler) is > > >>> > > that: > > >>> > > > > > > > > >>> > > > > > > shaping delay is always 0. (because all are > eligibility > > >>> > > arrivals) > > >>> > > > > > > > > >>> > > > > > > scheduling delay is always MAX (i.e., concurent full > > >>> burst > > >>> > > from all > > >>> > > > > > > eligibility arrivals on each hop) > > >>> > > > > > > > > >>> > > > > > > > > >>> > > > > > > Thus, the worst-case latency of gLBF and UBS is the same. > > >>> > > > > > > > > >>> > > > > > > Your example give a minimumal latency that may be > > >>> expierenced by > > >>> > > UBS, > > >>> > > > > but > > >>> > > > > > > > > >>> > > > > > > it is not the worst-case latency. In fact, your example > is a > > >>> simple > > >>> > > > > > > topology that only > > >>> > > > > > > > > >>> > > > > > > contains a line without fan-in, that cause scheduling > delay > > >>> almost > > >>> > > a > > >>> > > > > > > minimumal > > >>> > > > > > > > > >>> > > > > > > value due to no interfering flows. > > >>> > > > > > > > > >>> > > > > > > > > >>> > > > > > > Regards, > > >>> > > > > > > > > >>> > > > > > > PSF > > >>> > > > > > > > > >>> > > > > > > > > >>> > > > > > > > > >>> > > > > > > Original > > >>> > > > > > > *From: *JinooJoung <jjoung@smu.ac.kr> > > >>> > > > > > > *To: *彭少富10053815; > > >>> > > > > > > *Cc: *tte@cs.fau.de <tte@cs.fau.de>;detnet@ietf.org < > > >>> > > detnet@ietf.org>; > > >>> > > > > > > draft-eckert-detnet-glbf@ietf.org < > > >>> > > draft-eckert-detnet-glbf@ietf.org>; > > >>> > > > > > > *Date: *2023年07月21日 14:10 > > >>> > > > > > > *Subject: **Re: [Detnet] FYI: > > >>> draft-eckert-detnet-glbf-01.txt* > > >>> > > > > > > _______________________________________________ > > >>> > > > > > > detnet mailing list > > >>> > > > > > > detnet@ietf.org > > >>> > > > > > > https://www.ietf.org/mailman/listinfo/detnet > > >>> > > > > > > > > >>> > > > > > > Hello Toerless, > > >>> > > > > > > I have a comment on your argument. > > >>> > > > > > > This is not a question, so you don't have to answer. > > >>> > > > > > > > > >>> > > > > > > You argued that the gLBF + SP has the same latency bound > > >>> formula > > >>> > > with > > >>> > > > > UBS > > >>> > > > > > > (equivalently ATS IR + SP). > > >>> > > > > > > The IR is not a generalized gLBF, so they do not have the > > >>> same > > >>> > > bound. > > >>> > > > > > > > > >>> > > > > > > In short, ATS IR is a rate-based shaper so it enjoys "Pay > > >>> burst > > >>> > > only > > >>> > > > > once" > > >>> > > > > > > property. > > >>> > > > > > > gLBF is not. So it pays burst every node. > > >>> > > > > > > > > >>> > > > > > > Consider a simplest example, where there are only two > > >>> identical > > >>> > > flows > > >>> > > > > > > travelling the same path. > > >>> > > > > > > Every node and link in the path are identical. > > >>> > > > > > > > > >>> > > > > > > Case 1: Just FIFO > > >>> > > > > > > Case 2: gLBF + FIFO > > >>> > > > > > > > > >>> > > > > > > In the first node, two flows' max bursts arrive almost at > > >>> the same > > >>> > > time > > >>> > > > > > > but your flow is just a little late. > > >>> > > > > > > Then your last packet in the burst (packet of interest, > POI) > > >>> > > suffers > > >>> > > > > > > latency around 2B/R, where B is the burst size and R is > the > > >>> link > > >>> > > > > capacity. > > >>> > > > > > > This is true for both cases. > > >>> > > > > > > > > >>> > > > > > > In the next node: > > >>> > > > > > > In Case 1, the POI does not see any packet queued. so it > is > > >>> > > delayed by > > >>> > > > > its > > >>> > > > > > > own transmission delay. > > >>> > > > > > > In Case 2, the burst from the other flow, as well as > your own > > >>> > > burst, > > >>> > > > > > > awaits the POI. So the POI is again delayed around 2B/R. > > >>> > > > > > > > > >>> > > > > > > In the case of UBS, the max bursts are legitimate, so the > > >>> regulator > > >>> > > > > does > > >>> > > > > > > not do anything, > > >>> > > > > > > and the forwarding behavior is identical to Case 1. > > >>> > > > > > > > > >>> > > > > > > Best, > > >>> > > > > > > Jinoo > > >>> > > > > > > > > >>> > > > > > > On Fri, Jul 21, 2023 at 10:58 AM <peng.shaofu@zte.com.cn > > > > >>> wrote: > > >>> > > > > > > > > >>> > > > > > >> > > >>> > > > > > >> Hi Toerless, > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> Thanks for your response, and understand your busy > > >>> situation. > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> A quick reply is that gLBF is really an interested > proposal, > > >>> > > which is > > >>> > > > > > >> very > > >>> > > > > > >> > > >>> > > > > > >> similar to the function of Deadline on-time per hop. > Our > > >>> views > > >>> > > are > > >>> > > > > > >> > > >>> > > > > > >> consistent on this point. The key beneficial is to avoid > > >>> burst > > >>> > > > > cumulation. > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> The following example originated from the analysis of > > >>> deadline > > >>> > > on-time > > >>> > > > > > >> > > >>> > > > > > >> mode. I believe it also makes sense for gLBF. When you > have > > >>> free > > >>> > > time, > > >>> > > > > > >> > > >>> > > > > > >> may verify it. The result may be helpful both for gLBF > and > > >>> > > deadline > > >>> > > > > > >> > > >>> > > > > > >> on-time mode. Note that I didn't question mathematical > > >>> proof > > >>> > > about > > >>> > > > > UBS, > > >>> > > > > > >> > > >>> > > > > > >> which get the worst-case latency based on the > combination of > > >>> > > > > > >> > > >>> > > > > > >> "IR shaper + SP scheduler". > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> Regards, > > >>> > > > > > >> > > >>> > > > > > >> PSF > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> Original > > >>> > > > > > >> *From: *ToerlessEckert <tte@cs.fau.de> > > >>> > > > > > >> *To: *彭少富10053815; > > >>> > > > > > >> *Cc: *jjoung@smu.ac.kr <jjoung@smu.ac.kr>; > detnet@ietf.org < > > >>> > > > > > >> detnet@ietf.org>;draft-eckert-detnet-glbf@ietf.org < > > >>> > > > > > >> draft-eckert-detnet-glbf@ietf.org>; > > >>> > > > > > >> *Date: *2023年07月21日 06:07 > > >>> > > > > > >> *Subject: **Re: [Detnet] FYI: > > >>> draft-eckert-detnet-glbf-01.txt* > > >>> > > > > > >> > > >>> > > > > > >> Thanks folks for the question and discussion, I have > some > > >>> WG chair > > >>> > > > > vultures hovering over me > > >>> > > > > > >> > > >>> > > > > > >> making sure i prioritize building slides now (the worst > one > > >>> is > > >>> > > myself > > >>> > > > > ;-), so i will only > > >>> > > > > > >> give a brief answer and will get back to it later when i > > >>> had more > > >>> > > > > time. > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> The calculus that i used is from the [UBS] research > paper > > >>> from > > >>> > > > > Johannes Specht, aka: it has > > >>> > > > > > >> > > >>> > > > > > >> the mathematical proof, full reference is in the gLBF > draft. > > >>> > > There is > > >>> > > > > another, later proof of the > > >>> > > > > > >> > > >>> > > > > > >> calculus from Jean Yves Le Boudec in another research > paper > > >>> which > > >>> > > i'd > > >>> > > > > have to dig up, and > > >>> > > > > > >> > > >>> > > > > > >> depending on whom you ask one or the other is easier to > > >>> read. I > > >>> > > am on > > >>> > > > > the UBS research paper > > >>> > > > > > >> > > >>> > > > > > >> side because i have not studied Jean Yves calculus book. > > >>> But its > > >>> > > > > really beautifully simple > > >>> > > > > > >> > > >>> > > > > > >> that as soon as you think of flows with only burst-size > and > > >>> rate > > >>> > > (or > > >>> > > > > period) of those burst, > > >>> > > > > > >> > > >>> > > > > > >> then your delay through the queue is really just the > sum of > > >>> > > bursts. > > >>> > > > > And i just find beauty > > >>> > > > > > >> > > >>> > > > > > >> in simplicity. And that can not be the full answer to > > >>> Jinoo, but i > > >>> > > > > first need to read up more > > >>> > > > > > >> on his WRR options. > > >>> > > > > > >> > > >>> > > > > > >> The need for doing per-hop dampening is really as i said > > >>> from two > > >>> > > > > points: > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> 1. Unless we do per-hop dampening, we will not get such > a > > >>> simple > > >>> > > > > calculus and equally low latency. > > >>> > > > > > >> > > >>> > > > > > >> The two validation slides of the gLBF presentation show > > >>> that one > > >>> > > can > > >>> > > > > exceed the simple > > >>> > > > > > >> > > >>> > > > > > >> calculated bounded latency already with as few as 9 > flows > > >>> across > > >>> > > a > > >>> > > > > single hop and arriving > > >>> > > > > > >> > > >>> > > > > > >> into one single queue - unless there is per-hop > dampening > > >>> (or > > >>> > > > > per-flow-shaper). > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> 2. I can not imagine how to safely sell router equipment > > >>> and build > > >>> > > > > out all desirable topologies without > > >>> > > > > > >> > > >>> > > > > > >> every node is able to do the dampening. And i also see > it > > >>> as the > > >>> > > > > right next-generation challenge > > >>> > > > > > >> > > >>> > > > > > >> and option to make that happen in high speed hardware. > > >>> > > Specifically > > >>> > > > > in metro rings, every big aggregation > > >>> > > > > > >> > > >>> > > > > > >> ring node has potentially 100 incoming interfaces and > hence > > >>> can > > >>> > > > > create a lot of bursts onto ring interfaces. > > >>> > > > > > >> > > >>> > > > > > >> Cheers > > >>> > > > > > >> Toerless > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> P.S.: The validation picture in our slides was from our > > >>> Springer > > >>> > > > > Journal article, so > > >>> > > > > > >> > > >>> > > > > > >> i can not simply put a copy on the Internet now, but > ping > > >>> me in > > >>> > > PM if > > >>> > > > > you want an authors copy. > > >>> > > > > > >> > > >>> > > > > > >> On Wed, Jul 12, 2023 at 11:48:36AM +0800, > > >>> peng.shaofu@zte.com.cn > > >>> > > > > wrote: > > >>> > > > > > >> > Hi Jinoo, Toerless > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > Also thank Toerless for bringing us this interested > draft. > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > For the question Jinoo pointed out, I guess, based on > the > > >>> > > similar > > >>> > > > > analysis > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > of deadline on-time per hop, that even if all flows > > >>> departured > > >>> > > from > > >>> > > > > the damper > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > and arrived at the queueing subsystem at the same > time, > > >>> each > > >>> > > flow > > >>> > > > > can still > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > have its worst-case latency, but just consume the next > > >>> round of > > >>> > > > > budget (i.e., > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > the MAX value mentioned in the document). > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > However, consuming the next round of budget, means > that it > > >>> > > relies > > >>> > > > > on the > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > downstream node to compensate latency, and may result > a > > >>> jitter > > >>> > > with > > >>> > > > > MAX > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > (i.e., worst-case latency). Due to this reason, > deadline > > >>> on-time > > >>> > > > > per hop is > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > temperately removed in version-6, waiting for more > strict > > >>> > > proof and > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > optimization. > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > Anyway, gLBF can do same things that deadline on-time > per > > >>> hop > > >>> > > done. > > >>> > > > > The > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > following instuitive exaple is common for these two > > >>> solutions. > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > Assuming that at the last node, all received flows > have > > >>> > > expierenced > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > almost 0 queueding delay on the upstream nodes. > Traffic > > >>> class-8 > > >>> > > has > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > per hop worst case latency 80 us (just an example, > > >>> similar to > > >>> > > delay > > >>> > > > > level > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > of deadline), traffic class-7 has 70 us, ... ..., > traffic > > >>> > > class-1 > > >>> > > > > has 10 us. > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > Then, at time T0, traffic class-8 arrived at the last > > >>> node, it > > >>> > > will > > >>> > > > > dampen 80us; > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > at time T0+10us, traffic class-7 arrived, it will > dampen > > >>> 70us, > > >>> > > and > > >>> > > > > so on. At > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > T0+80us, all traffic class flows will departure from > the > > >>> damper, > > >>> > > > > and send to > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > the same outgoing port. So, an observed packet may > > >>> expierence > > >>> > > > > another > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > round of worst case lantecy if other higher priority > flows > > >>> > > > > existing, or > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > expierence best case latency (almost 0) if other > higher > > >>> priority > > >>> > > > > flows not > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > existing. That is, a jitter with value of worst case > > >>> latency > > >>> > > still > > >>> > > > > exists. > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > Regards, > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > PSF > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > Original > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > From: JinooJoung <jjoung@smu.ac.kr> > > >>> > > > > > >> > To: Toerless Eckert <tte@cs.fau.de>; > > >>> > > > > > >> > Cc: detnet@ietf.org <detnet@ietf.org>; > > >>> > > > > draft-eckert-detnet-glbf@ietf.org > > >>> > > > > > >> <draft-eckert-detnet-glbf@ietf.org>; > > >>> > > > > > >> > Date: 2023年07月09日 09:39 > > >>> > > > > > >> > Subject: Re: [Detnet] FYI: > draft-eckert-detnet-glbf-01.txt > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > _______________________________________________ > > >>> > > > > > >> > detnet mailing list > > >>> > > > > > >> > detnet@ietf.org > > >>> > > > > > >> > https://www.ietf.org/mailman/listinfo/detnet > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > Dear Toerless; thanks for the draft. > > >>> > > > > > >> > > >>> > > > > > >> > gLBF is an interesting approach, similar in concept > to the > > >>> > > Buffered > > >>> > > > > Network (BN) I have introduced in the ADN Framework document. > > >>> > > > > > >> > > >>> > > > > > >> > The difference seems that the BN buffers only once at > the > > >>> > > network > > >>> > > > > boundary, while gLBF buffers at every node. > > >>> > > > > > >> > > >>> > > > > > >> > Therefore in the BN, a buffer handles only a few > flows, > > >>> while in > > >>> > > > > the gLBF a buffer needs to face millions of flows. > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > The implementation complexity should be addressed in > the > > >>> future > > >>> > > > > draft, I think. > > >>> > > > > > >> > > > >>> > > > > > >> > I have a quick question below. > > >>> > > > > > >> > > > >>> > > > > > >> > +------------------------+ > > >>> +------------------------+ > > >>> > > > > > >> > | Node A | | Node B > > >>> > > | > > >>> > > > > > >> > | +-+ +-+ +-+ | | +-+ +-+ > +-+ > > >>> | > > >>> > > > > > >> > |-x-|D|-y-|F|---|Q|----z -|------ > > >>> |-x-|D|-y-|F|---|Q|----z- | > > >>> > > > > > >> > | +-+ +-+ +-+ | Link | +-+ +-+ +-+ > > >>> | > > >>> > > > > > >> > +------------------------+ > > >>> +------------------------+ > > >>> > > > > > >> > |<- A/B in-time latency ->| > > >>> > > > > > >> > |<--A/B on-time latency ------->| > > >>> > > > > > >> > > > >>> > > > > > >> > Figure 3: Forwarding with Damper and > > >>> measuring > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > In Figure 3, how can F and Q guarantee the nodal > latency > > >>> below > > >>> > > MAX? > > >>> > > > > > >> > > >>> > > > > > >> > Does the gLBF provide the same latency bound as that > of > > >>> UBS, as > > >>> > > it > > >>> > > > > is argued? > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > In UBS, an interleaved regulator (IR) works as the > damper > > >>> D in > > >>> > > the > > >>> > > > > gLBF. > > >>> > > > > > >> > > >>> > > > > > >> > IR is essentially a FIFO, whose HoQ packet is > examined and > > >>> > > leaves > > >>> > > > > if eligible. > > >>> > > > > > >> > > >>> > > > > > >> > A packet's eligible time can be earlier than the time > > >>> that it > > >>> > > > > became the HoQ. > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > However, in gLBF, a packet has a precise moment that > it > > >>> needs > > >>> > > to be > > >>> > > > > forwarded from D. > > >>> > > > > > >> > (Therefore, UBS is not a generalized gLBF.) > > >>> > > > > > >> > > >>> > > > > > >> > In the worst case, all the flows may want to send the > > >>> packets > > >>> > > from > > >>> > > > > D to F at the same time. > > >>> > > > > > >> > > >>> > > > > > >> > If it can be implemented as such, bursts may > accumulate, > > >>> and the > > >>> > > > > latency cannot be guaranteed. > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > If it cannot be implemented that way, you may > introduce > > >>> another > > >>> > > > > type of delay. > > >>> > > > > > >> > > > >>> > > > > > >> > Don't you need an additional mechanism for latency > > >>> guarantee? > > >>> > > > > > >> > > > >>> > > > > > >> > Thanks a lot in advance, I support this draft. > > >>> > > > > > >> > > > >>> > > > > > >> > Best, > > >>> > > > > > >> > Jinoo > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > > > >>> > > > > > >> > On Sat, Jul 8, 2023 at 12:05 AM Toerless Eckert < > > >>> tte@cs.fau.de> > > >>> > > > > wrote: > > >>> > > > > > >> > > > >>> > > > > > >> > Dear DetNet WG, > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > FYI on a newly posted bounded latency > method/proposal > > >>> draft > > >>> > > that > > >>> > > > > we call gLBF. > > >>> > > > > > >> > (guaranteed Latency Based Forwarding). > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > gLBF, as compared to TCQF and CSQF is proposed from > our > > >>> side > > >>> > > to be > > >>> > > > > a more long-term > > >>> > > > > > >> > > >>> > > > > > >> > solution, because it has not been validated with > > >>> high-speed > > >>> > > > > forwarding hardware and requires > > >>> > > > > > >> > > >>> > > > > > >> > new network header information for the damper value, > > >>> whereas > > >>> > > > > TCQF/CSQF of course can operate > > >>> > > > > > >> > > >>> > > > > > >> > without new headers, have proven high-speed > > >>> implementations PoC > > >>> > > > > and are therefore really > > >>> > > > > > >> > ready for adoption now. > > >>> > > > > > >> > > > >>> > > > > > >> > > >>> > > > > > >> > gLBF is a specific variant of the damper idea that is > > >>> meant to > > >>> > > be > > >>> > > > > compatible with the > > >>> > > > > > >> > > >>> > > > > > >> > TSN-ATS latency calculus so that it can use the same > > >>> > > > > controller-plane/path-computation > > >>> > > > > > >> > > >>> > > > > > >> > algorithms and implementations one would use for > > >>> TSN-ATS. It > > >>> > > also > > >>> > > > > allows to eliminate > > >>> > > > > > >> > > >>> > > > > > >> > the need for hop-by-hop clock synchronization and (we > > >>> hope) > > >>> > > should > > >>> > > > > be well implementable > > >>> > > > > > >> > in high-speed hardware. > > >>> > > > > > >> > > > >>> > > > > > >> > Any feedback welcome. > > >>> > > > > > >> > > > >>> > > > > > >> > Cheers > > >>> > > > > > >> > Toerless > > >>> > > > > > >> > > > >>> > > > > > >> > In-Reply-To: < > > >>> > > > > 168874067601.53296.4506535864118204933@ietfa.amsl.com> > > >>> > > > > > >> > > > >>> > > > > > >> > On Fri, Jul 07, 2023 at 07:37:56AM -0700, > > >>> > > internet-drafts@ietf.org > > >>> > > > > > >> wrote: > > >>> > > > > > >> > > > > >>> > > > > > >> > > A new version of I-D, > draft-eckert-detnet-glbf-01.txt > > >>> > > > > > >> > > has been successfully submitted by Toerless Eckert > and > > >>> > > posted to > > >>> > > > > the > > >>> > > > > > >> > > IETF repository. > > >>> > > > > > >> > > > > >>> > > > > > >> > > Name: draft-eckert-detnet-glbf > > >>> > > > > > >> > > Revision: 01 > > >>> > > > > > >> > > >>> > > > > > >> > > Title: Deterministic Networking > > >>> (DetNet) Data > > >>> > > > > Plane - guaranteed Latency Based Forwarding (gLBF) for > bounded > > >>> latency > > >>> > > with > > >>> > > > > low jitter and asynchronous forwarding in Deterministic > Networks > > >>> > > > > > >> > > Document date: 2023-07-07 > > >>> > > > > > >> > > Group: Individual Submission > > >>> > > > > > >> > > Pages: 39 > > >>> > > > > > >> > > URL: > > >>> > > > > > >> > > >>> https://www.ietf.org/archive/id/draft-eckert-detnet-glbf-01.txt > > >>> > > > > > >> > > Status: > > >>> > > > > > >> > https://datatracker.ietf.org/doc/draft-eckert-detnet-glbf/ > > >>> > > > > > >> > > Htmlized: > > >>> > > > > > >> > > >>> https://datatracker.ietf.org/doc/html/draft-eckert-detnet-glbf > > >>> > > > > > >> > > Diff: > > >>> > > > > > >> > > >>> > > > > >>> > https://author-tools.ietf.org/iddiff?url2=draft-eckert-detnet-glbf-01 > > >>> > > > > > >> > > > > >>> > > > > > >> > > Abstract: > > >>> > > > > > >> > > This memo proposes a mechanism called > "guaranteed > > >>> Latency > > >>> > > > > Based > > >>> > > > > > >> > > >>> > > > > > >> > > Forwarding" (gLBF) as part of DetNet for > hop-by-hop > > >>> packet > > >>> > > > > forwarding > > >>> > > > > > >> > > with per-hop deterministically bounded latency > and > > >>> minimal > > >>> > > > > jitter. > > >>> > > > > > >> > > > > >>> > > > > > >> > > gLBF is intended to be useful across a wide > range of > > >>> > > networks > > >>> > > > > and > > >>> > > > > > >> > > applications with need for high-precision > > >>> deterministic > > >>> > > > > networking > > >>> > > > > > >> > > >>> > > > > > >> > > services, including in-car networks or networks > > >>> used for > > >>> > > > > industrial > > >>> > > > > > >> > > automation across on factory floors, all the > way to > > >>> > > ++100Gbps > > >>> > > > > > >> > > country-wide networks. > > >>> > > > > > >> > > > > >>> > > > > > >> > > Contrary to other mechanisms, gLBF does not > require > > >>> > > network > > >>> > > > > wide > > >>> > > > > > >> > > >>> > > > > > >> > > clock synchronization, nor does it need to > maintain > > >>> > > per-flow > > >>> > > > > state at > > >>> > > > > > >> > > network nodes, avoiding drawbacks of other known > > >>> methods > > >>> > > while > > >>> > > > > > >> > > leveraging their advantages. > > >>> > > > > > >> > > > > >>> > > > > > >> > > Specifically, gLBF uses the queuing model and > > >>> calculus of > > >>> > > > > Urgency > > >>> > > > > > >> > > Based Scheduling (UBS, [UBS]), which is used by > TSN > > >>> > > > > Asynchronous > > >>> > > > > > >> > > Traffic Shaping [TSN-ATS]. gLBF is intended to > be a > > >>> > > plug-in > > >>> > > > > > >> > > replacement for TSN-ATN or as a parallel > mechanism > > >>> beside > > >>> > > > > TSN-ATS > > >>> > > > > > >> > > >>> > > > > > >> > > because it allows to keeping the same > > >>> controller-plane > > >>> > > design > > >>> > > > > which > > >>> > > > > > >> > > is selecting paths for TSN-ATS, sizing TSN-ATS > > >>> queues, > > >>> > > > > calculating > > >>> > > > > > >> > > latencies and admitting flows to calculated > paths > > >>> for > > >>> > > > > calculated > > >>> > > > > > >> > > latencies. > > >>> > > > > > >> > > > > >>> > > > > > >> > > >>> > > > > > >> > > In addition to reducing the jitter compared to > > >>> TSN-ATS by > > >>> > > > > additional > > >>> > > > > > >> > > >>> > > > > > >> > > buffering (dampening) in the network, gLBF also > > >>> eliminates > > >>> > > > > the need > > >>> > > > > > >> > > for per-flow, per-hop state maintenance > required by > > >>> > > TSN-ATS. > > >>> > > > > This > > >>> > > > > > >> > > avoids the need to signal per-flow state to > every > > >>> hop > > >>> > > from the > > >>> > > > > > >> > > controller-plane and associated scaling > problems. > > >>> It also > > >>> > > > > reduces > > >>> > > > > > >> > > implementation cost for high-speed networking > > >>> hardware > > >>> > > due to > > >>> > > > > the > > >>> > > > > > >> > > >>> > > > > > >> > > avoidance of additional high-speed speed > read/write > > >>> memory > > >>> > > > > access to > > >>> > > > > > >> > > retrieve, process and update per-flow state > > >>> variables for > > >>> > > a > > >>> > > > > large > > >>> > > > > > >> > > number of flows. > > >>> > > > > > >> > > > > >>> > > > > > >> > > >>> > > > > > >> > > > > >>> > > > > > >> > > > > >>> > > > > > >> > > > > >>> > > > > > >> > > The IETF Secretariat > > >>> > > > > > >> > > > >>> > > > > > >> > _______________________________________________ > > >>> > > > > > >> > detnet mailing list > > >>> > > > > > >> > detnet@ietf.org > > >>> > > > > > >> > https://www.ietf.org/mailman/listinfo/detnet > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> -- > > >>> > > > > > >> --- > > >>> > > > > > >> tte@cs.fau.de > > >>> > > > > > >> > > >>> > > > > > >> _______________________________________________ > > >>> > > > > > >> detnet mailing list > > >>> > > > > > >> detnet@ietf.org > > >>> > > > > > >> https://www.ietf.org/mailman/listinfo/detnet > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > >> > > >>> > > > > > > > > >>> > > > > > > >>> > > > > -- > > >>> > > > > --- > > >>> > > > > tte@cs.fau.de > > >>> > > > > > > >>> > > > > >>> > > -- > > >>> > > --- > > >>> > > tte@cs.fau.de > > >>> > > > > >>> > > >>> -- > > >>> --- > > >>> tte@cs.fau.de > > >>> > > >> > > >> > > > > > -- > --- > tte@cs.fau.de >
- [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt peng.shaofu
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt peng.shaofu
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt peng.shaofu
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt peng.shaofu
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt peng.shaofu
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt peng.shaofu
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt peng.shaofu
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Andrea Francini (Nokia)
- [Detnet] Andrea's comments on C-SCORE Jinoo Joung
- Re: [Detnet] Andrea's comments on C-SCORE Andrea Francini (Nokia)
- Re: [Detnet] Andrea's comments on C-SCORE Andrea Francini (Nokia)
- Re: [Detnet] Andrea's comments on C-SCORE Jinoo Joung
- Re: [Detnet] Andrea's comments on C-SCORE Andrea Francini (Nokia)
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Jinoo Joung
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt peng.shaofu
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt peng.shaofu
- Re: [Detnet] FYI: draft-eckert-detnet-glbf-01.txt Toerless Eckert