Re: [aqm] [tsvwg] New Version Notification for draft-baker-tsvwg-aqm-recommendation-00.txt

"Fred Baker (fred)" <fred@cisco.com> Thu, 14 March 2013 11:55 UTC

Return-Path: <fred@cisco.com>
X-Original-To: aqm@ietfa.amsl.com
Delivered-To: aqm@ietfa.amsl.com
Received: from localhost (localhost [127.0.0.1]) by ietfa.amsl.com (Postfix) with ESMTP id AD8B121F8F2F; Thu, 14 Mar 2013 04:55:25 -0700 (PDT)
X-Virus-Scanned: amavisd-new at amsl.com
X-Spam-Flag: NO
X-Spam-Score: -107.773
X-Spam-Level:
X-Spam-Status: No, score=-107.773 tagged_above=-999 required=5 tests=[RCVD_IN_DNSWL_HI=-8, SARE_SUB_OBFU_Q1=0.227, USER_IN_WHITELIST=-100]
Received: from mail.ietf.org ([12.22.58.30]) by localhost (ietfa.amsl.com [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id tDgQ8m2OdOYO; Thu, 14 Mar 2013 04:55:21 -0700 (PDT)
Received: from rcdn-iport-9.cisco.com (rcdn-iport-9.cisco.com [173.37.86.80]) by ietfa.amsl.com (Postfix) with ESMTP id 650F921F8F22; Thu, 14 Mar 2013 04:55:21 -0700 (PDT)
DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=cisco.com; i=@cisco.com; l=6852; q=dns/txt; s=iport; t=1363262121; x=1364471721; h=from:to:cc:subject:date:message-id:references: in-reply-to:content-id:content-transfer-encoding: mime-version; bh=dwdOr4tcuxvoFIoCPmTmnmHrI6y2T9QHLZopC/8dHTM=; b=nKhGIx9lteeLOIPlD6Hk+06xML3zKE8tHtHUdQPLITRSwv1+oAB9ku5j Ju6VQ9Rnp4S2c+/riW7iksdzz1wnaCmMo5z1QqC+0bJM2CjH3ut2qCAcX XrmxlB0qtqSc/ZipVJFDysZTR/nRPeZRCbZPycQW3MM3r4VbAG3DgTqXT A=;
X-IronPort-Anti-Spam-Filtered: true
X-IronPort-Anti-Spam-Result: AgAFAH65QVGtJV2a/2dsb2JhbABDxHmBYRZ0gikBAQECAQFsCwIQAgEIDgoKJDIlAgQOBQiIBgbBLY5dAjEHgl9hA4g+ilWUR4FUgTaCKA
X-IronPort-AV: E=Sophos;i="4.84,844,1355097600"; d="scan'208";a="184402837"
Received: from rcdn-core-3.cisco.com ([173.37.93.154]) by rcdn-iport-9.cisco.com with ESMTP; 14 Mar 2013 11:55:20 +0000
Received: from xhc-rcd-x13.cisco.com (xhc-rcd-x13.cisco.com [173.37.183.87]) by rcdn-core-3.cisco.com (8.14.5/8.14.5) with ESMTP id r2EBtK9C029555 (version=TLSv1/SSLv3 cipher=AES128-SHA bits=128 verify=FAIL); Thu, 14 Mar 2013 11:55:20 GMT
Received: from xmb-rcd-x09.cisco.com ([169.254.9.206]) by xhc-rcd-x13.cisco.com ([173.37.183.87]) with mapi id 14.02.0318.004; Thu, 14 Mar 2013 06:55:10 -0500
From: "Fred Baker (fred)" <fred@cisco.com>
To: Michael Welzl <michawe@ifi.uio.no>
Thread-Topic: [tsvwg] New Version Notification for draft-baker-tsvwg-aqm-recommendation-00.txt
Thread-Index: AQHOIAqa1tJ18EAwSkCm0kPJ4goPbg==
Date: Thu, 14 Mar 2013 11:55:09 +0000
Message-ID: <8C48B86A895913448548E6D15DA7553B7D1BBA@xmb-rcd-x09.cisco.com>
References: <20130313164833.27324.89314.idtracker@ietfa.amsl.com> <8C48B86A895913448548E6D15DA7553B7D045F@xmb-rcd-x09.cisco.com> <b6d8c6a56b1e4f066019c81da8a202a0.squirrel@www.erg.abdn.ac.uk> <8C48B86A895913448548E6D15DA7553B7D0BD6@xmb-rcd-x09.cisco.com> <fc97d79292a1cfb2efab45753f90351b.squirrel@www.erg.abdn.ac.uk> <4936E58C-A6AE-485D-BD6E-7E6D30493359@ifi.uio.no> <8C48B86A895913448548E6D15DA7553B7D1A35@xmb-rcd-x09.cisco.com>
In-Reply-To: <8C48B86A895913448548E6D15DA7553B7D1A35@xmb-rcd-x09.cisco.com>
Accept-Language: en-US
Content-Language: en-US
X-MS-Has-Attach:
X-MS-TNEF-Correlator:
x-originating-ip: [10.82.243.95]
Content-Type: text/plain; charset="iso-8859-1"
Content-ID: <31CD23BE7ECB3444A6BF64A5C1829D6C@emea.cisco.com>
Content-Transfer-Encoding: quoted-printable
MIME-Version: 1.0
Cc: "tsv-area@ietf.org" <TSV-area@ietf.org>, "aqm@ietf.org" <aqm@ietf.org>
Subject: Re: [aqm] [tsvwg] New Version Notification for draft-baker-tsvwg-aqm-recommendation-00.txt
X-BeenThere: aqm@ietf.org
X-Mailman-Version: 2.1.12
Precedence: list
List-Id: "Discussion list for active queue management and flow isolation." <aqm.ietf.org>
List-Unsubscribe: <https://www.ietf.org/mailman/options/aqm>, <mailto:aqm-request@ietf.org?subject=unsubscribe>
List-Archive: <http://www.ietf.org/mail-archive/web/aqm>
List-Post: <mailto:aqm@ietf.org>
List-Help: <mailto:aqm-request@ietf.org?subject=help>
List-Subscribe: <https://www.ietf.org/mailman/listinfo/aqm>, <mailto:aqm-request@ietf.org?subject=subscribe>
X-List-Received-Date: Thu, 14 Mar 2013 11:55:25 -0000

This will be my last note on the topic on tsvwg. I'd like to believe that by tomorrow we will have all joined aqm@.

On Mar 14, 2013, at 6:54 AM, Fred Baker (fred) <fred@cisco.com> wrote:

> 
> On Mar 14, 2013, at 12:43 AM, Michael Welzl <michawe@ifi.uio.no> wrote:
> 
>> Great, I turn on ECN, and that gives me more delay, just what I want!
>> And, even better: the more people use ECN, the more delay everybody gets.
>> 
>> Seriously, I see the incentive idea behind this two-level marking idea, but one has to carefully consider the increased delay vs. gained throughput trade-off in such a scheme.
> 
> I don't understand your comment. Fill me in please?
> 
> If I have a tail-drop queue, TCP using CUBIC or New Reno seeks to keep the queue full. If the queue has N positions (there are many ways to measure a queue, bear with this one for the purpose of the example), worst case delay approximates waiting for N messages, and general case delay is predicted by queuing theory to shoot for N when the link has 95% utilization or whatever.
> 
> With any AQM algorithm, the same queue will accept N messages in the event that it gets a burst, but will start marking/dropping at some point M significantly less than N, so that the queue depth tends to approximate M, not N. That's the whole point of any AQM algorithm. How M is chosen or predicted is of course different for various algorithms.
> 
> The pushback I generally hear about ECN is that people will mark traffic as ECN-capable in order to work around AQM algorithms signaling early; to make them signal later. I am told that people do abuse the EF code point in order to make their traffic go into a priority queue, so I can imagine people doing this as well.
> 
> What I am suggesting is that the AQM algorithm use the appropriate signal to make queue depth approximate M, whether than is ECN or loss depending on the traffic marking, and in the event of abuse do no worse than present it's-done-everywhere tail drop, in which the queue depth tends towards N.
> 
> M < N.
> 
> Remind me how saying that one will use ECN for ECN-capable traffic makes the average queue depth deeper? That makes no sense to me.


Following up on the comment about M and N. Let me show you some (old and low speed) slides that I use as examples in this class of discussion. It's not ECN vs loss, it's AQM vs tail-drop. However, since AQM-using-ECN and AQM-using-loss are examples of AQM, and the argument in the draft is to use both depending on the capabilities of the traffic, I think it's applicable.

    ftp://ftpeng.cisco.com/fred/dpreed/Bufferbloat_Masterclass.pdf

I'm looking at slides 11 and 12. Comments on the rest of the deck are interesting, but I'd suggest taking that private in order to not muddy this discussion.

The slides were pulled together as a simple and understandable example of what I said above about M and N. "AQM", in this case, is "RED"; other algorithms will display slightly different characteristics, but I assert that they will have the same basic behavior. The value and perhaps appearance of M will vary, but not the fact of it.

The test setup is two hosts talking across two low end routers connected with a 2 MBPS link. I could go faster; the charts would display timings appropriate to the line speeds and I would have to generate some multiple of the traffic to present the issue, but the structure of the behavior would be about the same. 

     +----+   +------+         +------+   +-----+
     |Left|   |Left  +---------+Right |   |Right|
     |Host|   |Router|         |Router|   |Host |
     +--+-+   +---+--+         +---+--+   +---+-+
        |         |                |          |
    ----+---------+---         ----+----------+---

On the "Left Host", I have a large file, large enough to take perhaps ten seconds to move to the "Right Host" in isolation. In one window on "Left", I run a ping -s to measure the RTT of traffic between the two hosts, and by extension the changing depth of the queue in "Left Router". In another window, I run a script. The script counts from one to fifteen. On each value, it simultaneously opens up that number of FTPs to copy said file to "Right Host". When the last FTP in the set completes, it waits a few seconds, rsh's to delete the now-unneeded copies from "Right", and goes on to the next value.

Afterwards, I did some simple calculations in Excel. I took successive and overlapping sets of ping measurements (ten, IIRC) and calculated the min, max, average (probably should have been median), and standard deviation of the set. I then plotted the values.

I ran the test twice, the difference being that I used RED or tail drop.

Slide 11 is the tail drop test. You see 15 bumps, measuring traffic from 1..15 FTPs moving said file in parallel. What is pretty obvious is that the variation in the queue is at the top of the queue; if N was five, we would be discussing variation around five; if it was five hundred, we would be discussing variation around five hundred. loss-based TCPs work Really Hard to fill the queue and keep it full, resulting in the potential queue depth being a predictor of end to end delay and variation in delay.

Slide 12 is the RED test. It's pretty obvious where I set min-thresh. I'm not, in this, suggesting a value for min-thresh, BTW; I set it somewhere to demonstrate that where I set it is the average queue depth in the scenario. You want it somewhere else, be my guest. What I observe is wider variation in queue depth, because I am using the configured depth of the queue to absorb bursts when they happen. But the average queue depth is at min-threshold. M, which is less than N.

I'll argue, and I think we probably generally agree, that the optimum target average queue depth is close to zero, and we'd like to use the buffer available to absorb and play out bursts when they happen. It would also be nice if the impact of AQM on the various data flows were more consistent, so that there was less variation in delay, and that might be true of some algorithms. But that's the target.

Coming back to ECN vs loss, ECN doesn't increase delay, it enables the same signal to be sent end to end, but sent more quickly and reliably (it is explicit and is known to be about congestion, as opposed to being implicit and possibly triggered by other factors). It targets the same mean queue depth. Signaling without losing traffic is better than losing traffic and calling that signaling, as the session completes more quickly and reliably and is then out of the way. I don't see the down side.