Re: [MMUSIC] draft-holmberg-mmusic-t140-usage-data-channel - multi-party

Paul Kyzivat <pkyzivat@alum.mit.edu> Wed, 28 August 2019 18:54 UTC

Return-Path: <pkyzivat@alum.mit.edu>
X-Original-To: mmusic@ietfa.amsl.com
Delivered-To: mmusic@ietfa.amsl.com
Received: from localhost (localhost [127.0.0.1]) by ietfa.amsl.com (Postfix) with ESMTP id 6F6B9120073 for <mmusic@ietfa.amsl.com>; Wed, 28 Aug 2019 11:54:09 -0700 (PDT)
X-Virus-Scanned: amavisd-new at amsl.com
X-Spam-Flag: NO
X-Spam-Score: -4.199
X-Spam-Level:
X-Spam-Status: No, score=-4.199 tagged_above=-999 required=5 tests=[BAYES_00=-1.9, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_NONE=0.001, SPF_PASS=-0.001, URIBL_BLOCKED=0.001] autolearn=ham autolearn_force=no
Received: from mail.ietf.org ([4.31.198.44]) by localhost (ietfa.amsl.com [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id jo_y2hyZgxsV for <mmusic@ietfa.amsl.com>; Wed, 28 Aug 2019 11:54:06 -0700 (PDT)
Received: from outgoing-alum.mit.edu (outgoing-alum.mit.edu [18.7.68.33]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ietfa.amsl.com (Postfix) with ESMTPS id 21B4712026E for <mmusic@ietf.org>; Wed, 28 Aug 2019 11:54:06 -0700 (PDT)
Received: from Kokiri.localdomain (c-24-62-227-142.hsd1.ma.comcast.net [24.62.227.142]) (authenticated bits=0) (User authenticated as pkyzivat@ALUM.MIT.EDU) by outgoing-alum.mit.edu (8.14.7/8.12.4) with ESMTP id x7SIs4Vh021769 (version=TLSv1/SSLv3 cipher=AES128-SHA bits=128 verify=NOT) for <mmusic@ietf.org>; Wed, 28 Aug 2019 14:54:04 -0400
To: mmusic@ietf.org
References: <49749CEF-41E8-4E87-8CC6-938DBDA0CEE7@ericsson.com> <HE1PR07MB3161CC8F09E178F2E982E5AC93A60@HE1PR07MB3161.eurprd07.prod.outlook.com> <a5057f34-f222-e290-df88-da6e96d56294@omnitor.se> <A48CA18B-567B-46B9-9501-4FF65D4C8CAE@ericsson.com> <6c6d13ad-8699-73b9-f8aa-8ef16c5b3451@omnitor.se> <FB8E32BC-89C2-41CE-BBE1-81C4E7A56E77@ericsson.com> <f59f9335-329d-7b26-dec3-0898051040ce@omnitor.se> <VI1PR07MB31675A98F9B08A07D984369293A10@VI1PR07MB3167.eurprd07.prod.outlook.com> <084e8ee3-56fe-b764-fbbe-8afbee72e8e3@omnitor.se> <255759e0-c2c0-981a-e530-e05f7553c26e@alum.mit.edu> <173fc580-3a30-913b-ebba-5b0cb59702c8@omnitor.se> <49da444c-c74f-b7b4-49be-72b2e4c3fbb9@alum.mit.edu> <3102c9de-0d99-0a9c-c3d3-1baf2a3d1b5d@omnitor.se> <HE1PR07MB316107D614D83B75EB44421293A30@HE1PR07MB3161.eurprd07.prod.outlook.com> <HE1PR07MB31613E4B56430C7829A4586193A30@HE1PR07MB3161.eurprd07.prod.outlook.com> <HE1PR07MB316197DD9F693CB226C4F60C93A30@HE1PR07MB3161.eurprd07.prod.outlook.com>
From: Paul Kyzivat <pkyzivat@alum.mit.edu>
Message-ID: <25c89003-7f18-b440-a058-b3e6d8d42fdd@alum.mit.edu>
Date: Wed, 28 Aug 2019 14:54:03 -0400
User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.13; rv:60.0) Gecko/20100101 Thunderbird/60.8.0
MIME-Version: 1.0
In-Reply-To: <HE1PR07MB316197DD9F693CB226C4F60C93A30@HE1PR07MB3161.eurprd07.prod.outlook.com>
Content-Type: text/plain; charset="utf-8"; format="flowed"
Content-Language: en-US
Content-Transfer-Encoding: 8bit
Archived-At: <https://mailarchive.ietf.org/arch/msg/mmusic/rKVGue1Fgnqc1StR8ec_8X6DVPU>
Subject: Re: [MMUSIC] draft-holmberg-mmusic-t140-usage-data-channel - multi-party
X-BeenThere: mmusic@ietf.org
X-Mailman-Version: 2.1.29
Precedence: list
List-Id: Multiparty Multimedia Session Control Working Group <mmusic.ietf.org>
List-Unsubscribe: <https://www.ietf.org/mailman/options/mmusic>, <mailto:mmusic-request@ietf.org?subject=unsubscribe>
List-Archive: <https://mailarchive.ietf.org/arch/browse/mmusic/>
List-Post: <mailto:mmusic@ietf.org>
List-Help: <mailto:mmusic-request@ietf.org?subject=help>
List-Subscribe: <https://www.ietf.org/mailman/listinfo/mmusic>, <mailto:mmusic-request@ietf.org?subject=subscribe>
X-List-Received-Date: Wed, 28 Aug 2019 18:54:15 -0000

ISTM that this discussion is getting pretty far afield from the proper 
scope of the draft. I think much of what is being discussed here belongs 
in a different draft, probably in a different wg, like AVT.

	Thanks,
	Paul

On 8/28/19 2:37 PM, Christer Holmberg wrote:
> ....and, a T140 extension would work both for a single data channel and multiple data channel, because the receiver will only need to look at the T140 data, and not care about what data channel it was received on.
> 
> Regards,
> 
> Christer
> 
> -----Alkuperäinen viesti-----
> Lähettäjä: mmusic <mmusic-bounces@ietf.org> Puolesta Christer Holmberg
> Lähetetty: keskiviikko 28. elokuuta 2019 21.36
> Vastaanottaja: Gunnar Hellström <gunnar.hellstrom@omnitor.se>; Paul Kyzivat <pkyzivat@alum.mit.edu>; mmusic@ietf.org
> Aihe: Re: [MMUSIC] draft-holmberg-mmusic-t140-usage-data-channel - multi-party
> 
> 
>> Also, whatever the solution is, I assume you also want it to work in
>> interworking cases, where the MCU might be using RTP-based text (RFC
>> 4103)? From that perspective a T.140 extension would be the best solution, because it would work both for T140 data channels and RTP streams without having to touch the T140 data.
> 
> .....and here I am talking about the case where you would use a single data channel for multiple participants.
> 
> Regards,
> 
> Christer
> 
> 
> -----Alkuperäinen viesti-----
> Lähettäjä: mmusic <mmusic-bounces@ietf.org> Puolesta Gunnar Hellström
> Lähetetty: keskiviikko 28. elokuuta 2019 21.26
> Vastaanottaja: Paul Kyzivat <pkyzivat@alum.mit.edu>; mmusic@ietf.org
> Aihe: Re: [MMUSIC] draft-holmberg-mmusic-t140-usage-data-channel - multi-party
> 
> Hi Paul, see comment at the end,
> 
> Den 2019-08-28 kl. 16:49, skrev Paul Kyzivat:
>> On 8/27/19 2:57 PM, Gunnar Hellström wrote:
>>> Hi Paul,
>>>
>>> Please see inline,
>>>
>>> Den 2019-08-27 kl. 20:39, skrev Paul Kyzivat:
>>>> On 8/26/19 3:59 PM, Gunnar Hellström wrote:
>>>>
>>>>> 4. A multi-party server S, combining a number of sources into one
>>>>> call to a participant A, with real-time text from each other
>>>>> participant (B,C,...) communicated in just one T140 data channel
>>>>> between S and A. There is a need to indicate source for each
>>>>> T140block sent to A. We currently have no way specified for that.
>>>>> An extension of T.140 could do it.
>>>>
>>>> Is there a reason to ever do this?
>>>>
>>>> In audio and video the "mixing" actually does something
>>>> irreversible. And it takes real work to do it, and doing so reduces
>>>> the bandwith considerably over what is required to transmit the
>>>> individual streams.
>>>>
>>>> For RTT none of that is true. There is very little impact in
>>>> bandwidth or processing in transmitting them all as separate channels.
>>>>
>>>> So why not just say "don't do that"?
>>>
>>> Yes, interesting and realistic thought. It would likely be the best
>>> choice for many practical cases.
>>>
>>> I am not sure however how it will work with a huge conference with
>>> hundreds of participants, and some of them occasionally asking for
>>> the floor and send a bit of RTT text. A server using one data channel
>>> per participant would either be required to establish an enormous
>>> amount of T140 data channels being prepared to send what has been
>>> received, or take the effort to establish a new T140 data channel to
>>> all users at the moment a user gets the floor, and then possibly
>>> close it again.
>>>
>>> What do you think about that case?
>>
>> Initially each user would only have one channel. Then he gets another
>> one added each time there is a new speaker who hasn't spoken before. I
>> don't know if there would be formal floor control, or if everyone
>> would always be allowed to talk. Formal floor control would provide an
>> early hint to establish the new channels, possibly preventing delays.
>> But adding a channel isn't an expensive operation and delaying
>> messages until it is done shouldn't be a problem.
>>
>> But is this a realistic issue? Do RTT conferences with hundreds of
>> speakers happen in practice?
> 
> It is realistic to the same degree as use of video and audio in multi-party multi-media conferences.
> 
> It is not realistic to let hundreds of participants talk simultaneously, but one or a very low number. It is realistic to let one at a time have the floor and speak, sending the audio to all the others, and have the opportunity to hand over the floor to someone else with a lose or strict protocol.
> 
> There is very little use for transmitting live video from hundreds of participants simultaneously, but it is realistic to show a low number of users and be prepared to switch who are seen.
> 
> Similarly it is not realistic to let hundreds of participants present new RTT simultaneously, but one or a very low number. It is realistic to let a few at a time transmit, sending the RTT to all the others. Anyone should have the opportunity to transmit RTT, controlled by either no protocol or a lose or a strict protocol.
> 
> RTT is sadly often characterized as media for accessibility and thereby expected to be used only in special cases. This is wrong from two points of view:
> 
> 1: If used for accessibility, in cases when one or some users have no or little use of audio or video for language communication because of a disability, then it is very important that all other participants can use the same media, so that communication can go directly to or from the
> one(s) who need RTT. This is the whole idea of accessibility, that users are not forced to limited technology corners, but are able to participate anywhere with the same technology as others and have full and equal participation. (sometimes a translating service between different modalities are still needed so that everybody are enabled to use the modality they prefer)
> 
> 2. In any multimedia conference there appears reasons to communicate something in text. The three real-time media: audio, video and RTT are needed together for alternating or simultaneous use. Talk, show and write for a complete and efficient communication. In that scenario, text messaging is a slow tool, causing delays and risk for losing the interest from  viewers. When anyone starts texting RTT, it is possible to start following the thoughts as they are expressed in text immediately from start, while for messaging the typing user needs to complete the message and risking to enter the message far too late to be valid in the discussion.
> 
> 3. Automatic subtitling by speech-to-text services and even language translation before presentation would be useless without a form of RTT.
> It is now realistic to use such services to enhance multimedia conferences and make the contents available for both people with other languages, people in noisy areas and for deaf or hard-of-hearing users.
> This is a rapidly increasing use. Also for this case it is realistic only with a few actively transmitting users, while the result can be if interest to distribute to many.
> 
> This said, I anyway agree that the usage that will develop most rapidly will likely be for meetings with not more than five participants having audio, video and RTT available and most commonly only three taking turns on sending RTT. Some of these calls will be in emergency services and interpreting (or relay) services, while there is also a need for user-to-user conferences.
> 
> I would also like to know which MCU model is most realistic for these applications. Is it the one with separate text channels per active source opened and closed when there is a need, or is it the single text channel with in-line identification of the source for each new data channel message?
> 
> 
> Regards
> 
> Gunnar
> 
>>
>>
>>      Thanks,
>>      Paul
>>
>> _______________________________________________
>> mmusic mailing list
>> mmusic@ietf.org
>> https://www.ietf.org/mailman/listinfo/mmusic
> 
> _______________________________________________
> mmusic mailing list
> mmusic@ietf.org
> https://www.ietf.org/mailman/listinfo/mmusic
> 
> _______________________________________________
> mmusic mailing list
> mmusic@ietf.org
> https://www.ietf.org/mailman/listinfo/mmusic
> 
> _______________________________________________
> mmusic mailing list
> mmusic@ietf.org
> https://www.ietf.org/mailman/listinfo/mmusic
>