Return-Path: <gunnar.hellstrom@omnitor.se>
X-Original-To: mmusic@ietfa.amsl.com
Delivered-To: mmusic@ietfa.amsl.com
Received: from localhost (localhost [127.0.0.1]) by ietfa.amsl.com (Postfix)
 with ESMTP id BCA1D21F9075 for <mmusic@ietfa.amsl.com>;
 Mon, 11 Mar 2013 14:45:44 -0700 (PDT)
X-Virus-Scanned: amavisd-new at amsl.com
X-Spam-Flag: NO
X-Spam-Score: -2.299
X-Spam-Level: 
X-Spam-Status: No, score=-2.299 tagged_above=-999 required=5 tests=[AWL=-0.300,
 BAYES_00=-2.599, HTML_MESSAGE=0.001, J_CHICKENPOX_14=0.6]
Received: from mail.ietf.org ([12.22.58.30]) by localhost (ietfa.amsl.com
 [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id nIcHn3eTrxAS for
 <mmusic@ietfa.amsl.com>; Mon, 11 Mar 2013 14:45:40 -0700 (PDT)
Received: from vsp-authed-01-02.binero.net (vsp-authed02.binero.net
 [195.74.38.226]) by ietfa.amsl.com (Postfix) with SMTP id A9C5221F8EDA for
 <mmusic@ietf.org>; Mon, 11 Mar 2013 14:45:32 -0700 (PDT)
Received: from smtp01.binero.se (unknown [195.74.38.28]) by
 vsp-authed-01-02.binero.net (Halon Mail Gateway) with ESMTP for
 <mmusic@ietf.org>; Mon, 11 Mar 2013 22:44:49 +0100 (CET)
Received: from [192.168.50.38] (h79n2fls31o933.telia.com [212.181.137.79])
 (Authenticated sender: gunnar.hellstrom@omnitor.se) by
 smtp-04-01.atm.binero.net (Postfix) with ESMTPA id 6EF3D3A0F7 for
 <mmusic@ietf.org>; Mon, 11 Mar 2013 22:44:48 +0100 (CET)
Message-ID: <513E504F.1010209@omnitor.se>
Date: Mon, 11 Mar 2013 22:44:47 +0100
From: Gunnar Hellstrom <gunnar.hellstrom@omnitor.se>
User-Agent: Mozilla/5.0 (Windows NT 6.1; WOW64;
 rv:17.0) Gecko/20130215 Thunderbird/17.0.3
MIME-Version: 1.0
To: mmusic@ietf.org
References: <p0624060ecd63af26fe28@dhcp-42ec.meeting.ietf.org>
In-Reply-To: <p0624060ecd63af26fe28@dhcp-42ec.meeting.ietf.org>
Content-Type: multipart/alternative;
 boundary="------------040800080507030709090604"
Subject: Re: [MMUSIC] draft-gellens-negotiating-human-language-02
X-BeenThere: mmusic@ietf.org
X-Mailman-Version: 2.1.12
Precedence: list
List-Id: Multiparty Multimedia Session Control Working Group <mmusic.ietf.org>
List-Unsubscribe: <https://www.ietf.org/mailman/options/mmusic>,
 <mailto:mmusic-request@ietf.org?subject=unsubscribe>
List-Archive: <http://www.ietf.org/mail-archive/web/mmusic>
List-Post: <mailto:mmusic@ietf.org>
List-Help: <mailto:mmusic-request@ietf.org?subject=help>
List-Subscribe: <https://www.ietf.org/mailman/listinfo/mmusic>,
 <mailto:mmusic-request@ietf.org?subject=subscribe>
X-List-Received-Date: Mon, 11 Mar 2013 21:45:45 -0000

This is a multi-part message in MIME format.
--------------040800080507030709090604
Content-Type: text/plain; charset=ISO-8859-1; format=flowed
Content-Transfer-Encoding: 8bit

Before this discussion got its home in mmusic, we discussed quite 
similar topics as you, Dale brought up now.

It was about what was needed to be expressed by the parameters and if 
SDP or SIP was the right place. And in the case of SIP, if RFC 3840 / 
3841 could be a suitable mechanism for routing and decisions on the 
parameters.

Here is part of that discussion that we need to capture.


I see some complication that might be needed in order to reflect 
reality. At least they should be discussed.

And I am also seeing some different ways to specify it.

The complications to discuss are:

*1. Level of preference.*

There may be a need for specifying levels of preference for languages.  
I might strongly prefer to talk English, but have some useful capability 
in French. I want to display that preference and that capability with 
that difference, so that I get English whenever possible, but get the 
call connected even if English is not available at all but French is.

I would assume that two levels are sufficient, but that can be 
discussed:  Preferred and capable.
>
> The draft already proposes that languages be listed in order of 
> preference, which should handle the example you mention: you list 
> English first and French second.  The called party selects English if 
> it is capable and falls back to French if English is not and French 
> is.  This seems much simpler and is a common way of handling 
> situations where there is a preference.  It would be good to keep the 
> mechanism as simple as possible.
> Yes, I am afraid of complicating this beyond the point when users do 
> not manage to get their settings right.
> Still, I do not think that the order is sufficient as level of 
> preference indicator. You may want to indicate capability for one 
> modality but preference for another. ( as my example, capability for 
> ASL, but preference for talking and reading )

If you have a capability for ASL but preference for talking and reading, 
you could initially offer two media streams: a voice with English and a 
text with English.  If accepted, you have your preferred 
communications.  If those are rejected you could then offer video with 
ASL.  Would that handle the case?
No, video is still very valuable for judging the emergency case. Or 
seeing a friend. So, if you support it you want to offer it. But the 
decision on languages and modalities may end up in video being not 
important for language communication.


>
>
>>>> *2. Directionality*
>>>> There is a need for a direction of the language preference. 
>>>> "Transmit, receive or both". or   "Produce, perceive or both". That 
>>>> is easy to understand for the relay service examples.
>>>> A hard-of-hearing user may declare:
>>>>
>>>> Text, capable, produce, English
>>>> Text, prefer, perceive, English
>>>> Audio, prefer, produce, English
>>>> Audio, capable, perceive, English    ( tricky, a typical 
>>>> hard-of-hearing user may have benefit of receiving audio, while it 
>>>> is not usable enough for reliable perception. I do not want to see 
>>>> this eternally complex, but I see a need for refined expressions here)
>>>> video, capable, both, ASL
>>>>
>>>> This should be understood as that the user prefers to speak and get 
>>>> text back, and has benefit of getting voice in parallel with text.  
>>>> ASL signing can be an alternative if the other party has 
>>>> corresponding capability or preference.
>>>>
>>>
>>> The draft does support this (and even mentions some of these 
>>> specific uses) because it proposes an SDP media attribute, and media 
>>> can be specified to be send, receive, or both.
>> No, that is not the same. You want the media to flow, but by the 
>> parameter you want to indicate your preference for how to use it.  
>> You do not want to turn off incoming audio just because you prefer to 
>> talk but read text.
>
> Yes, I see, thanks for the clarification.  Does this need to be part 
> of the session setup?  If you establish all media streams that you 
> wish to use, can you then just use them as you prefer?  I will consult 
> with the NENA accessibility committee on this.
No, there are specific services providing service with one direction but 
not the other. The information is needed for decision on what assisting 
service to invoke. One such service is the captioned telephony, that 
adds rapidly created speech-to-text in parallel with the voice. They 
provide just that. A user will have a very strong preference for getting 
just that service, but could accept with much lower preference to get a 
direct conversation with the far end in combined text and voice.
>
>
>>>> I think it would be useful to move most of the introduction to a 
>>>> structured use case chapter and express the different cases 
>>>> according to a template. Thast can then be used to test if proposed 
>>>> approaches will work.
>>>
>>> I'm not sure I fully understand what you mean by "structured" in 
>>> "structured use case" or "template."  Can you be more specific?
>> I mean just a simple template for how the use case descriptions are 
>> written.
>>
>> E.g.
>> A title indicating what case we have.
>> Description of the calling user and its capabilities and preferences.
>> Description of the answering user and its capabilities and preferences
>> Description of a possible assisting service and its capabilities and 
>> preferences
>> Description of the calling user's indications.
>> Description of the answering user's indications.
>> The resulting decision and outcome
>>>
>>>
>>>> *3.  Specify language and modality at SIP Media tag level instead.
>>>> *There could be some benfits to declare these parameters at the SIP 
>>>> media tag level instead of SDP level.
>>>> A call center can then register with their capabilities already at 
>>>> the SIP REGISTER time, and the caller preferences / callee 
>>>> capabilities mechanism from RFC 3840/ 3841 can be used to select 
>>>> modalities and languages and route the call to the best capable 
>>>> person or combination of person and assisting interpreting.
>>>
>>> Maybe, but one advantage of using SDP is that the ESInet can take 
>>> language and media needs into account during policy-based routing.  
>>> For example, in some European countries emergency calls placed by a 
>>> speaker of language x in country y may be routed to a PSAP in a 
>>> country where x is the native language.  Or, there might be regional 
>>> or national text relay or sign language interpreter services as 
>>> opposed to PSAP-level capabilities.
>> Is there a complete specification for how policy based routing is 
>> thought to work? Where?
>> Does it not use RFC 3840/3841?
>> That procedure is already supported by SIP servers. Using SDP 
>> requires new SIP server programming.
>
> NENA has a document under development.  I thought it was able to take 
> SDP into account but I'll look into it, and I'm sure Brian will have 
> something to say.
Yes, I think I have seen that. But it needs to come into IETF to be 
possible to refer to.
>
>
>>>> But, on the other hand, then we need a separate specification of 
>>>> what modality the parameters indicate, because the language tags 
>>>> only distinguish between signed and other languages, and "other" 
>>>> seems to mean either spoken or written without any difference.
>>>>
>>>
>>> The SDP media already indicates the type (audio, video, text).
>> Yes, convenient. But there is no knowledge about the parameters until 
>> at call time. It could be better to know the callee capabilities in 
>> advance if available. Then middle boxes can do the routing instead of 
>> the far end. There may be many terminals competing for the call and 
>> the comparison about who to get it should be done by a sip server 
>> instead of an endpoint.
>
> I think call time is the right time.  For emergency calls, it isolates 
> the decision making about how to process calls requiring text, sign 
> language, foreign language, etc. to the ESInet and PSAPs, which is I 
> think the right place.  The processing rules in the ESInet can then be 
> changed without involving any carrier.  The capabilities of an entity 
> may vary based on dynamic factors (time of day, load, etc.) so the 
> decision as to how to support a need may be best made by the ESInet or 
> PSAP in the case of an emergency call, or called party for 
> non-emergency calls.  For example, at some times or under some loads, 
> emergency calls may be routed to a specific PSAP that is not the 
> geographically indicated one.  Likewise, a non-emergency call to a 
> call center may be routed to a center in a country that has support 
> for the language or media needed.
The decision is of course made at call time. With the RFC 3840/3841 
method, the different agents and services available register their 
availability and capabilities when they go on duty, and unregister when 
they stop, so that their information is available at call time.

>
> Further, it is often the case that the cost of relay, interpretation, 
> or translation services is affected by which entity invokes the service.
Yes, that is a complicating policy issue.
>
>
>>>> *4. Problem that 3GPP specifies that it is the UAs only who specify 
>>>> and act on these parameters.
>>>> *I think it is a problem that 3GPP inserted the restriction that 
>>>> the language and modality negotiation shall only bother the 
>>>> involved UAs.
>>>> It would be more natural that it is a service provider between them 
>>>> who detect the differences and make the decision to invoke a relay 
>>>> service for the relay case.
>>>> How do you propose to solve that? Let the service provider behave 
>>>> as a B2BUA, who then can behave as both a UA and a service provider?
>>>
>>> What do you mean by "service provider?"  In the case of a voice 
>>> service provider such as a cellular carrier or a VoIP provider, I 
>>> think this should be entirely transparent.  The voice service 
>>> provider knows it is an emergency call and routes to an ESInet.  It 
>>> is then up to the ESInet and the PSAPs to handle the call as they wish.
>> It can be a service provider for just the function to make advanced 
>> call invocation based on language preferences. The same type of 
>> decisions, call connections and assisting service invocation are 
>> needed in everyday calls as in emergency calls. But it can also be a 
>> service provider for emergency services and the user is registered by 
>> that service provider. They can make decisions on the call. E.g. 
>> detect that it is an emergency call requiring interpreter, and 
>> therefore connect to both the PSAP and interpreter at the same time 
>> to save time.
>
> I think it's best to make these decisions at the end, not the middle.  
> In the case of emergency calls, the ESInet can route to a particular 
> PSAP, the PSAP may bridge in translation or interpretation services, 
> etc.  In the case of non-emergency calls, the call center may support 
> some capabilities locally at some hours but route to a different call 
> center at other times.
The end is not decided until you have evaluated the alternative possible 
ends and decided who has the right capability and preference.



There is another issue with using sdp for decisions. SIP Message is 
included in the set of methods to handle in emergency calls in RFC 6443. 
It can be used within sessions to carry text messages if other media are 
used as well. It is no favored way to have text communication, but 
possible. SIP message has no sdp.  I know that the 3GPP sections  about 
emergency calling in TS 22.101 points towards using MSRP for text 
messaging, so it should not be an issue for 3GPP. Can we neglect SIP 
Message from the discussion and aim at solving it only for real-time 
conversational media?  I do not urge for solving it for SIP Message, I 
just wanted to point out that result by basing the mechanism on SDP.






Will there be a possibility for remote participation on Thursday. I am 
sorry I am not there, but would like to participate if possible.
/Gunnar

------------------------------------------------------------------------
Gunnar Hellström
Omnitor
gunnar.hellstrom@omnitor.se
+46708204288
On 2013-03-11 16:57, Randall Gellens wrote:
> [[[ resending without Cc list ]]]
>
> Hi Dale,
>
> At 11:00 AM -0500 2/25/13, Dale R. Worley wrote:
>
>>  (It's not clear to me what the proper mailing list is to discuss this
>>  draft.  From the headers of the messages, it appears that the primary
>>  list is ietf@ietf.org, but the first message in this thread about that
>>  draft already has a "Re:" in the subject line, so the discussion
>>  started somewhere else.)
>
> There has been some discussion among those listed in the CC header of 
> this message.  I think the mmusic list is probably the right place to 
> continue the discussion and was planning on doing so more formally 
> with the next revision of the draft.
>
> By the way, the draft was updated and is now at -02: 
> http://www.ietf.org/internet-drafts/draft-gellens-negotiating-human-language-02.txt
>
> There is a face-to-face discussion Thursday 11:30-1:00 at The 
> Tropicale (the cafe in the Caribe Royal).  Please let me know if you 
> can make it.
>
>>  (Also, it's not clear why Randall's messages are coming through in
>>  HTML.)
>
> My apologies; I have gotten in the habit when replying to messages 
> that have style to allow Eudora to send my reply styled as well.
>
>
>>  But onward to questions of substance:
>>
>>  - Why SDP and not SIP?
>>
>>  I'd like to see a more thorough exploration of why language
>>  negotiation is to be handled in SDP rather than SIP.  (SIP, like HTTP,
>>  uses the Content-Language header to specify languages.)  In principle,
>>  specifying data that may be used in call-routing should be done in the
>>  SIP layer, but it's well-accepted in the SIP world that call routing
>>  may be affected by the SDP content as well (e.g., media types).
>
> I think it fits more naturally in SDP since the language is related to 
> the media, e.g., English for audio and ASL for video.
>
>
>>  And some discussion and comparison should be done with the SIP/HTTP
>>  Content-Language header (used to specify the language of the
>>  communications) and the SIP Accept-Language header (used to specify
>>  the language of text components of SIP messages), particularly given
>>  that Accept-Language has different set of language specifiers and a
>>  richer syntax for specifying preferences.  In any case, preference
>>  should be given to reusing one of the existing syntaxes for specifying
>>  language preferences.
>
> I think the semantics of Content-Language and Accept-Language are 
> different from the semantics here, especially when setting up a 
> session with, as an example, an audio stream using English and a video 
> stream using ASL.  (But I can see clients using a default value to set 
> both the SDP language attribute and the HTTP Content-Language, unless 
> configured differently.)
>
> As for reusing existing mechanisms, the draft does contain two 
> alternative proposals, one to re-use the existing 'language' SDP 
> attribute, and one to define a new attribute.
>
>>  - Dependency between media descriptions?
>>
>>     Another example would be a user who is able to speak but is deaf or
>>     hard-of-hearing and requires a voice stream plus a text stream
>>     (known as voice carry over).  Making language a media attribute
>>     allows the standard session negotiation mechanism to handle this by
>>     providing the information and mechanism for the endpoints to make
>>     appropriate decisions.
>>
>>  This scenario suggests that there might be dependency or interaction
>>  between language specifications for different media descriptions.
>>  Whether this is needed should be determined and documented.
>>
>>  - Specifying preference levels?
>>
>>     For example, some users may be able to speak several languages, but
>>     have a preference.
>>
>>  This might argue for describing degrees of preference using "q"
>>  parameters (as in the SIP Accept-Language header).
>>
>>  - Expressing multiple languages in answers
>>
>>     (While it is true that a conversation among multilingual people
>>     often involves multiple languages, it does not seem useful enough
>>     as a general facility to warrant complicating the desired semantics
>>     of the SDP attribute to allow negotiation of multiple simultaneous
>>     languages within an interactive media stream.)
>>
>>  Why shouldn't an answer be able to indicate multiple languages?  At
>>  the least, this might provide the offerer with useful information.
>
> You raise good questions that I think need more discussion.  I am 
> hoping to keep the work as simple as possible and not add additional 
> complexity, which argues for not solving every aspect of the problem, 
> but only those that must be solved immediately.
>
>>
>>  - Reusing a=lang
>>
>>  Searching, I can only find these descriptions of the use of
>>  "a=lang:...":
>>
>>      RFC 4566
>>      draft-saintandre-sip-xmpp-chat
>>      draft-gellens-negotiating-human-language
>>
>>  So it looks like "a=lang:..." is entirely unused at the present and is
>>  safe to be redefined.
>
>
>
>
>


--------------040800080507030709090604
Content-Type: text/html; charset=ISO-8859-1
Content-Transfer-Encoding: 7bit

<html>
  <head>
    <meta content="text/html; charset=ISO-8859-1"
      http-equiv="Content-Type">
  </head>
  <body text="#000000" bgcolor="#FFFFFF">
    <div class="moz-cite-prefix">Before this discussion got its home in
      mmusic, we discussed quite similar topics as you, Dale brought up
      now.<br>
      <br>
      It was about what was needed to be expressed by the parameters and
      if SDP or SIP was the right place. And in the case of SIP, if RFC
      3840 / 3841 could be a suitable mechanism for routing and
      decisions on the parameters.<br>
      <br>
      Here is part of that discussion that we need to capture.<br>
      <br>
      <br>
      I see some complication that might be needed in order to reflect
      reality. At least they should be discussed.<br>
      <br>
      And I am also seeing some different ways to specify it.<br>
      <br>
      The complications to discuss are:<br>
      <br>
      <b>1. Level of preference.</b><br>
      <br>
      There may be a need for specifying levels of preference for
      languages.&nbsp; I might strongly prefer to talk English, but have some
      useful capability in French. I want to display that preference and
      that capability with that difference, so that I get English
      whenever possible, but get the call connected even if English is
      not available at all but French is.<br>
      <br>
      I would assume that two levels are sufficient, but that can be
      discussed:&nbsp; Preferred and capable.<br>
      <blockquote type="cite" cite=""><br>
      </blockquote>
      <blockquote type="cite" cite="">The draft already proposes that
        languages be listed in order of preference, which should handle
        the example you mention: you list English first and French
        second.&nbsp; The called party selects English if it is capable and
        falls back to French if English is not and French is.&nbsp; This
        seems much simpler and is a common way of handling situations
        where there is a preference.&nbsp; It would be good to keep the
        mechanism as simple as possible.<br>
      </blockquote>
      <blockquote type="cite" cite="">Yes, I am afraid of complicating
        this beyond the point when users do not manage to get their
        settings right.</blockquote>
      <blockquote type="cite" cite="">Still, I do not think that the
        order is sufficient as level of preference indicator. You may
        want to indicate capability for one modality but preference for
        another. ( as my example, capability for ASL, but preference for
        talking and reading )</blockquote>
      <div><br>
      </div>
      <div>If you have a capability for ASL but preference for talking
        and reading, you could initially offer two media streams: a
        voice with English and a text with English.&nbsp; If accepted, you
        have your preferred communications.&nbsp; If those are rejected you
        could then offer video with ASL.&nbsp; Would that handle the case?</div>
      No, video is still very valuable for judging the emergency case.
      Or seeing a friend. So, if you support it you want to offer it.
      But the decision on languages and modalities may end up in video
      being not important for language communication.<br>
      <br>
      <br>
      <blockquote cite="mid:p06240603cd5069df24cd@%5B10.177.136.135%5D"
        type="cite">
        <div><br>
        </div>
        <div><br>
        </div>
        <blockquote type="cite" cite="">
          <blockquote type="cite" cite="">
            <blockquote type="cite" cite=""><b>2. Directionality</b><br>
              There is a need for a direction of the language
              preference. "Transmit, receive or both". or&nbsp;&nbsp; "Produce,
              perceive or both". That is easy to understand for the
              relay service examples.<br>
              A hard-of-hearing user may declare:<br>
              <br>
              Text, capable, produce, English<br>
              Text, prefer, perceive, English<br>
              Audio, prefer, produce, English<br>
              Audio, capable, perceive, English&nbsp;&nbsp;&nbsp; ( tricky, a typical
              hard-of-hearing user may have benefit of receiving audio,
              while it is not usable enough for reliable perception. I
              do not want to see this eternally complex, but I see a
              need for refined expressions here)<br>
              video, capable, both, ASL&nbsp;<br>
              <br>
              This should be understood as that the user prefers to
              speak and get text back, and has benefit of getting voice
              in parallel with text.&nbsp; ASL signing can be an alternative
              if the other party has corresponding capability or
              preference.</blockquote>
            <blockquote type="cite" cite=""><br>
            </blockquote>
          </blockquote>
          <blockquote type="cite" cite=""><br>
          </blockquote>
          <blockquote type="cite" cite="">The draft does support this
            (and even mentions some of these specific uses) because it
            proposes an SDP media attribute, and media can be specified
            to be send, receive, or both.<br>
          </blockquote>
        </blockquote>
        <blockquote type="cite" cite="">No, that is not the same. You
          want the media to flow, but by the parameter you want to
          indicate your preference for how to use it.&nbsp; You do not want
          to turn off incoming audio just because you prefer to talk but
          read text.</blockquote>
        <div><br>
        </div>
        <div>Yes, I see, thanks for the clarification.&nbsp; Does this need
          to be part of the session setup?&nbsp; If you establish all media
          streams that you wish to use, can you then just use them as
          you prefer?&nbsp; I will consult with the NENA accessibility
          committee on this.</div>
      </blockquote>
      No, there are specific services providing service with one
      direction but not the other. The information is needed for
      decision on what assisting service to invoke. One such service is
      the captioned telephony, that adds rapidly created speech-to-text
      in parallel with the voice. They provide just that. A user will
      have a very strong preference for getting just that service, but
      could accept with much lower preference to get a direct
      conversation with the far end in combined text and voice.<br>
      <blockquote cite="mid:p06240603cd5069df24cd@%5B10.177.136.135%5D"
        type="cite">
        <div><br>
        </div>
        <div><br>
        </div>
        <blockquote type="cite" cite="">
          <blockquote type="cite" cite="">
            <blockquote type="cite" cite="">I think it would be useful&nbsp;
              to move most of the introduction to a structured use case
              chapter and express the different cases according to a
              template. Thast can then be used to test if proposed
              approaches will work.<br>
            </blockquote>
          </blockquote>
          <blockquote type="cite" cite=""><br>
          </blockquote>
          <blockquote type="cite" cite="">I'm not sure I fully
            understand what you mean by "structured" in "structured use
            case" or "template."&nbsp; Can you be more specific?<br>
          </blockquote>
        </blockquote>
        <blockquote type="cite" cite="">I mean just a simple template
          for how the use case descriptions are written.<br>
          <br>
          E.g.<br>
          A title indicating what case we have.<br>
          Description of the calling user and its capabilities and
          preferences.<br>
          Description of the answering user and its capabilities and
          preferences<br>
          Description of a possible assisting service and its
          capabilities and preferences<br>
          Description of the calling user's indications.<br>
          Description of the answering user's indications.<br>
          The resulting decision and outcome<br>
          <blockquote type="cite" cite=""><br>
          </blockquote>
          <blockquote type="cite" cite=""><br>
            <blockquote type="cite" cite=""><b>3.&nbsp; Specify language and
                modality at SIP Media tag level instead.<br>
              </b>There could be some benfits to declare these
              parameters at the SIP media tag level instead of SDP
              level.<br>
              A call center can then register with their capabilities
              already at the SIP REGISTER time, and the caller
              preferences / callee capabilities mechanism from RFC 3840/
              3841 can be used to select modalities and languages and
              route the call to the best capable person or combination
              of person and assisting interpreting.<br>
            </blockquote>
          </blockquote>
          <blockquote type="cite" cite=""><br>
          </blockquote>
          <blockquote type="cite" cite="">Maybe, but one advantage of
            using SDP is that the ESInet can take language and media
            needs into account during policy-based routing.&nbsp; For
            example, in some European countries emergency calls placed
            by a speaker of language x in country y may be routed to a
            PSAP in a country where x is the native language.&nbsp; Or, there
            might be regional or national text relay or sign language
            interpreter services as opposed to PSAP-level capabilities.<br>
          </blockquote>
        </blockquote>
        <blockquote type="cite" cite="">Is there a complete
          specification for how policy based routing is thought to work?
          Where?<br>
          Does it not use RFC 3840/3841?<br>
          That procedure is already supported by SIP servers. Using SDP
          requires new SIP server programming.</blockquote>
        <div><br>
        </div>
        <div>NENA has a document under development.&nbsp; I thought it was
          able to take SDP into account but I'll look into it, and I'm
          sure Brian will have something to say.</div>
      </blockquote>
      Yes, I think I have seen that. But it needs to come into IETF to
      be possible to refer to.<br>
      <blockquote cite="mid:p06240603cd5069df24cd@%5B10.177.136.135%5D"
        type="cite">
        <div><br>
        </div>
        <div><br>
        </div>
        <blockquote type="cite" cite="">
          <blockquote type="cite" cite="">
            <blockquote type="cite" cite="">But, on the other hand, then
              we need a separate specification of what modality the
              parameters indicate, because the language tags only
              distinguish between signed and other languages, and
              "other" seems to mean either spoken or written without any
              difference.</blockquote>
            <blockquote type="cite" cite=""><br>
            </blockquote>
          </blockquote>
          <blockquote type="cite" cite=""><br>
          </blockquote>
          <blockquote type="cite" cite="">The SDP media already
            indicates the type (audio, video, text).<br>
          </blockquote>
        </blockquote>
        <blockquote type="cite" cite="">Yes, convenient. But there is no
          knowledge about the parameters until at call time. It could be
          better to know the callee capabilities in advance if
          available. Then middle boxes can do the routing instead of the
          far end. There may be many terminals competing for the call
          and the comparison about who to get it should be done by a sip
          server instead of an endpoint.</blockquote>
        <div><br>
        </div>
        <div>I think call time is the right time.&nbsp; For emergency calls,
          it isolates the decision making about how to process calls
          requiring text, sign language, foreign language, etc. to the
          ESInet and PSAPs, which is I think the right place.&nbsp; The
          processing rules in the ESInet can then be changed without
          involving any carrier.&nbsp; The capabilities of an entity may vary
          based on dynamic factors (time of day, load, etc.) so the
          decision as to how to support a need may be best made by the
          ESInet or PSAP in the case of an emergency call, or called
          party for non-emergency calls.&nbsp; For example, at some times or
          under some loads, emergency calls may be routed to a specific
          PSAP that is not the geographically indicated one.&nbsp; Likewise,
          a non-emergency call to a call center may be routed to a
          center in a country that has support for the language or media
          needed.</div>
      </blockquote>
      The decision is of course made at call time. With the RFC
      3840/3841 method, the different agents and services available
      register their availability and capabilities when they go on duty,
      and unregister when they stop, so that their information is
      available at call time. <br>
      &nbsp;<br>
      <blockquote cite="mid:p06240603cd5069df24cd@%5B10.177.136.135%5D"
        type="cite">
        <div><br>
        </div>
        <div>Further, it is often the case that the cost of relay,
          interpretation, or translation services is affected by which
          entity invokes the service. </div>
      </blockquote>
      Yes, that is a complicating policy issue. <br>
      <blockquote cite="mid:p06240603cd5069df24cd@%5B10.177.136.135%5D"
        type="cite">
        <div><br>
        </div>
        <div><br>
        </div>
        <blockquote type="cite" cite="">
          <blockquote type="cite" cite="">
            <blockquote type="cite" cite=""><b>4. Problem that 3GPP
                specifies that it is the UAs only who specify and act on
                these parameters.<br>
              </b>I think it is a problem that 3GPP inserted the
              restriction that the language and modality negotiation
              shall only bother the involved UAs.<br>
              It would be more natural that it is a service provider
              between them who detect the differences and make the
              decision to invoke a relay service for the relay case.<br>
              How do you propose to solve that? Let the service provider
              behave as a B2BUA, who then can behave as both a UA and a
              service provider?<br>
            </blockquote>
          </blockquote>
          <blockquote type="cite" cite=""><br>
          </blockquote>
          <blockquote type="cite" cite="">What do you mean by "service
            provider?"&nbsp; In the case of a voice service provider such as
            a cellular carrier or a VoIP provider, I think this should
            be entirely transparent.&nbsp; The voice service provider knows
            it is an emergency call and routes to an ESInet.&nbsp; It is then
            up to the ESInet and the PSAPs to handle the call as they
            wish.<br>
          </blockquote>
        </blockquote>
        <blockquote type="cite" cite="">It can be a service provider for
          just the function to make advanced call invocation based on
          language preferences. The same type of decisions, call
          connections and assisting service invocation are needed in
          everyday calls as in emergency calls. But it can also be a
          service provider for emergency services and the user is
          registered by that service provider. They can make decisions
          on the call. E.g. detect that it is an emergency call
          requiring interpreter, and therefore connect to both the PSAP
          and interpreter at the same time to save time.</blockquote>
        <div><br>
        </div>
        <div>I think it's best to make these decisions at the end, not
          the middle.&nbsp; In the case of emergency calls, the ESInet can
          route to a particular PSAP, the PSAP may bridge in translation
          or interpretation services, etc.&nbsp; In the case of non-emergency
          calls, the call center may support some capabilities locally
          at some hours but route to a different call center at other
          times.</div>
      </blockquote>
      The end is not decided until you have evaluated the alternative
      possible ends and decided who has the right capability and
      preference. <br>
      <br>
      <br>
      <br>
      There is another issue with using sdp for decisions. SIP Message
      is included in the set of methods to handle in emergency calls in
      RFC 6443. It can be used within sessions to carry text messages if
      other media are used as well. It is no favored way to have text
      communication, but possible. SIP message has no sdp.&nbsp; I know that
      the 3GPP sections&nbsp; about emergency calling in TS 22.101 points
      towards using MSRP for text messaging, so it should not be an
      issue for 3GPP. Can we neglect SIP Message from the discussion and
      aim at solving it only for real-time conversational media?&nbsp; I do
      not urge for solving it for SIP Message, I just wanted to point
      out that result by basing the mechanism on SDP.<br>
      <br>
      &nbsp; <br>
      <br>
      <br>
      <br>
      <br>
      Will there be a possibility for remote participation on Thursday.
      I am sorry I am not there, but would like to participate if
      possible.<br>
      /Gunnar<br>
      <br>
      <div class="moz-signature">
        <hr>
        Gunnar Hellstr&ouml;m<br>
        Omnitor<br>
        <a class="moz-txt-link-abbreviated" href="mailto:gunnar.hellstrom@omnitor.se">gunnar.hellstrom@omnitor.se</a><br>
        +46708204288<br>
      </div>
      On 2013-03-11 16:57, Randall Gellens wrote:<br>
    </div>
    <blockquote
      cite="mid:p0624060ecd63af26fe28@dhcp-42ec.meeting.ietf.org"
      type="cite">[[[ resending without Cc list ]]]
      <br>
      <br>
      Hi Dale,
      <br>
      <br>
      At 11:00 AM -0500 2/25/13, Dale R. Worley wrote:
      <br>
      <br>
      <blockquote type="cite">&nbsp;(It's not clear to me what the proper
        mailing list is to discuss this
        <br>
        &nbsp;draft.&nbsp; From the headers of the messages, it appears that the
        primary
        <br>
        &nbsp;list is <a class="moz-txt-link-abbreviated" href="mailto:ietf@ietf.org">ietf@ietf.org</a>, but the first message in this thread
        about that
        <br>
        &nbsp;draft already has a "Re:" in the subject line, so the
        discussion
        <br>
        &nbsp;started somewhere else.)
        <br>
      </blockquote>
      <br>
      There has been some discussion among those listed in the CC header
      of this message.&nbsp; I think the mmusic list is probably the right
      place to continue the discussion and was planning on doing so more
      formally with the next revision of the draft.
      <br>
      <br>
      By the way, the draft was updated and is now at -02:
<a class="moz-txt-link-freetext" href="http://www.ietf.org/internet-drafts/draft-gellens-negotiating-human-language-02.txt">http://www.ietf.org/internet-drafts/draft-gellens-negotiating-human-language-02.txt</a><br>
      <br>
      There is a face-to-face discussion Thursday 11:30-1:00 at The
      Tropicale (the cafe in the Caribe Royal).&nbsp; Please let me know if
      you can make it.
      <br>
      <br>
      <blockquote type="cite">&nbsp;(Also, it's not clear why Randall's
        messages are coming through in
        <br>
        &nbsp;HTML.)
        <br>
      </blockquote>
      <br>
      My apologies; I have gotten in the habit when replying to messages
      that have style to allow Eudora to send my reply styled as well.
      <br>
      <br>
      <br>
      <blockquote type="cite">&nbsp;But onward to questions of substance:
        <br>
        <br>
        &nbsp;- Why SDP and not SIP?
        <br>
        <br>
        &nbsp;I'd like to see a more thorough exploration of why language
        <br>
        &nbsp;negotiation is to be handled in SDP rather than SIP.&nbsp; (SIP,
        like HTTP,
        <br>
        &nbsp;uses the Content-Language header to specify languages.)&nbsp; In
        principle,
        <br>
        &nbsp;specifying data that may be used in call-routing should be done
        in the
        <br>
        &nbsp;SIP layer, but it's well-accepted in the SIP world that call
        routing
        <br>
        &nbsp;may be affected by the SDP content as well (e.g., media types).
        <br>
      </blockquote>
      <br>
      I think it fits more naturally in SDP since the language is
      related to the media, e.g., English for audio and ASL for video.
      <br>
      <br>
      <br>
      <blockquote type="cite">&nbsp;And some discussion and comparison should
        be done with the SIP/HTTP
        <br>
        &nbsp;Content-Language header (used to specify the language of the
        <br>
        &nbsp;communications) and the SIP Accept-Language header (used to
        specify
        <br>
        &nbsp;the language of text components of SIP messages), particularly
        given
        <br>
        &nbsp;that Accept-Language has different set of language specifiers
        and a
        <br>
        &nbsp;richer syntax for specifying preferences.&nbsp; In any case,
        preference
        <br>
        &nbsp;should be given to reusing one of the existing syntaxes for
        specifying
        <br>
        &nbsp;language preferences.
        <br>
      </blockquote>
      <br>
      I think the semantics of Content-Language and Accept-Language are
      different from the semantics here, especially when setting up a
      session with, as an example, an audio stream using English and a
      video stream using ASL.&nbsp; (But I can see clients using a default
      value to set both the SDP language attribute and the HTTP
      Content-Language, unless configured differently.)
      <br>
      <br>
      As for reusing existing mechanisms, the draft does contain two
      alternative proposals, one to re-use the existing 'language' SDP
      attribute, and one to define a new attribute.
      <br>
      <br>
      <blockquote type="cite">&nbsp;- Dependency between media descriptions?
        <br>
        <br>
        &nbsp;&nbsp;&nbsp; Another example would be a user who is able to speak but is
        deaf or
        <br>
        &nbsp;&nbsp;&nbsp; hard-of-hearing and requires a voice stream plus a text
        stream
        <br>
        &nbsp;&nbsp;&nbsp; (known as voice carry over).&nbsp; Making language a media
        attribute
        <br>
        &nbsp;&nbsp;&nbsp; allows the standard session negotiation mechanism to handle
        this by
        <br>
        &nbsp;&nbsp;&nbsp; providing the information and mechanism for the endpoints to
        make
        <br>
        &nbsp;&nbsp;&nbsp; appropriate decisions.
        <br>
        <br>
        &nbsp;This scenario suggests that there might be dependency or
        interaction
        <br>
        &nbsp;between language specifications for different media
        descriptions.
        <br>
        &nbsp;Whether this is needed should be determined and documented.
        <br>
        <br>
        &nbsp;- Specifying preference levels?
        <br>
        <br>
        &nbsp;&nbsp;&nbsp; For example, some users may be able to speak several
        languages, but
        <br>
        &nbsp;&nbsp;&nbsp; have a preference.
        <br>
        <br>
        &nbsp;This might argue for describing degrees of preference using "q"
        <br>
        &nbsp;parameters (as in the SIP Accept-Language header).
        <br>
        <br>
        &nbsp;- Expressing multiple languages in answers
        <br>
        <br>
        &nbsp;&nbsp;&nbsp; (While it is true that a conversation among multilingual
        people
        <br>
        &nbsp;&nbsp;&nbsp; often involves multiple languages, it does not seem useful
        enough
        <br>
        &nbsp;&nbsp;&nbsp; as a general facility to warrant complicating the desired
        semantics
        <br>
        &nbsp;&nbsp;&nbsp; of the SDP attribute to allow negotiation of multiple
        simultaneous
        <br>
        &nbsp;&nbsp;&nbsp; languages within an interactive media stream.)
        <br>
        <br>
        &nbsp;Why shouldn't an answer be able to indicate multiple
        languages?&nbsp; At
        <br>
        &nbsp;the least, this might provide the offerer with useful
        information.
        <br>
      </blockquote>
      <br>
      You raise good questions that I think need more discussion.&nbsp; I am
      hoping to keep the work as simple as possible and not add
      additional complexity, which argues for not solving every aspect
      of the problem, but only those that must be solved immediately.
      <br>
      <br>
      <blockquote type="cite">
        <br>
        &nbsp;- Reusing a=lang
        <br>
        <br>
        &nbsp;Searching, I can only find these descriptions of the use of
        <br>
        &nbsp;"a=lang:...":
        <br>
        <br>
        &nbsp;&nbsp;&nbsp;&nbsp; RFC 4566
        <br>
        &nbsp;&nbsp;&nbsp;&nbsp; draft-saintandre-sip-xmpp-chat
        <br>
        &nbsp;&nbsp;&nbsp;&nbsp; draft-gellens-negotiating-human-language
        <br>
        <br>
        &nbsp;So it looks like "a=lang:..." is entirely unused at the present
        and is
        <br>
        &nbsp;safe to be redefined.
        <br>
      </blockquote>
      <br>
      <br>
      <br>
      <br>
      <br>
    </blockquote>
    <br>
  </body>
</html>

--------------040800080507030709090604--
