Re: [dc] draft-khasnabish-vmmi-problems-00.txt

Vishwas Manral <vishwas.ietf@gmail.com> Sat, 21 January 2012 01:23 UTC

Return-Path: <vishwas.ietf@gmail.com>
X-Original-To: dc@ietfa.amsl.com
Delivered-To: dc@ietfa.amsl.com
Received: from localhost (localhost [127.0.0.1]) by ietfa.amsl.com (Postfix) with ESMTP id 97EF021F86DE for <dc@ietfa.amsl.com>; Fri, 20 Jan 2012 17:23:28 -0800 (PST)
X-Virus-Scanned: amavisd-new at amsl.com
X-Spam-Flag: NO
X-Spam-Score: -3.398
X-Spam-Level:
X-Spam-Status: No, score=-3.398 tagged_above=-999 required=5 tests=[AWL=0.201, BAYES_00=-2.599, RCVD_IN_DNSWL_LOW=-1]
Received: from mail.ietf.org ([12.22.58.30]) by localhost (ietfa.amsl.com [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id WmsfHigal1qA for <dc@ietfa.amsl.com>; Fri, 20 Jan 2012 17:23:27 -0800 (PST)
Received: from mail-tul01m020-f172.google.com (mail-tul01m020-f172.google.com [209.85.214.172]) by ietfa.amsl.com (Postfix) with ESMTP id A3FA221F86C5 for <dc@ietf.org>; Fri, 20 Jan 2012 17:23:27 -0800 (PST)
Received: by obbwc12 with SMTP id wc12so1683254obb.31 for <dc@ietf.org>; Fri, 20 Jan 2012 17:23:27 -0800 (PST)
DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc:content-type:content-transfer-encoding; bh=c6Pe4EhcAX2ergyuqHvShkvjc86zp7OBUrpYIcO+/Qs=; b=BlscJP9RLsdR6ZiaKqHI2Y1qi9GE4DkMiDnDroMlr6b7EMIxGZPvqbAiv1rbS/pMqQ 3ejXcvUmjgW4gx8Z70a7s3NT9a3qm55E8aphlY/uTrsAZScai1wcO2Y/NbyflDS1njbG wJfLfv5qqyZlJvdO0eWuEnZCMZ9Urfa1UCmnY=
MIME-Version: 1.0
Received: by 10.182.222.102 with SMTP id ql6mr149213obc.2.1327109005926; Fri, 20 Jan 2012 17:23:25 -0800 (PST)
Received: by 10.182.28.196 with HTTP; Fri, 20 Jan 2012 17:23:25 -0800 (PST)
In-Reply-To: <FF8EC204-C4B0-4690-B692-905F672D60D3@asgaard.org>
References: <CAH==cJxfmae0u0bSF4cn_haLgY1T-vnw2102PApzYtj5Aty=GQ@mail.gmail.com> <CANtnpwhFJ746ooi9GUCxfBqsOXu14hDka0D9inhh5pPq3U_ZTA@mail.gmail.com> <201201171540.q0HFeNan008591@cichlid.raleigh.ibm.com> <CANtnpwjexDPazOXLYHHjn3+JDi-o49Bv5ptDExAZHAA8Ra2m-A@mail.gmail.com> <201201191419.q0JEJTLF010649@cichlid.raleigh.ibm.com> <1326989277.2513.4.camel@ecliptic.extremenetworks.com> <618BE8B40039924EB9AED233D4A09C5102CB2291@XMB-BGL-416.cisco.com> <406B8B5D-E1E5-4DF4-8DE2-D7D2A699430A@asgaard.org> <4F18CE61.6030002@gmail.com> <618BE8B40039924EB9AED233D4A09C5102CB2330@XMB-BGL-416.cisco.com> <4F18EF4A.3060308@gmail.com> <618BE8B40039924EB9AED233D4A09C5102CB234C@XMB-BGL-416.cisco.com> <4F18FB72.2090900@joelhalpern.com> <618BE8B40039924EB9AED233D4A09C5102CB2380@XMB-BGL-416.cisco.com> <4F19034E.1070802@gmail.com> <CAOyVPHTbxB=QYC3Qw0ybL=5RN7VefSENV4iiBBOpXbCn58oi=Q@mail.gmail.com> <4F19F939.2020804@gmail.com> <DF0D6664-9FD5-4EF0-A03F-86C1921D9D01@asgaard.org> <CAOyVPHQh2yb5iP9-bH6NOzamW6FaK0cYwpfqfqns7TZVTpmY5g@mail.gmail.com> <FF8EC204-C4B0-4690-B692-905F672D60D3@asgaard.org>
Date: Fri, 20 Jan 2012 17:23:25 -0800
Message-ID: <CAOyVPHQPBmhpKKLs=GnaOe9EwJMB0pSBAuS458mNDo8J6tC3Lw@mail.gmail.com>
From: Vishwas Manral <vishwas.ietf@gmail.com>
To: Christopher LILJENSTOLPE <cdl@asgaard.org>
Content-Type: text/plain; charset="UTF-8"
Content-Transfer-Encoding: quoted-printable
Cc: Melinda Shore <melinda.shore@gmail.com>, dc@ietf.org
Subject: Re: [dc] draft-khasnabish-vmmi-problems-00.txt
X-BeenThere: dc@ietf.org
X-Mailman-Version: 2.1.12
Precedence: list
List-Id: IETF Data Center Mailing List <dc.ietf.org>
List-Unsubscribe: <https://www.ietf.org/mailman/options/dc>, <mailto:dc-request@ietf.org?subject=unsubscribe>
List-Archive: <http://www.ietf.org/mail-archive/web/dc>
List-Post: <mailto:dc@ietf.org>
List-Help: <mailto:dc-request@ietf.org?subject=help>
List-Subscribe: <https://www.ietf.org/mailman/listinfo/dc>, <mailto:dc-request@ietf.org?subject=subscribe>
X-List-Received-Date: Sat, 21 Jan 2012 01:23:28 -0000

Hi Chris,

It is not about crashing alone, it could be a wrong entry added into
the table that could cause the packets to be routed wrongly. The idea
is about increasing the complexity in a single point of failure.

I guess this is not the most critical discussion and we can probably
move forward.

Thanks,
Vishwas


On 1/20/12, Christopher LILJENSTOLPE <cdl@asgaard.org> wrote:
> Greetings Vishwas,
>
> 	And I guess I am saying that I'm not sure I agree.  If a single process
> running can kill all the other processes on a given system (especially if it
> is in "user space" - which I would assume for control plane functions), I
> would say that there are more substantial issues with the architecture of
> that particular VM distribution.  Similar answer if someone actually put a
> control-plane engine (think routing protocols) in kernel space.  We, as a
> standards organization can't save the world from less-than-inteligent
> developers....
>
>
> 	Chris
>
> On 20Jan2012, at 15.47, Vishwas Manral wrote:
>
>> Hi Chistopher,
>>
>> I totally agree.
>>
>> The point I was making was that the hypervisor is the single point of
>> failure which can cause all the guest OS to fail. The more complex you
>> make the functionality the higer the chances of failure. So we should
>> work with that thought in mind.
>>
>> Thanks,
>> Vishwas
>>
>> On 1/20/12, Christopher LILJENSTOLPE <cdl@asgaard.org> wrote:
>>> Greetings,
>>>
>>> 	Another way of looking at it is a hypervisor is really an operating
>>> system
>>> distribution.  Many of the hypervisors out there have the network
>>> "switch"
>>> as a separate process (actually I believe all of them do).  So, if we are
>>> saying that networking intel doesn't belong in an OS distribution, that
>>> is a
>>> departure from current thinking :)
>>>
>>> 	Chris
>>>
>>> On 20Jan2012, at 15.31, Melinda Shore wrote:
>>>
>>>> On 01/20/2012 02:18 PM, Vishwas Manral wrote:
>>>>> An interesting thing to note is that the more the functionality you
>>>>> put in the hypervisor, the more you stress the single point of failure
>>>>> in the virtualized system.
>>>>
>>>> Well, there are a few ways to look at it.  For example,
>>>> the fewer components you've got the larger the mean time
>>>> between failures.  But aside from that it's been a
>>>> general rule of thumb that you want to minimize the
>>>> impact of failed components on non-failed components
>>>> (the fate sharing principle).
>>>>
>>>> At any rate the hypervisor (at least the ones with which
>>>> I'm familiar) basically *are* network devices - they
>>>> function as a switch, or even a NAT.  If you're going to
>>>> suggest that they can't be used to terminate control plane
>>>> sessions I hope there's a more compelling reason for it
>>>> than what has been offered so far.
>>>>
>>>> Melinda
>>>> _______________________________________________
>>>> dc mailing list
>>>> dc@ietf.org
>>>> https://www.ietf.org/mailman/listinfo/dc
>>>
>>> --
>>> 李柯睿
>>> Check my PGP key here: https://www.asgaard.org/~cdl/cdl.asc
>>> Current vCard here: https://www.asgaard.org/~cdl/cdl.vcf
>>>
>>>
>
> --
> 李柯睿
> Check my PGP key here: https://www.asgaard.org/~cdl/cdl.asc
> Current vCard here: https://www.asgaard.org/~cdl/cdl.vcf
>
>