All of lore.kernel.org
 help / color / mirror / Atom feed
From: Liu Yuan <namei.unix@gmail.com>
To: Badari Pulavarty <pbadari@us.ibm.com>
Cc: kvm@vger.kernel.org, Dongsu Park <dongsu.park@profitbricks.com>
Subject: Re: [RFC PATCH]vhost-blk: In-kernel accelerator for virtio block device
Date: Tue, 16 Aug 2011 13:44:13 +0800	[thread overview]
Message-ID: <4E4A03AD.8000806@gmail.com> (raw)
In-Reply-To: <4E489DF7.3050707@us.ibm.com>

On 08/15/2011 12:17 PM, Badari Pulavarty wrote:
> On 8/14/2011 8:20 PM, Liu Yuan wrote:
>> On 08/13/2011 12:12 AM, Badari Pulavarty wrote:
>>> On 8/12/2011 4:40 AM, Liu Yuan wrote:
>>>> On 08/12/2011 04:27 PM, Liu Yuan wrote:
>>>>> On 08/12/2011 12:50 PM, Badari Pulavarty wrote:
>>>>>> On 8/10/2011 8:19 PM, Liu Yuan wrote:
>>>>>>> On 08/11/2011 11:01 AM, Liu Yuan wrote:
>>>>>>>>
>>>>>>>>> It looks like the patch wouldn't work for testing multiple 
>>>>>>>>> devices.
>>>>>>>>>
>>>>>>>>> vhost_blk_open() does
>>>>>>>>> +       used_info_cachep = KMEM_CACHE(used_info, 
>>>>>>>>> SLAB_HWCACHE_ALIGN |
>>>>>>>>> SLAB_PANIC);
>>>>>>>>>
>>>>>>>>
>>>>>>>> This is weird. how do you open multiple device?I just opened 
>>>>>>>> the device with following command:
>>>>>>>>
>>>>>>>> -drive file=/dev/sda6,if=virtio,cache=none,aio=native -drive 
>>>>>>>> file=~/data0.img,if=virtio,cache=none,aio=native -drive 
>>>>>>>> file=~/data1.img,if=virtio,cache=none,aio=native
>>>>>>>>
>>>>>>>> And I didn't meet any problem.
>>>>>>>>
>>>>>>>> this would tell qemu to open three devices, and pass three FDs 
>>>>>>>> to three instances of vhost_blk module.
>>>>>>>> So KMEM_CACHE() is okay in vhost_blk_open().
>>>>>>>>
>>>>>>>
>>>>>>> Oh, you are right. KMEM_CACHE() is in the wrong place. it is 
>>>>>>> three instances vhost worker threads created. Hmmm, but I didn't 
>>>>>>> meet any problem when opening it and running it. So strange. 
>>>>>>> I'll go to figure it out.
>>>>>>>
>>>>>>>>> When opening second device, we get panic since 
>>>>>>>>> used_info_cachep is
>>>>>>>>> already created. Just to make progress I moved this call to
>>>>>>>>> vhost_blk_init().
>>>>>>>>>
>>>>>>>>> I don't see any host panics now. With single block device (dd),
>>>>>>>>> it seems to work fine. But when I start testing multiple block
>>>>>>>>> devices I quickly run into hangs in the guest. I see following
>>>>>>>>> messages in the guest from virtio_ring.c:
>>>>>>>>>
>>>>>>>>> virtio_blk virtio2: requests: id 0 is not a head !
>>>>>>>>> virtio_blk virtio1: requests: id 0 is not a head !
>>>>>>>>> virtio_blk virtio4: requests: id 1 is not a head !
>>>>>>>>> virtio_blk virtio3: requests: id 39 is not a head !
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Badari
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>> vq->data[] is initialized by guest virtio-blk driver and 
>>>>>>>> vhost_blk is unware of it. it looks like used ID passed
>>>>>>>> over by vhost_blk to guest virtio_blk is wrong, but, it should 
>>>>>>>> not happen. :|
>>>>>>>>
>>>>>>>> And I can't reproduce this on my laptop. :(
>>>>>>>>
>>>>>> Finally, found the issue  :)
>>>>>>
>>>>>> Culprit is:
>>>>>>
>>>>>> +static struct io_event events[MAX_EVENTS];
>>>>>>
>>>>>> With multiple devices, multiple threads could be executing 
>>>>>> handle_completion() (one for
>>>>>> each fd) at the same time. "events" array is global :( Need to 
>>>>>> make it one per device/fd.
>>>>>>
>>>>>> For test, I changed MAX_EVENTS to 32 and moved "events" array to 
>>>>>> be local (stack)
>>>>>> to handle_completion(). Tests are running fine.
>>>>>>
>>>>>> Your laptop must have single processor, hence you have only one 
>>>>>> thread executing handle_completion()
>>>>>> at any time..
>>>>>>
>>>>>> Thanks,
>>>>>> Badari
>>>>>>
>>>>>>
>>>>> Good catch, this is rather cool!....Yup, I develop it mostly in a 
>>>>> nested KVM environment. and the L2 host  only runs single 
>>>>> processor :(
>>>>>
>>>>> Thanks,
>>>>> Yuan
>>>> By the way, MAX_EVENTS should be 128, as much as guest virtio_blk 
>>>> driver can batch-submit,
>>>> causing array overflow.
>>>> I have had turned on the debug, and had seen as much as over 100 
>>>> requests batched from guest OS.
>>>>
>>>
>>> Hmm.. I am not sure why you see over 100 outstanding events per fd.  
>>> Max events could be as high as
>>> number of number of outstanding IOs.
>>>
>>> Anyway, instead of putting it on stack, I kmalloced it now.
>>>
>>> Dongsu Park, Here is the complete patch.
>>>
>>> Thanks
>>> Badari
>>>
>>>
>> In the physical machine, there is a queue depth posted by block 
>> device driver to limit the
>> pending requests number, normally it is 31. But virtio driver doesn't 
>> post it in the guest OS.
>> So nothing prvents OS batch-submitting requests more than 31.
>>
>> I have noticed over 100 pending requests during guest OS 
>> initilization and it is reproducible.
>>
>> BTW, how is perf number for vhost-blk in your environment?
>
> Right now I am doing "dd" tests to test out the functionality and 
> stability.
>
> I plan to collect FFSB benchmark results across 6-virtio-blk/vhost-blk 
> disks with
> all profiles - seq read, seq write, random read, random write with 
> blocksizes varying
> from 4k to 1MB.
>
> I will start the test tomorrow. It will take few days to run thru all 
> the scenarios.
> I don't have an easy way to collect host CPU consumption - but for now 
> lets
> focus on throughput and latency. I will share the results in few days.
>
> Thanks
> Badari
>
>

Awesome! Thanks for your work and data.

Yuan

  reply	other threads:[~2011-08-16  5:44 UTC|newest]

Thread overview: 54+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2011-07-28 14:29 [RFC PATCH]vhost-blk: In-kernel accelerator for virtio block device Liu Yuan
2011-07-28 14:29 ` [RFC PATCH] vhost-blk: An in-kernel accelerator for virtio-blk Liu Yuan
2011-07-28 14:47   ` Christoph Hellwig
2011-07-29 11:19     ` Liu Yuan
2011-07-28 15:18   ` Stefan Hajnoczi
2011-07-28 15:22   ` Michael S. Tsirkin
2011-07-29 15:09     ` Liu Yuan
2011-08-01  6:25     ` Liu Yuan
2011-08-01  8:12       ` Michael S. Tsirkin
2011-08-01  8:55         ` Liu Yuan
2011-08-01 10:26           ` Michael S. Tsirkin
2011-08-11 19:59     ` Dongsu Park
2011-08-12  8:56       ` Alan Cox
2011-07-28 14:29 ` [RFC PATCH] vhost: Enable vhost-blk support Liu Yuan
2011-07-28 15:44 ` [RFC PATCH]vhost-blk: In-kernel accelerator for virtio block device Stefan Hajnoczi
2011-07-29  4:48   ` Stefan Hajnoczi
2011-07-29  7:59     ` Liu Yuan
2011-07-29 10:55       ` Christoph Hellwig
2011-07-29  7:22   ` Liu Yuan
2011-07-29  9:06     ` Stefan Hajnoczi
2011-07-29 12:01       ` Liu Yuan
2011-07-29 12:29         ` Stefan Hajnoczi
2011-07-29 12:50           ` Stefan Hajnoczi
2011-07-29 14:45             ` Liu Yuan
2011-07-29 14:50               ` Liu Yuan
2011-07-29 15:25         ` Sasha Levin
2011-08-01  8:17           ` Avi Kivity
2011-08-01  9:18             ` Liu Yuan
2011-08-01  9:37               ` Avi Kivity
2011-07-29 18:12     ` Badari Pulavarty
2011-08-01  5:46       ` Liu Yuan
2011-08-01  8:12         ` Christoph Hellwig
2011-08-04 21:58         ` Badari Pulavarty
2011-08-05  7:56           ` Liu Yuan
2011-08-05 11:04           ` Liu Yuan
2011-08-05 18:02             ` Badari Pulavarty
2011-08-08  1:35               ` Liu Yuan
2011-08-08  5:04                 ` Badari Pulavarty
2011-08-08  7:31                   ` Liu Yuan
2011-08-08 17:16                     ` Badari Pulavarty
2011-08-10  2:19                       ` Liu Yuan
2011-08-10 20:37                         ` Badari Pulavarty
2011-08-11  3:01                           ` Liu Yuan
2011-08-11  3:19                             ` Liu Yuan
2011-08-11 23:51                               ` Badari Pulavarty
2011-08-12  4:50                               ` Badari Pulavarty
2011-08-12  6:46                                 ` Dongsu Park
2011-08-12  8:27                                 ` Liu Yuan
2011-08-12 11:40                                   ` Liu Yuan
2011-08-12 16:12                                     ` Badari Pulavarty
2011-08-15  3:20                                       ` Liu Yuan
2011-08-15  4:17                                         ` Badari Pulavarty
2011-08-16  5:44                                           ` Liu Yuan [this message]
2011-09-07 13:36                                           ` Liu Yuan

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4E4A03AD.8000806@gmail.com \
    --to=namei.unix@gmail.com \
    --cc=dongsu.park@profitbricks.com \
    --cc=kvm@vger.kernel.org \
    --cc=pbadari@us.ibm.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.