All of lore.kernel.org
 help / color / mirror / Atom feed
From: Adam Chang <adamchang at qnap.com>
To: spdk@lists.01.org
Subject: Re: [SPDK] Error when issue IO in QEMU to vhost scsi NVMe
Date: Fri, 10 Aug 2018 13:01:41 +0800	[thread overview]
Message-ID: <CANvoUxh4t8=DPfdP2-3SZ4Lo8HMWqUnDjqfRTgZ2wrU9XS9s+A@mail.gmail.com> (raw)
In-Reply-To: FBE7E039FA50BF47A673AD0BD3CD56A8461B0C4F@HASMSX106.ger.corp.intel.com

[-- Attachment #1: Type: text/plain, Size: 86824 bytes --]

Hi:
Here are My host environment
==================================================================
Host OS: Ubuntu 18.04 x86_64
Linux Kernel: 4.15.0-30
CPU: Intel i7 8700K
Memory: 32GB
NVME SSD: Intel Optane Memory 32GB
==================================================================
configuration for building QEMU:
==================================================================
./configure --prefix=/usr --target-list=x86_64-softmmu --enable-kvm
--enable-debug --enable-debug-info --enable-modules --enable-linux-aio
--enable-vnc --enable-trace-backends=log --enable-numa --disable-werror
--disable-strip --with-sdlabi=2.0
==================================================================

configuration for building SPDK:
==================================================================
./configure --enable-debug
==================================================================

I checked the dmesg, it showed the following error
==================================================================
[ 4640.033876] DMAR: intel_iommu_map: iommu width (39) is not sufficient
for the mapped address (7fdce8200000)
==================================================================

I add log option when start the vhost target as follow:
==================================================================
./app/vhost/vhost -S /var/tmp -m 0x3 -L vhost vhost_scsi &
==================================================================

And here are my vhost log
==================================================================
 VHOST_CONFIG: new vhost user connection is 18
VHOST_CONFIG: new device, handle is 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_GET_FEATURES
VHOST_CONFIG: /var/tmp/vhost.0: read message
VHOST_USER_GET_PROTOCOL_FEATURES
VHOST_CONFIG: /var/tmp/vhost.0: read message
VHOST_USER_SET_PROTOCOL_FEATURES
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_GET_QUEUE_NUM
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_OWNER
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_GET_FEATURES
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_CALL
VHOST_CONFIG: vring call idx:0 file:25
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_CALL
VHOST_CONFIG: vring call idx:1 file:26
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_CALL
VHOST_CONFIG: vring call idx:2 file:27
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_FEATURES
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_NUM
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_BASE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_ADDR
VHOST_CONFIG: guest memory region 0, size: 0x40000000
guest physical addr: 0x0
guest virtual  addr: 0x7f4754600000
host  virtual  addr: 0x7fdce8000000
mmap addr : 0x7fdce8000000
mmap size : 0x40000000
mmap align: 0x200000
mmap off  : 0x0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_KICK
VHOST_CONFIG: vring kick idx:2 file:29
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
Cannot set up DMA mapping, error 14
vhost.c: 541:spdk_vhost_dev_mem_register: *WARNING*: Failed to register
memory region 0. Future vtophys translation might fail.
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_CALL
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: vring call idx:0 file:30
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_CALL
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: vring call idx:1 file:25
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_CALL
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: vring call idx:2 file:26
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_GET_VRING_BASE
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: vring base idx:2 file:259
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_FEATURES
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_MEM_TABLE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_NUM
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_BASE
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_ADDR
VHOST_CONFIG: guest memory region 0, size: 0x40000000
guest physical addr: 0x0
guest virtual  addr: 0x7f4754600000
host  virtual  addr: 0x7fdce8000000
mmap addr : 0x7fdce8000000
mmap size : 0x40000000
mmap align: 0x200000
mmap off  : 0x0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_KICK
VHOST_CONFIG: vring kick idx:0 file:27
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_NUM
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_BASE
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_ADDR
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_KICK
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: vring kick idx:1 file:28
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_NUM
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_BASE
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_ADDR
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_KICK
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: vring kick idx:2 file:29
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_CALL
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: vring call idx:0 file:31
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_CALL
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: vring call idx:1 file:30
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_CALL
vhost_scsi.c:1141:destroy_device_poller_cb: *INFO*: Stopping poller for
vhost controller vhost.0
VHOST_CONFIG: vring call idx:2 file:25
VHOST_CONFIG: virtio is now ready for processing.
vhost.c: 537:spdk_vhost_dev_mem_register: *INFO*: Registering VM memory for
vtophys translation - 0x7fdce8000000 len:0x40000000
vhost_scsi.c:1099:spdk_vhost_scsi_start: *INFO*: Started poller for vhost
controller vhost.0 on lcore 0
nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
vtophys(0x7fdd1c467000) failed
nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95
nsid:1 lba:0 len:8
nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
(00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
bdev_nvme.c:1511:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
vtophys(0x7fdd1c467000) failed
nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95
nsid:1 lba:0 len:8
nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
(00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
bdev_nvme.c:1511:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
==================================================================

Thanks,
Adam Chang.

On Thu, Aug 9, 2018 at 9:55 PM Stojaczyk, DariuszX <
dariuszx.stojaczyk(a)intel.com> wrote:

> Thanks,
> The address that vtophys fails on should be mapped. Something went wrong,
> but that vhost log is not particularly helpful because it comes from a
> non-debug app.
> I could ask you to enable debug (./configure --enable-debug), but could
> you frst provide the dmesg? Are there any errors?
> D.
>
> > -----Original Message-----
> > From: SPDK [mailto:spdk-bounces(a)lists.01.org] On Behalf Of Adam Chang
> > Sent: Thursday, August 9, 2018 12:56 PM
> > To: Storage Performance Development Kit <spdk(a)lists.01.org>
> > Subject: Re: [SPDK] Error when issue IO in QEMU to vhost scsi NVMe
> >
> > Hi:
> > I have added "-numa node,memdev=mem0" in QEMU command line, but still had
> > same error message.
> > Here are my modified QEMU command argument
> >
> >
> >  ==================================================================
> > taskset -c 2,3,4,5 qemu-system-x86_64 -enable-kvm -m 1G \
> >         -name bread,debug-threads=on \
> >         -daemonize \
> >         -pidfile /var/log/bread.pid \
> >         -cpu host\
> >         -smp 4,sockets=1,cores=4,threads=1 \
> >         -object memory-backend-file,id=mem0,size=1G,mem-
> > path=/dev/hugepages,share=on,prealloc=yes,host-nodes=0,policy=bind -numa
> > node,memdev=mem0\
> >         -drive
> > file=../ubuntu.img,media=disk,cache=unsafe,aio=threads,format=qcow2\
> > -chardev socket,id=char0,path=/var/tmp/vhost.0 \
> > -device vhost-user-scsi-pci,id=scsi0,chardev=char0\
> >         -machine usb=on \
> >         -device usb-tablet \
> >         -device usb-mouse \
> >         -device usb-kbd \
> >         -vnc :2 \
> > -net nic,model=virtio\
> > -net user,hostfwd=tcp::2222-:22
> >  ==================================================================
> >
> > And the following is the vhost log from QEMU starting:
> > ==================================================================
> > VHOST_CONFIG: new vhost user connection is 18
> > VHOST_CONFIG: new device, handle is 0
> > VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_GET_FEATURES
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_GET_PROTOCOL_FEATURES
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_PROTOCOL_FEATURES
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_GET_QUEUE_NUM
> > VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_OWNER
> > VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_GET_FEATURES
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_CALL
> > VHOST_CONFIG: vring call idx:0 file:25
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_CALL
> > VHOST_CONFIG: vring call idx:1 file:26
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_CALL
> > VHOST_CONFIG: vring call idx:2 file:27
> > VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_FEATURES
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_NUM
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_BASE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_ADDR
> > VHOST_CONFIG: guest memory region 0, size: 0x40000000
> > guest physical addr: 0x0
> > guest virtual  addr: 0x7fa1a4a00000
> > host  virtual  addr: 0x7f8fb4000000
> > mmap addr : 0x7f8fb4000000
> > mmap size : 0x40000000
> > mmap align: 0x200000
> > mmap off  : 0x0
> > VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_KICK
> > VHOST_CONFIG: vring kick idx:2 file:29
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_CALL
> > VHOST_CONFIG: vring call idx:0 file:30
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_CALL
> > VHOST_CONFIG: vring call idx:1 file:25
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_CALL
> > VHOST_CONFIG: vring call idx:2 file:26
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_GET_VRING_BASE
> > VHOST_CONFIG: vring base idx:2 file:259
> > VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_FEATURES
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_MEM_TABLE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_NUM
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_BASE
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_ADDR
> > VHOST_CONFIG: guest memory region 0, size: 0x40000000
> > guest physical addr: 0x0
> > guest virtual  addr: 0x7fa1a4a00000
> > host  virtual  addr: 0x7f8fb4000000
> > mmap addr : 0x7f8fb4000000
> > mmap size : 0x40000000
> > mmap align: 0x200000
> > mmap off  : 0x0
> > VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_KICK
> > VHOST_CONFIG: vring kick idx:0 file:27
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_NUM
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_BASE
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_ADDR
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_KICK
> > VHOST_CONFIG: vring kick idx:1 file:28
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_NUM
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_BASE
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_ADDR
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message VHOST_USER_SET_VRING_KICK
> > VHOST_CONFIG: vring kick idx:2 file:29
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_CALL
> > VHOST_CONFIG: vring call idx:0 file:31
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_CALL
> > VHOST_CONFIG: vring call idx:1 file:30
> > VHOST_CONFIG: virtio is now ready for processing.
> > VHOST_CONFIG: /var/tmp/vhost.0: read message
> > VHOST_USER_SET_VRING_CALL
> > VHOST_CONFIG: vring call idx:2 file:25
> > VHOST_CONFIG: virtio is now ready for processing.
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc8000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:24 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc8000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:24 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc8000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:24 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc8000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:24 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc8000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:24 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc8000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:24 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7fc9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:0 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe7f2a000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> > vtophys(0x7f8fe98a9000) failed
> > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1
> > cid:95 nsid:1 lba:57149312 len:8
> > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID FIELD
> > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc = -22
> >
> > ==================================================================
> >
> > Thanks,
> > Adam Chang.
> >
> > On Thu, Aug 9, 2018 at 4:07 PM Stojaczyk, DariuszX
> > <dariuszx.stojaczyk(a)intel.com <mailto:dariuszx.stojaczyk(a)intel.com> >
> wrote:
> >
> >
> >       Can you provide a full vhost log?
> >       D.
> >
> >       > -----Original Message-----
> >       > From: SPDK [mailto:spdk-bounces(a)lists.01.org <mailto:spdk-
> > bounces(a)lists.01.org> ] On Behalf Of Adam Chang
> >       > Sent: Thursday, August 9, 2018 4:05 AM
> >       > To: spdk(a)lists.01.org <mailto:spdk(a)lists.01.org>
> >       > Subject: [SPDK] Error when issue IO in QEMU to vhost scsi NVMe
> >       >
> >       > Hi all:
> >       >       I just create NVMe bdev and vhost-scsi controller which
> can be
> > accessed by
> >       > QEMU, but it occurred error when IO issued from VM.
> >       >       Here are my steps for SPDK configuration
> >       >
> >       > Host OS:Ubuntu 18.04, Kernel 4.15.0-30
> >       > Guest OS: Ubuntu 18.04
> >       > QEMU: 2.12.0
> >       > SPDK: v18.07
> >       >
> >       > 1)  sudo HUGEMEM=4096 scripts/setup.sh
> >       >
> >       > 0000:05:00.0 (8086 2522): nvme -> vfio-pci
> >       >
> >       > Current user memlock limit: 4116 MB
> >       >
> >       > This is the maximum amount of memory you will be
> >       > able to use with DPDK and VFIO if run as current user.
> >       > To change this, please adjust limits.conf memlock limit for
> current user.
> >       >
> >       > 2) sudo ./app/vhost/vhost -S /var/tmp -m 0x3 &
> >       >
> >       > [ DPDK EAL parameters: vhost -c 0x3 -m 1024 --legacy-mem --file-
> >       > prefix=spdk_pid1921 ]
> >       > EAL: Detected 12 lcore(s)
> >       > EAL: Detected 1 NUMA nodes
> >       > EAL: Multi-process socket /var/run/dpdk/spdk_pid1921/mp_socket
> >       > EAL: No free hugepages reported in hugepages-1048576kB
> >       > EAL: Probing VFIO support...
> >       > EAL: VFIO support initialized
> >       > app.c: 530:spdk_app_start: *NOTICE*: Total cores available: 2
> >       > reactor.c: 718:spdk_reactors_init: *NOTICE*: Occupied cpu socket
> mask
> > is 0x1
> >       > reactor.c: 492:_spdk_reactor_run: *NOTICE*: Reactor started on
> core 1
> > on socket
> >       > 0
> >       > reactor.c: 492:_spdk_reactor_run: *NOTICE*: Reactor started on
> core 0
> > on socket
> >       > 0
> >       >
> >       > 3) sudo ./scripts/rpc.py construct_vhost_scsi_controller
> --cpumask 0x1
> > vhost.0
> >       > EAL: PCI device 0000:05:00.0 on NUMA socket 0
> >       > EAL:   probe driver: 8086:2522 spdk_nvme
> >       > EAL:   using IOMMU type 1 (Type 1)
> >       > Nvme0n1
> >       >
> >       > 4) sudo ./scripts/rpc.py add_vhost_scsi_lun vhost.0 0 Nvme0n1
> >       > 5) start qemu:
> >       > taskset qemu-system-x86_64 -enable-kvm -m 1G \
> >       >         -name bread,debug-threads=on \
> >       >         -daemonize \
> >       >         -pidfile /var/log/bread.pid \
> >       >         -cpu host\
> >       >         -smp 4,sockets=1,cores=4,threads=1 \
> >       >         -object memory-backend-file,id=mem0,size=1G,mem-
> >       > path=/dev/hugepages,share=on -numa node,memdev=mem0\
> >       >         -drive
> >       >
> > file=../ubuntu.img,media=disk,cache=unsafe,aio=threads,format=qcow2\
> >       > -chardev socket,id=char0,path=/var/tmp/vhost.0 \
> >       > -device vhost-user-scsi-pci,id=scsi0,chardev=char0\
> >       >         -machine usb=on \
> >       >         -device usb-tablet \
> >       >         -device usb-mouse \
> >       >         -device usb-kbd \
> >       >         -vnc :2 \
> >       >     -net nic,model=virtio\
> >       >     -net user,hostfwd=tcp::2222-:22
> >       >
> >       > then when I use fio to test the vhost nvme disk in guest VM, I
> got the
> > following
> >       > error message in host console.
> >       >
> > ==================================================================
> >       > =========
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:32
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:32
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:32
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:32
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:32
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:32
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:8
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:8
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:8
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:8
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:8
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       > nvme_pcie.c:1706:nvme_pcie_prp_list_append: *ERROR*:
> >       > vtophys(0x7f8fed64d000) failed
> >       > nvme_qpair.c: 137:nvme_io_qpair_print_command: *NOTICE*: READ
> > sqid:1
> >       > cid:95 nsid:1 lba:0 len:8
> >       > nvme_qpair.c: 306:nvme_qpair_print_completion: *NOTICE*: INVALID
> > FIELD
> >       > (00/02) sqid:1 cid:95 cdw0:0 sqhd:0000 p:0 m:0 dnr:1
> >       > bdev_nvme.c:1521:bdev_nvme_queue_cmd: *ERROR*: readv failed: rc
> > = -22
> >       >
> > ==================================================================
> >       > =========
> >       >
> >       > I  used the lsblk to check block device information in guest,
> and could
> > see the
> >       > nvme disk with sdb.
> >       > >lsblk --output
> > "NAME,KNAME,MODEL,HCTL,SIZE,VENDOR,SUBSYSTEMS"
> >       >
> > ==================================================================
> >       > =========
> >       >
> >       > NAME   KNAME  MODEL            HCTL         SIZE VENDOR
>  SUBSYSTEMS
> >       > fd0    fd0                                    4K
> block:platform
> >       > loop0  loop0                               12.2M          block
> >       > loop1  loop1                               86.6M          block
> >       > loop2  loop2                                1.6M          block
> >       > loop3  loop3                                3.3M          block
> >       > loop4  loop4                                 21M          block
> >       > loop5  loop5                                2.3M          block
> >       > loop6  loop6                                 13M          block
> >       > loop7  loop7                                3.7M          block
> >       > loop8  loop8                                2.3M          block
> >       > loop9  loop9                               86.9M          block
> >       > loop10 loop10                              34.7M          block
> >       > loop11 loop11                                87M          block
> >       > loop12 loop12                             140.9M          block
> >       > loop13 loop13                                13M          block
> >       > loop14 loop14                               140M          block
> >       > loop15 loop15                             139.5M          block
> >       > loop16 loop16                               3.7M          block
> >       > loop17 loop17                              14.5M          block
> >       > sda    sda    QEMU HARDDISK    0:0:0:0       32G ATA
> block:scsi:pci
> >       >   sda1 sda1                                  32G
> block:scsi:pci
> >       > sdb    sdb    NVMe disk        2:0:0:0     27.3G INTEL
> block:scsi:virtio:pci
> >       > sr0    sr0    QEMU DVD-ROM     1:0:0:0     1024M QEMU
>  block:scsi:pci
> >       >
> > ==================================================================
> >       > =========
> >       >
> >       >
> >       > Does anyone can give me help how to solve this problem ?
> >       >
> >       > Thanks.
> >       > Adam Chang
> >       _______________________________________________
> >       SPDK mailing list
> >       SPDK(a)lists.01.org <mailto:SPDK(a)lists.01.org>
> >       https://lists.01.org/mailman/listinfo/spdk
> >
>
> _______________________________________________
> SPDK mailing list
> SPDK(a)lists.01.org
> https://lists.01.org/mailman/listinfo/spdk
>

[-- Attachment #2: attachment.html --]
[-- Type: text/html, Size: 98516 bytes --]

             reply	other threads:[~2018-08-10  5:01 UTC|newest]

Thread overview: 9+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-08-10  5:01 Adam Chang [this message]
  -- strict thread matches above, loose matches on Subject: below --
2018-08-10  8:54 [SPDK] Error when issue IO in QEMU to vhost scsi NVMe Adam Chang
2018-08-10  5:14 Stojaczyk, DariuszX
2018-08-09 13:55 Stojaczyk, DariuszX
2018-08-09 12:42 Wodkowski, PawelX
2018-08-09 10:56 Adam Chang
2018-08-09  8:07 Stojaczyk, DariuszX
2018-08-09  6:20 Wodkowski, PawelX
2018-08-09  2:04 Adam Chang

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='CANvoUxh4t8=DPfdP2-3SZ4Lo8HMWqUnDjqfRTgZ2wrU9XS9s+A@mail.gmail.com' \
    --to=spdk@lists.01.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.