All of lore.kernel.org
 help / color / mirror / Atom feed
* Issue with Ceph File System and LIO
@ 2015-12-15  6:08 Eric Eastman
  2015-12-15  9:26 ` Mike Christie
                   ` (2 more replies)
  0 siblings, 3 replies; 20+ messages in thread
From: Eric Eastman @ 2015-12-15  6:08 UTC (permalink / raw)
  To: Ceph Development

I am testing Linux Target SCSI, LIO, with a Ceph File System backstore
and I am seeing this error on my LIO gateway.  I am using Ceph v9.2.0
on a 4.4rc4 Kernel, on Trusty, using a kernel mounted Ceph File
System.  A file on the Ceph File System is exported via iSCSI to a
VMware ESXi 5.0 server, and I am seeing this error when doing a lot of
I/O on the ESXi server.   Is this a LIO or a Ceph issue?

[Tue Dec 15 00:46:55 2015] ------------[ cut here ]------------
[Tue Dec 15 00:46:55 2015] WARNING: CPU: 0 PID: 1123421 at
/home/kernel/COD/linux/fs/ceph/addr.c:125
ceph_set_page_dirty+0x230/0x240 [ceph]()
[Tue Dec 15 00:46:55 2015] Modules linked in: iptable_filter ip_tables
x_tables xfs rbd iscsi_target_mod vhost_scsi tcm_qla2xxx ib_srpt
tcm_fc tcm_usb_gadget tcm_loop target_core_file target_core_iblock
target_core_pscsi target_core_user target_core_mod ipmi_devintf vhost
qla2xxx ib_cm ib_sa ib_mad ib_core ib_addr libfc scsi_transport_fc
libcomposite udc_core uio configfs ipmi_ssif ttm drm_kms_helper
gpio_ich drm i2c_algo_bit fb_sys_fops coretemp syscopyarea ipmi_si
sysfillrect ipmi_msghandler sysimgblt kvm acpi_power_meter 8250_fintek
irqbypass hpilo shpchp input_leds serio_raw lpc_ich i7core_edac
edac_core mac_hid ceph libceph libcrc32c fscache bonding lp parport
mlx4_en vxlan ip6_udp_tunnel udp_tunnel ptp pps_core hid_generic
usbhid hid hpsa mlx4_core psmouse bnx2 scsi_transport_sas fjes [last
unloaded: target_core_mod]
[Tue Dec 15 00:46:55 2015] CPU: 0 PID: 1123421 Comm: iscsi_trx
Tainted: G        W I     4.4.0-040400rc4-generic #201512061930
[Tue Dec 15 00:46:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
P64 01/22/2015
[Tue Dec 15 00:46:55 2015]  0000000000000000 00000000fdc0ce43
ffff880bf38c38c0 ffffffff813c8ab4
[Tue Dec 15 00:46:55 2015]  0000000000000000 ffff880bf38c38f8
ffffffff8107d772 ffffea00127a8680
[Tue Dec 15 00:46:55 2015]  ffff8804e52c1448 ffff8804e52c15b0
ffff8804e52c10f0 0000000000000200
[Tue Dec 15 00:46:55 2015] Call Trace:
[Tue Dec 15 00:46:55 2015]  [<ffffffff813c8ab4>] dump_stack+0x44/0x60
[Tue Dec 15 00:46:55 2015]  [<ffffffff8107d772>] warn_slowpath_common+0x82/0xc0
[Tue Dec 15 00:46:55 2015]  [<ffffffff8107d8ba>] warn_slowpath_null+0x1a/0x20
[Tue Dec 15 00:46:55 2015]  [<ffffffffc01fadb0>]
ceph_set_page_dirty+0x230/0x240 [ceph]
[Tue Dec 15 00:46:55 2015]  [<ffffffff81188770>] ?
pagecache_get_page+0x150/0x1c0
[Tue Dec 15 00:46:55 2015]  [<ffffffffc01fe338>] ?
ceph_pool_perm_check+0x48/0x700 [ceph]
[Tue Dec 15 00:46:55 2015]  [<ffffffff8119301d>] set_page_dirty+0x3d/0x70
[Tue Dec 15 00:46:55 2015]  [<ffffffffc01fcd7e>]
ceph_write_end+0x5e/0x180 [ceph]
[Tue Dec 15 00:46:55 2015]  [<ffffffff813dc006>] ?
iov_iter_copy_from_user_atomic+0x156/0x220
[Tue Dec 15 00:46:55 2015]  [<ffffffff81187bc4>]
generic_perform_write+0x114/0x1c0
[Tue Dec 15 00:46:55 2015]  [<ffffffffc01f818a>]
ceph_write_iter+0xf8a/0x1050 [ceph]
[Tue Dec 15 00:46:55 2015]  [<ffffffffc0205983>] ?
ceph_put_cap_refs+0x143/0x320 [ceph]
[Tue Dec 15 00:46:55 2015]  [<ffffffff810b10ba>] ?
check_preempt_wakeup+0xfa/0x220
[Tue Dec 15 00:46:55 2015]  [<ffffffff811a7eec>] ? zone_statistics+0x7c/0xa0
[Tue Dec 15 00:46:55 2015]  [<ffffffff813dd2ee>] ? copy_page_to_iter+0x5e/0xa0
[Tue Dec 15 00:46:55 2015]  [<ffffffff816e5d22>] ?
skb_copy_datagram_iter+0x122/0x250
[Tue Dec 15 00:46:55 2015]  [<ffffffff812053f6>] vfs_iter_write+0x76/0xc0
[Tue Dec 15 00:46:55 2015]  [<ffffffffc02cbf88>]
fd_do_rw.isra.5+0xd8/0x1e0 [target_core_file]
[Tue Dec 15 00:46:55 2015]  [<ffffffffc02cc155>]
fd_execute_rw+0xc5/0x2a0 [target_core_file]
[Tue Dec 15 00:46:55 2015]  [<ffffffffc04696f2>]
sbc_execute_rw+0x22/0x30 [target_core_mod]
[Tue Dec 15 00:46:55 2015]  [<ffffffffc04681ef>]
__target_execute_cmd+0x1f/0x70 [target_core_mod]
[Tue Dec 15 00:46:55 2015]  [<ffffffffc0468da5>]
target_execute_cmd+0x195/0x2a0 [target_core_mod]
[Tue Dec 15 00:46:55 2015]  [<ffffffffc05db89a>]
iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
[Tue Dec 15 00:46:55 2015]  [<ffffffffc05e4aea>]
iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
[Tue Dec 15 00:46:55 2015]  [<ffffffffc05eafbd>]
iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
[Tue Dec 15 00:46:55 2015]  [<ffffffff8101566c>] ? __switch_to+0x1dc/0x5a0
[Tue Dec 15 00:46:55 2015]  [<ffffffffc05eaaa0>] ?
iscsi_target_tx_thread+0x1e0/0x1e0 [iscsi_target_mod]
[Tue Dec 15 00:46:55 2015]  [<ffffffff8109c8b8>] kthread+0xd8/0xf0
[Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
kthread_create_on_node+0x1a0/0x1a0
[Tue Dec 15 00:46:55 2015]  [<ffffffff817fc58f>] ret_from_fork+0x3f/0x70
[Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
kthread_create_on_node+0x1a0/0x1a0
[Tue Dec 15 00:46:55 2015] ---[ end trace 4079437668c77cbb ]---
[Tue Dec 15 00:47:45 2015] ABORT_TASK: Found referenced iSCSI task_tag: 95784927
[Tue Dec 15 00:47:45 2015] ABORT_TASK: ref_tag: 95784927 already
complete, skipping

If it is a Ceph File System issue, let me know and I will open a bug.

Thanks

Eric

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-15  6:08 Issue with Ceph File System and LIO Eric Eastman
@ 2015-12-15  9:26 ` Mike Christie
  2015-12-15 10:12   ` John Spray
  2015-12-15 12:05 ` Yan, Zheng
  2015-12-18 21:54 ` Mike Christie
  2 siblings, 1 reply; 20+ messages in thread
From: Mike Christie @ 2015-12-15  9:26 UTC (permalink / raw)
  To: Eric Eastman, Ceph Development

On 12/15/2015 12:08 AM, Eric Eastman wrote:
> I am testing Linux Target SCSI, LIO, with a Ceph File System backstore
> and I am seeing this error on my LIO gateway.  I am using Ceph v9.2.0
> on a 4.4rc4 Kernel, on Trusty, using a kernel mounted Ceph File
> System.  A file on the Ceph File System is exported via iSCSI to a
> VMware ESXi 5.0 server, and I am seeing this error when doing a lot of
> I/O on the ESXi server.   Is this a LIO or a Ceph issue?
> 
> [Tue Dec 15 00:46:55 2015] ------------[ cut here ]------------
> [Tue Dec 15 00:46:55 2015] WARNING: CPU: 0 PID: 1123421 at
> /home/kernel/COD/linux/fs/ceph/addr.c:125
> ceph_set_page_dirty+0x230/0x240 [ceph]()
> [Tue Dec 15 00:46:55 2015] Modules linked in: iptable_filter ip_tables
> x_tables xfs rbd iscsi_target_mod vhost_scsi tcm_qla2xxx ib_srpt
> tcm_fc tcm_usb_gadget tcm_loop target_core_file target_core_iblock
> target_core_pscsi target_core_user target_core_mod ipmi_devintf vhost
> qla2xxx ib_cm ib_sa ib_mad ib_core ib_addr libfc scsi_transport_fc
> libcomposite udc_core uio configfs ipmi_ssif ttm drm_kms_helper
> gpio_ich drm i2c_algo_bit fb_sys_fops coretemp syscopyarea ipmi_si
> sysfillrect ipmi_msghandler sysimgblt kvm acpi_power_meter 8250_fintek
> irqbypass hpilo shpchp input_leds serio_raw lpc_ich i7core_edac
> edac_core mac_hid ceph libceph libcrc32c fscache bonding lp parport
> mlx4_en vxlan ip6_udp_tunnel udp_tunnel ptp pps_core hid_generic
> usbhid hid hpsa mlx4_core psmouse bnx2 scsi_transport_sas fjes [last
> unloaded: target_core_mod]
> [Tue Dec 15 00:46:55 2015] CPU: 0 PID: 1123421 Comm: iscsi_trx
> Tainted: G        W I     4.4.0-040400rc4-generic #201512061930
> [Tue Dec 15 00:46:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
> P64 01/22/2015
> [Tue Dec 15 00:46:55 2015]  0000000000000000 00000000fdc0ce43
> ffff880bf38c38c0 ffffffff813c8ab4
> [Tue Dec 15 00:46:55 2015]  0000000000000000 ffff880bf38c38f8
> ffffffff8107d772 ffffea00127a8680
> [Tue Dec 15 00:46:55 2015]  ffff8804e52c1448 ffff8804e52c15b0
> ffff8804e52c10f0 0000000000000200
> [Tue Dec 15 00:46:55 2015] Call Trace:
> [Tue Dec 15 00:46:55 2015]  [<ffffffff813c8ab4>] dump_stack+0x44/0x60
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d772>] warn_slowpath_common+0x82/0xc0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d8ba>] warn_slowpath_null+0x1a/0x20
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fadb0>]
> ceph_set_page_dirty+0x230/0x240 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff81188770>] ?
> pagecache_get_page+0x150/0x1c0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fe338>] ?
> ceph_pool_perm_check+0x48/0x700 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8119301d>] set_page_dirty+0x3d/0x70
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fcd7e>]
> ceph_write_end+0x5e/0x180 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dc006>] ?
> iov_iter_copy_from_user_atomic+0x156/0x220
> [Tue Dec 15 00:46:55 2015]  [<ffffffff81187bc4>]
> generic_perform_write+0x114/0x1c0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01f818a>]
> ceph_write_iter+0xf8a/0x1050 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0205983>] ?
> ceph_put_cap_refs+0x143/0x320 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff810b10ba>] ?
> check_preempt_wakeup+0xfa/0x220
> [Tue Dec 15 00:46:55 2015]  [<ffffffff811a7eec>] ? zone_statistics+0x7c/0xa0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dd2ee>] ? copy_page_to_iter+0x5e/0xa0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff816e5d22>] ?
> skb_copy_datagram_iter+0x122/0x250
> [Tue Dec 15 00:46:55 2015]  [<ffffffff812053f6>] vfs_iter_write+0x76/0xc0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cbf88>]
> fd_do_rw.isra.5+0xd8/0x1e0 [target_core_file]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cc155>]
> fd_execute_rw+0xc5/0x2a0 [target_core_file]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04696f2>]
> sbc_execute_rw+0x22/0x30 [target_core_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04681ef>]
> __target_execute_cmd+0x1f/0x70 [target_core_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0468da5>]
> target_execute_cmd+0x195/0x2a0 [target_core_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05db89a>]
> iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05e4aea>]
> iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eafbd>]
> iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8101566c>] ? __switch_to+0x1dc/0x5a0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eaaa0>] ?
> iscsi_target_tx_thread+0x1e0/0x1e0 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c8b8>] kthread+0xd8/0xf0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
> kthread_create_on_node+0x1a0/0x1a0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff817fc58f>] ret_from_fork+0x3f/0x70
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
> kthread_create_on_node+0x1a0/0x1a0
> [Tue Dec 15 00:46:55 2015] ---[ end trace 4079437668c77cbb ]---
> [Tue Dec 15 00:47:45 2015] ABORT_TASK: Found referenced iSCSI task_tag: 95784927
> [Tue Dec 15 00:47:45 2015] ABORT_TASK: ref_tag: 95784927 already
> complete, skipping
> 


For writes, LIO just allocates pages using GFP_KERNEL, passes them to
sock_recvmsg to read the data into them, then passes them to the fs
using the function you see above, vfs_iter_write. So it does not do
anything fancy.

Do we need to send specific types of pages to ceph?


^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-15  9:26 ` Mike Christie
@ 2015-12-15 10:12   ` John Spray
  0 siblings, 0 replies; 20+ messages in thread
From: John Spray @ 2015-12-15 10:12 UTC (permalink / raw)
  To: Mike Christie; +Cc: Eric Eastman, Ceph Development, Zheng Yan

On Tue, Dec 15, 2015 at 9:26 AM, Mike Christie <mchristi@redhat.com> wrote:
> On 12/15/2015 12:08 AM, Eric Eastman wrote:
>> I am testing Linux Target SCSI, LIO, with a Ceph File System backstore
>> and I am seeing this error on my LIO gateway.  I am using Ceph v9.2.0
>> on a 4.4rc4 Kernel, on Trusty, using a kernel mounted Ceph File
>> System.  A file on the Ceph File System is exported via iSCSI to a
>> VMware ESXi 5.0 server, and I am seeing this error when doing a lot of
>> I/O on the ESXi server.   Is this a LIO or a Ceph issue?
>>
>> [Tue Dec 15 00:46:55 2015] ------------[ cut here ]------------
>> [Tue Dec 15 00:46:55 2015] WARNING: CPU: 0 PID: 1123421 at
>> /home/kernel/COD/linux/fs/ceph/addr.c:125
>> ceph_set_page_dirty+0x230/0x240 [ceph]()
>> [Tue Dec 15 00:46:55 2015] Modules linked in: iptable_filter ip_tables
>> x_tables xfs rbd iscsi_target_mod vhost_scsi tcm_qla2xxx ib_srpt
>> tcm_fc tcm_usb_gadget tcm_loop target_core_file target_core_iblock
>> target_core_pscsi target_core_user target_core_mod ipmi_devintf vhost
>> qla2xxx ib_cm ib_sa ib_mad ib_core ib_addr libfc scsi_transport_fc
>> libcomposite udc_core uio configfs ipmi_ssif ttm drm_kms_helper
>> gpio_ich drm i2c_algo_bit fb_sys_fops coretemp syscopyarea ipmi_si
>> sysfillrect ipmi_msghandler sysimgblt kvm acpi_power_meter 8250_fintek
>> irqbypass hpilo shpchp input_leds serio_raw lpc_ich i7core_edac
>> edac_core mac_hid ceph libceph libcrc32c fscache bonding lp parport
>> mlx4_en vxlan ip6_udp_tunnel udp_tunnel ptp pps_core hid_generic
>> usbhid hid hpsa mlx4_core psmouse bnx2 scsi_transport_sas fjes [last
>> unloaded: target_core_mod]
>> [Tue Dec 15 00:46:55 2015] CPU: 0 PID: 1123421 Comm: iscsi_trx
>> Tainted: G        W I     4.4.0-040400rc4-generic #201512061930
>> [Tue Dec 15 00:46:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
>> P64 01/22/2015
>> [Tue Dec 15 00:46:55 2015]  0000000000000000 00000000fdc0ce43
>> ffff880bf38c38c0 ffffffff813c8ab4
>> [Tue Dec 15 00:46:55 2015]  0000000000000000 ffff880bf38c38f8
>> ffffffff8107d772 ffffea00127a8680
>> [Tue Dec 15 00:46:55 2015]  ffff8804e52c1448 ffff8804e52c15b0
>> ffff8804e52c10f0 0000000000000200
>> [Tue Dec 15 00:46:55 2015] Call Trace:
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff813c8ab4>] dump_stack+0x44/0x60
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d772>] warn_slowpath_common+0x82/0xc0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d8ba>] warn_slowpath_null+0x1a/0x20
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fadb0>]
>> ceph_set_page_dirty+0x230/0x240 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff81188770>] ?
>> pagecache_get_page+0x150/0x1c0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fe338>] ?
>> ceph_pool_perm_check+0x48/0x700 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8119301d>] set_page_dirty+0x3d/0x70
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fcd7e>]
>> ceph_write_end+0x5e/0x180 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dc006>] ?
>> iov_iter_copy_from_user_atomic+0x156/0x220
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff81187bc4>]
>> generic_perform_write+0x114/0x1c0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01f818a>]
>> ceph_write_iter+0xf8a/0x1050 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0205983>] ?
>> ceph_put_cap_refs+0x143/0x320 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff810b10ba>] ?
>> check_preempt_wakeup+0xfa/0x220
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff811a7eec>] ? zone_statistics+0x7c/0xa0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dd2ee>] ? copy_page_to_iter+0x5e/0xa0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff816e5d22>] ?
>> skb_copy_datagram_iter+0x122/0x250
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff812053f6>] vfs_iter_write+0x76/0xc0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cbf88>]
>> fd_do_rw.isra.5+0xd8/0x1e0 [target_core_file]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cc155>]
>> fd_execute_rw+0xc5/0x2a0 [target_core_file]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04696f2>]
>> sbc_execute_rw+0x22/0x30 [target_core_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04681ef>]
>> __target_execute_cmd+0x1f/0x70 [target_core_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0468da5>]
>> target_execute_cmd+0x195/0x2a0 [target_core_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05db89a>]
>> iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05e4aea>]
>> iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eafbd>]
>> iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8101566c>] ? __switch_to+0x1dc/0x5a0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eaaa0>] ?
>> iscsi_target_tx_thread+0x1e0/0x1e0 [iscsi_target_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c8b8>] kthread+0xd8/0xf0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
>> kthread_create_on_node+0x1a0/0x1a0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff817fc58f>] ret_from_fork+0x3f/0x70
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
>> kthread_create_on_node+0x1a0/0x1a0
>> [Tue Dec 15 00:46:55 2015] ---[ end trace 4079437668c77cbb ]---
>> [Tue Dec 15 00:47:45 2015] ABORT_TASK: Found referenced iSCSI task_tag: 95784927
>> [Tue Dec 15 00:47:45 2015] ABORT_TASK: ref_tag: 95784927 already
>> complete, skipping

Looks likely to be a kclient bug, as it's in the newish
pool_perm_check path.  Perhaps we don't usually see this because we'd
usually hit the permissions checks earlier (or during a read).

CCing zyan, who will have a better idea than me.

Eric: you should probably go ahead and open a ticket for this.

John

>>
>
>
> For writes, LIO just allocates pages using GFP_KERNEL, passes them to
> sock_recvmsg to read the data into them, then passes them to the fs
> using the function you see above, vfs_iter_write. So it does not do
> anything fancy.
>
> Do we need to send specific types of pages to ceph?
>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-15  6:08 Issue with Ceph File System and LIO Eric Eastman
  2015-12-15  9:26 ` Mike Christie
@ 2015-12-15 12:05 ` Yan, Zheng
  2015-12-15 16:51   ` Eric Eastman
  2015-12-18 21:54 ` Mike Christie
  2 siblings, 1 reply; 20+ messages in thread
From: Yan, Zheng @ 2015-12-15 12:05 UTC (permalink / raw)
  To: Eric Eastman; +Cc: Ceph Development

On Tue, Dec 15, 2015 at 2:08 PM, Eric Eastman
<eric.eastman@keepertech.com> wrote:
> I am testing Linux Target SCSI, LIO, with a Ceph File System backstore
> and I am seeing this error on my LIO gateway.  I am using Ceph v9.2.0
> on a 4.4rc4 Kernel, on Trusty, using a kernel mounted Ceph File
> System.  A file on the Ceph File System is exported via iSCSI to a
> VMware ESXi 5.0 server, and I am seeing this error when doing a lot of
> I/O on the ESXi server.   Is this a LIO or a Ceph issue?
>
> [Tue Dec 15 00:46:55 2015] ------------[ cut here ]------------
> [Tue Dec 15 00:46:55 2015] WARNING: CPU: 0 PID: 1123421 at
> /home/kernel/COD/linux/fs/ceph/addr.c:125

could you confirm that addr.c:125 is WARN_ON(!PageLocked(page));

Regards
Yan, Zheng

> ceph_set_page_dirty+0x230/0x240 [ceph]()
> [Tue Dec 15 00:46:55 2015] Modules linked in: iptable_filter ip_tables
> x_tables xfs rbd iscsi_target_mod vhost_scsi tcm_qla2xxx ib_srpt
> tcm_fc tcm_usb_gadget tcm_loop target_core_file target_core_iblock
> target_core_pscsi target_core_user target_core_mod ipmi_devintf vhost
> qla2xxx ib_cm ib_sa ib_mad ib_core ib_addr libfc scsi_transport_fc
> libcomposite udc_core uio configfs ipmi_ssif ttm drm_kms_helper
> gpio_ich drm i2c_algo_bit fb_sys_fops coretemp syscopyarea ipmi_si
> sysfillrect ipmi_msghandler sysimgblt kvm acpi_power_meter 8250_fintek
> irqbypass hpilo shpchp input_leds serio_raw lpc_ich i7core_edac
> edac_core mac_hid ceph libceph libcrc32c fscache bonding lp parport
> mlx4_en vxlan ip6_udp_tunnel udp_tunnel ptp pps_core hid_generic
> usbhid hid hpsa mlx4_core psmouse bnx2 scsi_transport_sas fjes [last
> unloaded: target_core_mod]
> [Tue Dec 15 00:46:55 2015] CPU: 0 PID: 1123421 Comm: iscsi_trx
> Tainted: G        W I     4.4.0-040400rc4-generic #201512061930
> [Tue Dec 15 00:46:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
> P64 01/22/2015
> [Tue Dec 15 00:46:55 2015]  0000000000000000 00000000fdc0ce43
> ffff880bf38c38c0 ffffffff813c8ab4
> [Tue Dec 15 00:46:55 2015]  0000000000000000 ffff880bf38c38f8
> ffffffff8107d772 ffffea00127a8680
> [Tue Dec 15 00:46:55 2015]  ffff8804e52c1448 ffff8804e52c15b0
> ffff8804e52c10f0 0000000000000200
> [Tue Dec 15 00:46:55 2015] Call Trace:
> [Tue Dec 15 00:46:55 2015]  [<ffffffff813c8ab4>] dump_stack+0x44/0x60
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d772>] warn_slowpath_common+0x82/0xc0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d8ba>] warn_slowpath_null+0x1a/0x20
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fadb0>]
> ceph_set_page_dirty+0x230/0x240 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff81188770>] ?
> pagecache_get_page+0x150/0x1c0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fe338>] ?
> ceph_pool_perm_check+0x48/0x700 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8119301d>] set_page_dirty+0x3d/0x70
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fcd7e>]
> ceph_write_end+0x5e/0x180 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dc006>] ?
> iov_iter_copy_from_user_atomic+0x156/0x220
> [Tue Dec 15 00:46:55 2015]  [<ffffffff81187bc4>]
> generic_perform_write+0x114/0x1c0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01f818a>]
> ceph_write_iter+0xf8a/0x1050 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0205983>] ?
> ceph_put_cap_refs+0x143/0x320 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff810b10ba>] ?
> check_preempt_wakeup+0xfa/0x220
> [Tue Dec 15 00:46:55 2015]  [<ffffffff811a7eec>] ? zone_statistics+0x7c/0xa0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dd2ee>] ? copy_page_to_iter+0x5e/0xa0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff816e5d22>] ?
> skb_copy_datagram_iter+0x122/0x250
> [Tue Dec 15 00:46:55 2015]  [<ffffffff812053f6>] vfs_iter_write+0x76/0xc0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cbf88>]
> fd_do_rw.isra.5+0xd8/0x1e0 [target_core_file]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cc155>]
> fd_execute_rw+0xc5/0x2a0 [target_core_file]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04696f2>]
> sbc_execute_rw+0x22/0x30 [target_core_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04681ef>]
> __target_execute_cmd+0x1f/0x70 [target_core_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0468da5>]
> target_execute_cmd+0x195/0x2a0 [target_core_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05db89a>]
> iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05e4aea>]
> iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eafbd>]
> iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8101566c>] ? __switch_to+0x1dc/0x5a0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eaaa0>] ?
> iscsi_target_tx_thread+0x1e0/0x1e0 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c8b8>] kthread+0xd8/0xf0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
> kthread_create_on_node+0x1a0/0x1a0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff817fc58f>] ret_from_fork+0x3f/0x70
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
> kthread_create_on_node+0x1a0/0x1a0
> [Tue Dec 15 00:46:55 2015] ---[ end trace 4079437668c77cbb ]---
> [Tue Dec 15 00:47:45 2015] ABORT_TASK: Found referenced iSCSI task_tag: 95784927
> [Tue Dec 15 00:47:45 2015] ABORT_TASK: ref_tag: 95784927 already
> complete, skipping
>
> If it is a Ceph File System issue, let me know and I will open a bug.
>
> Thanks
>
> Eric
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-15 12:05 ` Yan, Zheng
@ 2015-12-15 16:51   ` Eric Eastman
  2015-12-16 14:29     ` Yan, Zheng
  0 siblings, 1 reply; 20+ messages in thread
From: Eric Eastman @ 2015-12-15 16:51 UTC (permalink / raw)
  To: Yan, Zheng, spray, mchristi; +Cc: Ceph Development

I have opened ticket: 14086

On Tue, Dec 15, 2015 at 5:05 AM, Yan, Zheng <ukernel@gmail.com> wrote:
> On Tue, Dec 15, 2015 at 2:08 PM, Eric Eastman
>> [Tue Dec 15 00:46:55 2015] ------------[ cut here ]------------
>> [Tue Dec 15 00:46:55 2015] WARNING: CPU: 0 PID: 1123421 at
>> /home/kernel/COD/linux/fs/ceph/addr.c:125
>
> could you confirm that addr.c:125 is WARN_ON(!PageLocked(page));

I am using the generic kernel from:
http://kernel.ubuntu.com/~kernel-ppa/mainline/v4.4-rc4-wily
and assuming they did not change anything, from the 4.4rc4 source tree
I pulled down shows:

124         ret = __set_page_dirty_nobuffers(page);
125         WARN_ON(!PageLocked(page));
126         WARN_ON(!page->mapping);


modinfo ceph
filename:       /lib/modules/4.4.0-040400rc4-generic/kernel/fs/ceph/ceph.ko
license:        GPL
description:    Ceph filesystem for Linux
author:         Patience Warnick <patience@newdream.net>
author:         Yehuda Sadeh <yehuda@hq.newdream.net>
author:         Sage Weil <sage@newdream.net>
alias:          fs-ceph
srcversion:     E94BA78C2D998705FE2C600
depends:        libceph,fscache
intree:         Y
vermagic:       4.4.0-040400rc4-generic SMP mod_unload modversions

This error has shown up about 20 times in 12 hours, since I started
the ESXi test.

Thanks,
Eric

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-15 16:51   ` Eric Eastman
@ 2015-12-16 14:29     ` Yan, Zheng
  2015-12-16 15:15       ` Eric Eastman
  0 siblings, 1 reply; 20+ messages in thread
From: Yan, Zheng @ 2015-12-16 14:29 UTC (permalink / raw)
  To: Eric Eastman; +Cc: spray, mchristi, Ceph Development

[-- Attachment #1: Type: text/plain, Size: 1522 bytes --]

On Wed, Dec 16, 2015 at 12:51 AM, Eric Eastman
<eric.eastman@keepertech.com> wrote:
> I have opened ticket: 14086
>
> On Tue, Dec 15, 2015 at 5:05 AM, Yan, Zheng <ukernel@gmail.com> wrote:
>> On Tue, Dec 15, 2015 at 2:08 PM, Eric Eastman
>>> [Tue Dec 15 00:46:55 2015] ------------[ cut here ]------------
>>> [Tue Dec 15 00:46:55 2015] WARNING: CPU: 0 PID: 1123421 at
>>> /home/kernel/COD/linux/fs/ceph/addr.c:125
>>
>> could you confirm that addr.c:125 is WARN_ON(!PageLocked(page));
>
> I am using the generic kernel from:
> http://kernel.ubuntu.com/~kernel-ppa/mainline/v4.4-rc4-wily
> and assuming they did not change anything, from the 4.4rc4 source tree
> I pulled down shows:
>
> 124         ret = __set_page_dirty_nobuffers(page);
> 125         WARN_ON(!PageLocked(page));
> 126         WARN_ON(!page->mapping);
>
>
> modinfo ceph
> filename:       /lib/modules/4.4.0-040400rc4-generic/kernel/fs/ceph/ceph.ko
> license:        GPL
> description:    Ceph filesystem for Linux
> author:         Patience Warnick <patience@newdream.net>
> author:         Yehuda Sadeh <yehuda@hq.newdream.net>
> author:         Sage Weil <sage@newdream.net>
> alias:          fs-ceph
> srcversion:     E94BA78C2D998705FE2C600
> depends:        libceph,fscache
> intree:         Y
> vermagic:       4.4.0-040400rc4-generic SMP mod_unload modversions
>
> This error has shown up about 20 times in 12 hours, since I started
> the ESXi test.
>

This warning is really strange. Could you try the attached debug patch.

Regards
Yan, Zheng

[-- Attachment #2: cephfs.patch --]
[-- Type: application/octet-stream, Size: 1468 bytes --]

diff --git a/fs/ceph/addr.c b/fs/ceph/addr.c
index b7d218a..988bda4 100644
--- a/fs/ceph/addr.c
+++ b/fs/ceph/addr.c
@@ -1149,16 +1149,19 @@ static int ceph_write_begin(struct file *file, struct address_space *mapping,
 		page = grab_cache_page_write_begin(mapping, index, 0);
 		if (!page)
 			return -ENOMEM;
+		WARN_ON(!PageLocked(page));
 		*pagep = page;
 
 		dout("write_begin file %p inode %p page %p %d~%d\n", file,
 		     inode, page, (int)pos, (int)len);
 
 		r = ceph_update_writeable_page(file, pos, len, page);
-		if (r < 0)
+		if (r < 0) {
 			page_cache_release(page);
-		else
+		} else {
+			WARN_ON(!PageLocked(page));
 			*pagep = page;
+		}
 	} while (r == -EAGAIN);
 
 	return r;
@@ -1378,11 +1381,13 @@ static int ceph_page_mkwrite(struct vm_area_struct *vma, struct vm_fault *vmf)
 
 	ret = VM_FAULT_NOPAGE;
 	if ((off > size) ||
-	    (page->mapping != inode->i_mapping))
+	    (page->mapping != inode->i_mapping)) {
+		unlock_page(page);
 		goto out;
+	}
 
 	ret = ceph_update_writeable_page(vma->vm_file, off, len, page);
-	if (ret == 0) {
+	if (ret >= 0) {
 		/* success.  we'll keep the page locked. */
 		set_page_dirty(page);
 		ret = VM_FAULT_LOCKED;
@@ -1393,8 +1398,6 @@ static int ceph_page_mkwrite(struct vm_area_struct *vma, struct vm_fault *vmf)
 			ret = VM_FAULT_SIGBUS;
 	}
 out:
-	if (ret != VM_FAULT_LOCKED)
-		unlock_page(page);
 	if (ret == VM_FAULT_LOCKED ||
 	    ci->i_inline_version != CEPH_INLINE_NONE) {
 		int dirty;

^ permalink raw reply related	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-16 14:29     ` Yan, Zheng
@ 2015-12-16 15:15       ` Eric Eastman
  2015-12-17  8:56         ` Eric Eastman
  0 siblings, 1 reply; 20+ messages in thread
From: Eric Eastman @ 2015-12-16 15:15 UTC (permalink / raw)
  To: Yan, Zheng; +Cc: Ceph Development

>>
> This warning is really strange. Could you try the attached debug patch.
>
> Regards
> Yan, Zheng

I will try the patch and get back to the list.

Eric

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-16 15:15       ` Eric Eastman
@ 2015-12-17  8:56         ` Eric Eastman
  2015-12-17  9:38           ` Yan, Zheng
  2015-12-17 16:14           ` Minfei Huang
  0 siblings, 2 replies; 20+ messages in thread
From: Eric Eastman @ 2015-12-17  8:56 UTC (permalink / raw)
  To: Yan, Zheng; +Cc: Ceph Development

I patched the 4.4rc4 kernel source and restarted the test.  Shortly
after starting it, this showed up in dmesg:

[Thu Dec 17 03:29:55 2015] WARNING: CPU: 0 PID: 2547 at
fs/ceph/addr.c:1162 ceph_write_begin+0xfb/0x120 [ceph]()
[Thu Dec 17 03:29:55 2015] Modules linked in: iscsi_target_mod
vhost_scsi tcm_qla2xxx ib_srpt tcm_fc tcm_usb_gadget tcm_loop
target_core_file target_core_iblock target_core_pscsi target_core_user
target_core_mod ipmi_devintf vhost qla2xxx ib_cm ib_sa ib_mad ib_core
ib_addr libfc scsi_transport_fc libcomposite udc_core uio configfs ttm
ipmi_ssif drm_kms_helper drm coretemp kvm gpio_ich i2c_algo_bit
i7core_edac fb_sys_fops syscopyarea edac_core sysfillrect sysimgblt
ipmi_si input_leds hpilo ipmi_msghandler shpchp acpi_power_meter
irqbypass serio_raw 8250_fintek lpc_ich mac_hid ceph bonding libceph
lp parport libcrc32c fscache mlx4_en vxlan ip6_udp_tunnel udp_tunnel
ptp pps_core hid_generic usbhid hid mlx4_core hpsa psmouse bnx2 fjes
scsi_transport_sas [last unloaded: target_core_mod]
[Thu Dec 17 03:29:55 2015] CPU: 0 PID: 2547 Comm: iscsi_trx Tainted: G
       W I     4.4.0-rc4-ede1 #1
[Thu Dec 17 03:29:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
P64 01/22/2015
[Thu Dec 17 03:29:55 2015]  ffffffffc020cd47 ffff8805f1e97958
ffffffff813ad644 0000000000000000
[Thu Dec 17 03:29:55 2015]  ffff8805f1e97990 ffffffff81079702
ffff8805f1e97a50 00000000015dd000
[Thu Dec 17 03:29:55 2015]  ffff880c034df800 0000000000000200
ffffea0000b26a80 ffff8805f1e979a0
[Thu Dec 17 03:29:55 2015] Call Trace:
[Thu Dec 17 03:29:55 2015]  [<ffffffff813ad644>] dump_stack+0x44/0x60
[Thu Dec 17 03:29:55 2015]  [<ffffffff81079702>] warn_slowpath_common+0x82/0xc0
[Thu Dec 17 03:29:55 2015]  [<ffffffff810797fa>] warn_slowpath_null+0x1a/0x20
[Thu Dec 17 03:29:55 2015]  [<ffffffffc01e53bb>]
ceph_write_begin+0xfb/0x120 [ceph]
[Thu Dec 17 03:29:55 2015]  [<ffffffff8117c8df>]
generic_perform_write+0xbf/0x1a0
[Thu Dec 17 03:29:55 2015]  [<ffffffffc01dff9c>]
ceph_write_iter+0xf5c/0x1010 [ceph]
[Thu Dec 17 03:29:55 2015]  [<ffffffff810a888c>] ? __enqueue_entity+0x6c/0x70
[Thu Dec 17 03:29:55 2015]  [<ffffffff813c0003>] ?
iov_iter_get_pages+0x113/0x210
[Thu Dec 17 03:29:55 2015]  [<ffffffff816b6802>] ?
skb_copy_datagram_iter+0x122/0x250
[Thu Dec 17 03:29:55 2015]  [<ffffffff811f6c93>] vfs_iter_write+0x63/0xa0
[Thu Dec 17 03:29:55 2015]  [<ffffffffc03c3f29>]
fd_do_rw.isra.5+0xc9/0x1b0 [target_core_file]
[Thu Dec 17 03:29:55 2015]  [<ffffffffc03c40d5>]
fd_execute_rw+0xc5/0x2a0 [target_core_file]
[Thu Dec 17 03:29:55 2015]  [<ffffffffc0445e72>]
sbc_execute_rw+0x22/0x30 [target_core_mod]
[Thu Dec 17 03:29:55 2015]  [<ffffffffc04449cf>]
__target_execute_cmd+0x1f/0x70 [target_core_mod]
[Thu Dec 17 03:29:55 2015]  [<ffffffffc0445525>]
target_execute_cmd+0x195/0x2a0 [target_core_mod]
[Thu Dec 17 03:29:55 2015]  [<ffffffffc05c978a>]
iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
[Thu Dec 17 03:29:55 2015]  [<ffffffffc05d28da>]
iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
[Thu Dec 17 03:29:55 2015]  [<ffffffffc05d8c4d>]
iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
[Thu Dec 17 03:29:55 2015]  [<ffffffff8101463d>] ? __switch_to+0x1cd/0x570
[Thu Dec 17 03:29:55 2015]  [<ffffffffc05d8730>] ?
iscsi_target_tx_thread+0x1c0/0x1c0 [iscsi_target_mod]
[Thu Dec 17 03:29:55 2015]  [<ffffffff810974c9>] kthread+0xc9/0xe0
[Thu Dec 17 03:29:55 2015]  [<ffffffff81097400>] ?
kthread_create_on_node+0x180/0x180
[Thu Dec 17 03:29:55 2015]  [<ffffffff817c794f>] ret_from_fork+0x3f/0x70
[Thu Dec 17 03:29:55 2015]  [<ffffffff81097400>] ?
kthread_create_on_node+0x180/0x180
[Thu Dec 17 03:29:55 2015] ---[ end trace 382a45986961da4e ]---

There are WARNINGs on both line 125 and 1162. I will attached the
whole set of dmesg output to the tracker ticket 14086

I wanted to note that file system snapshots are enabled and being used
on this file system.

Thanks
Eric

On Wed, Dec 16, 2015 at 8:15 AM, Eric Eastman
<eric.eastman@keepertech.com> wrote:
>>>
>> This warning is really strange. Could you try the attached debug patch.
>>
>> Regards
>> Yan, Zheng
>
> I will try the patch and get back to the list.
>
> Eric

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-17  8:56         ` Eric Eastman
@ 2015-12-17  9:38           ` Yan, Zheng
  2015-12-17 19:49             ` Eric Eastman
  2015-12-17 16:14           ` Minfei Huang
  1 sibling, 1 reply; 20+ messages in thread
From: Yan, Zheng @ 2015-12-17  9:38 UTC (permalink / raw)
  To: Eric Eastman; +Cc: Ceph Development

[-- Attachment #1: Type: text/plain, Size: 4490 bytes --]

On Thu, Dec 17, 2015 at 4:56 PM, Eric Eastman
<eric.eastman@keepertech.com> wrote:
> I patched the 4.4rc4 kernel source and restarted the test.  Shortly
> after starting it, this showed up in dmesg:
>
> [Thu Dec 17 03:29:55 2015] WARNING: CPU: 0 PID: 2547 at
> fs/ceph/addr.c:1162 ceph_write_begin+0xfb/0x120 [ceph]()
> [Thu Dec 17 03:29:55 2015] Modules linked in: iscsi_target_mod
> vhost_scsi tcm_qla2xxx ib_srpt tcm_fc tcm_usb_gadget tcm_loop
> target_core_file target_core_iblock target_core_pscsi target_core_user
> target_core_mod ipmi_devintf vhost qla2xxx ib_cm ib_sa ib_mad ib_core
> ib_addr libfc scsi_transport_fc libcomposite udc_core uio configfs ttm
> ipmi_ssif drm_kms_helper drm coretemp kvm gpio_ich i2c_algo_bit
> i7core_edac fb_sys_fops syscopyarea edac_core sysfillrect sysimgblt
> ipmi_si input_leds hpilo ipmi_msghandler shpchp acpi_power_meter
> irqbypass serio_raw 8250_fintek lpc_ich mac_hid ceph bonding libceph
> lp parport libcrc32c fscache mlx4_en vxlan ip6_udp_tunnel udp_tunnel
> ptp pps_core hid_generic usbhid hid mlx4_core hpsa psmouse bnx2 fjes
> scsi_transport_sas [last unloaded: target_core_mod]
> [Thu Dec 17 03:29:55 2015] CPU: 0 PID: 2547 Comm: iscsi_trx Tainted: G
>        W I     4.4.0-rc4-ede1 #1
> [Thu Dec 17 03:29:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
> P64 01/22/2015
> [Thu Dec 17 03:29:55 2015]  ffffffffc020cd47 ffff8805f1e97958
> ffffffff813ad644 0000000000000000
> [Thu Dec 17 03:29:55 2015]  ffff8805f1e97990 ffffffff81079702
> ffff8805f1e97a50 00000000015dd000
> [Thu Dec 17 03:29:55 2015]  ffff880c034df800 0000000000000200
> ffffea0000b26a80 ffff8805f1e979a0
> [Thu Dec 17 03:29:55 2015] Call Trace:
> [Thu Dec 17 03:29:55 2015]  [<ffffffff813ad644>] dump_stack+0x44/0x60
> [Thu Dec 17 03:29:55 2015]  [<ffffffff81079702>] warn_slowpath_common+0x82/0xc0
> [Thu Dec 17 03:29:55 2015]  [<ffffffff810797fa>] warn_slowpath_null+0x1a/0x20
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc01e53bb>]
> ceph_write_begin+0xfb/0x120 [ceph]
> [Thu Dec 17 03:29:55 2015]  [<ffffffff8117c8df>]
> generic_perform_write+0xbf/0x1a0
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc01dff9c>]
> ceph_write_iter+0xf5c/0x1010 [ceph]
> [Thu Dec 17 03:29:55 2015]  [<ffffffff810a888c>] ? __enqueue_entity+0x6c/0x70
> [Thu Dec 17 03:29:55 2015]  [<ffffffff813c0003>] ?
> iov_iter_get_pages+0x113/0x210
> [Thu Dec 17 03:29:55 2015]  [<ffffffff816b6802>] ?
> skb_copy_datagram_iter+0x122/0x250
> [Thu Dec 17 03:29:55 2015]  [<ffffffff811f6c93>] vfs_iter_write+0x63/0xa0
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc03c3f29>]
> fd_do_rw.isra.5+0xc9/0x1b0 [target_core_file]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc03c40d5>]
> fd_execute_rw+0xc5/0x2a0 [target_core_file]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc0445e72>]
> sbc_execute_rw+0x22/0x30 [target_core_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc04449cf>]
> __target_execute_cmd+0x1f/0x70 [target_core_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc0445525>]
> target_execute_cmd+0x195/0x2a0 [target_core_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05c978a>]
> iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05d28da>]
> iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05d8c4d>]
> iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffff8101463d>] ? __switch_to+0x1cd/0x570
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05d8730>] ?
> iscsi_target_tx_thread+0x1c0/0x1c0 [iscsi_target_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffff810974c9>] kthread+0xc9/0xe0
> [Thu Dec 17 03:29:55 2015]  [<ffffffff81097400>] ?
> kthread_create_on_node+0x180/0x180
> [Thu Dec 17 03:29:55 2015]  [<ffffffff817c794f>] ret_from_fork+0x3f/0x70
> [Thu Dec 17 03:29:55 2015]  [<ffffffff81097400>] ?
> kthread_create_on_node+0x180/0x180
> [Thu Dec 17 03:29:55 2015] ---[ end trace 382a45986961da4e ]---


Could you please try the apply the new incremental patch and try again.


Regards
Yan, Zheng


>
> There are WARNINGs on both line 125 and 1162. I will attached the
> whole set of dmesg output to the tracker ticket 14086
>
> I wanted to note that file system snapshots are enabled and being used
> on this file system.
>
> Thanks
> Eric
>
> On Wed, Dec 16, 2015 at 8:15 AM, Eric Eastman
> <eric.eastman@keepertech.com> wrote:
>>>>
>>> This warning is really strange. Could you try the attached debug patch.
>>>
>>> Regards
>>> Yan, Zheng
>>
>> I will try the patch and get back to the list.
>>
>> Eric

[-- Attachment #2: cephfs1.patch --]
[-- Type: application/octet-stream, Size: 1390 bytes --]

diff --git a/fs/ceph/addr.c b/fs/ceph/addr.c
index b7d218a..d5ea0f3 100644
--- a/fs/ceph/addr.c
+++ b/fs/ceph/addr.c
@@ -1081,18 +1081,24 @@ retry_locked:
 			r = wait_event_interruptible(ci->i_cap_wq,
 			       context_is_writeable_or_written(inode, snapc));
 			ceph_put_snap_context(snapc);
-			if (r == -ERESTARTSYS)
-				return r;
-			return -EAGAIN;
+			if (r != -ERESTARTSYS)
+				r = -EAGAIN;
+			pr_err("update_writeable_page no snapc %p %d\n", page, r);
+			return r;
 		}
 		ceph_put_snap_context(oldest);
 
 		/* yay, writeable, do it now (without dropping page lock) */
 		dout(" page %p snapc %p not current, but oldest\n",
 		     page, snapc);
-		if (!clear_page_dirty_for_io(page))
+		if (!clear_page_dirty_for_io(page)) {
+			if (!PageLocked(page))
+				pr_err("clear_page_dirty_for_io unlocks %p\n", page);
 			goto retry_locked;
+		}
 		r = writepage_nounlock(page, NULL);
+		if (!PageLocked(page))
+			pr_err("writepage_nounlock unlocks %p %d\n", page, r);
 		if (r < 0)
 			goto fail_nosnap;
 		goto retry_locked;
@@ -1123,11 +1129,14 @@ retry_locked:
 
 	/* we need to read it. */
 	r = readpage_nounlock(file, page);
+	if (!PageLocked(page))
+		pr_err("readpage_nounlock unlocks %p %d\n", page, r);
 	if (r < 0)
 		goto fail_nosnap;
 	goto retry_locked;
 fail_nosnap:
 	unlock_page(page);
+	pr_err("update_writeable_page failed %p %d\n", page, r);
 	return r;
 }
 

^ permalink raw reply related	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-17  8:56         ` Eric Eastman
  2015-12-17  9:38           ` Yan, Zheng
@ 2015-12-17 16:14           ` Minfei Huang
  1 sibling, 0 replies; 20+ messages in thread
From: Minfei Huang @ 2015-12-17 16:14 UTC (permalink / raw)
  To: Eric Eastman; +Cc: Yan, Zheng, Ceph Development

Hi.

It may be helpful to address this issue, if we flip the debug.

Thanks
Minfei

On 12/17/15 at 01:56P, Eric Eastman wrote:
> I patched the 4.4rc4 kernel source and restarted the test.  Shortly
> after starting it, this showed up in dmesg:
> 
> [Thu Dec 17 03:29:55 2015] WARNING: CPU: 0 PID: 2547 at
> fs/ceph/addr.c:1162 ceph_write_begin+0xfb/0x120 [ceph]()
> [Thu Dec 17 03:29:55 2015] Modules linked in: iscsi_target_mod
> vhost_scsi tcm_qla2xxx ib_srpt tcm_fc tcm_usb_gadget tcm_loop
> target_core_file target_core_iblock target_core_pscsi target_core_user
> target_core_mod ipmi_devintf vhost qla2xxx ib_cm ib_sa ib_mad ib_core
> ib_addr libfc scsi_transport_fc libcomposite udc_core uio configfs ttm
> ipmi_ssif drm_kms_helper drm coretemp kvm gpio_ich i2c_algo_bit
> i7core_edac fb_sys_fops syscopyarea edac_core sysfillrect sysimgblt
> ipmi_si input_leds hpilo ipmi_msghandler shpchp acpi_power_meter
> irqbypass serio_raw 8250_fintek lpc_ich mac_hid ceph bonding libceph
> lp parport libcrc32c fscache mlx4_en vxlan ip6_udp_tunnel udp_tunnel
> ptp pps_core hid_generic usbhid hid mlx4_core hpsa psmouse bnx2 fjes
> scsi_transport_sas [last unloaded: target_core_mod]
> [Thu Dec 17 03:29:55 2015] CPU: 0 PID: 2547 Comm: iscsi_trx Tainted: G
>        W I     4.4.0-rc4-ede1 #1
> [Thu Dec 17 03:29:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
> P64 01/22/2015
> [Thu Dec 17 03:29:55 2015]  ffffffffc020cd47 ffff8805f1e97958
> ffffffff813ad644 0000000000000000
> [Thu Dec 17 03:29:55 2015]  ffff8805f1e97990 ffffffff81079702
> ffff8805f1e97a50 00000000015dd000
> [Thu Dec 17 03:29:55 2015]  ffff880c034df800 0000000000000200
> ffffea0000b26a80 ffff8805f1e979a0
> [Thu Dec 17 03:29:55 2015] Call Trace:
> [Thu Dec 17 03:29:55 2015]  [<ffffffff813ad644>] dump_stack+0x44/0x60
> [Thu Dec 17 03:29:55 2015]  [<ffffffff81079702>] warn_slowpath_common+0x82/0xc0
> [Thu Dec 17 03:29:55 2015]  [<ffffffff810797fa>] warn_slowpath_null+0x1a/0x20
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc01e53bb>]
> ceph_write_begin+0xfb/0x120 [ceph]
> [Thu Dec 17 03:29:55 2015]  [<ffffffff8117c8df>]
> generic_perform_write+0xbf/0x1a0
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc01dff9c>]
> ceph_write_iter+0xf5c/0x1010 [ceph]
> [Thu Dec 17 03:29:55 2015]  [<ffffffff810a888c>] ? __enqueue_entity+0x6c/0x70
> [Thu Dec 17 03:29:55 2015]  [<ffffffff813c0003>] ?
> iov_iter_get_pages+0x113/0x210
> [Thu Dec 17 03:29:55 2015]  [<ffffffff816b6802>] ?
> skb_copy_datagram_iter+0x122/0x250
> [Thu Dec 17 03:29:55 2015]  [<ffffffff811f6c93>] vfs_iter_write+0x63/0xa0
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc03c3f29>]
> fd_do_rw.isra.5+0xc9/0x1b0 [target_core_file]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc03c40d5>]
> fd_execute_rw+0xc5/0x2a0 [target_core_file]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc0445e72>]
> sbc_execute_rw+0x22/0x30 [target_core_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc04449cf>]
> __target_execute_cmd+0x1f/0x70 [target_core_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc0445525>]
> target_execute_cmd+0x195/0x2a0 [target_core_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05c978a>]
> iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05d28da>]
> iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05d8c4d>]
> iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffff8101463d>] ? __switch_to+0x1cd/0x570
> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05d8730>] ?
> iscsi_target_tx_thread+0x1c0/0x1c0 [iscsi_target_mod]
> [Thu Dec 17 03:29:55 2015]  [<ffffffff810974c9>] kthread+0xc9/0xe0
> [Thu Dec 17 03:29:55 2015]  [<ffffffff81097400>] ?
> kthread_create_on_node+0x180/0x180
> [Thu Dec 17 03:29:55 2015]  [<ffffffff817c794f>] ret_from_fork+0x3f/0x70
> [Thu Dec 17 03:29:55 2015]  [<ffffffff81097400>] ?
> kthread_create_on_node+0x180/0x180
> [Thu Dec 17 03:29:55 2015] ---[ end trace 382a45986961da4e ]---
> 
> There are WARNINGs on both line 125 and 1162. I will attached the
> whole set of dmesg output to the tracker ticket 14086
> 
> I wanted to note that file system snapshots are enabled and being used
> on this file system.
> 
> Thanks
> Eric
> 
> On Wed, Dec 16, 2015 at 8:15 AM, Eric Eastman
> <eric.eastman@keepertech.com> wrote:
> >>>
> >> This warning is really strange. Could you try the attached debug patch.
> >>
> >> Regards
> >> Yan, Zheng
> >
> > I will try the patch and get back to the list.
> >
> > Eric
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-17  9:38           ` Yan, Zheng
@ 2015-12-17 19:49             ` Eric Eastman
  2015-12-18  4:05               ` Yan, Zheng
  0 siblings, 1 reply; 20+ messages in thread
From: Eric Eastman @ 2015-12-17 19:49 UTC (permalink / raw)
  To: Yan, Zheng; +Cc: Ceph Development

With cephfs.patch and cephfs1.patch applied and I am now seeing:

[Thu Dec 17 14:27:59 2015] ------------[ cut here ]------------
[Thu Dec 17 14:27:59 2015] WARNING: CPU: 0 PID: 3036 at
fs/ceph/addr.c:1171 ceph_write_begin+0xfb/0x120 [ceph]()
[Thu Dec 17 14:27:59 2015] Modules linked in: iscsi_target_mod
vhost_scsi tcm_qla2xxx ib_srpt tcm_fc tcm_usb_gadget tcm_loop
target_core_file target_core_iblock target_core_pscsi target_core_user
target_core_mod ipmi_devintf vhost qla2xxx ib_cm ib_sa ib_mad ib_core
ib_addr libfc scsi_transport_fc libcomposite udc_core uio configfs ttm
drm_kms_helper drm ipmi_ssif coretemp gpio_ich i2c_algo_bit kvm
fb_sys_fops syscopyarea sysfillrect sysimgblt shpchp input_leds ceph
irqbypass i7core_edac serio_raw hpilo edac_core ipmi_si
ipmi_msghandler 8250_fintek lpc_ich acpi_power_meter libceph mac_hid
libcrc32c fscache bonding lp parport mlx4_en vxlan ip6_udp_tunnel
udp_tunnel ptp pps_core hid_generic usbhid hid mlx4_core hpsa psmouse
bnx2 fjes scsi_transport_sas [last unloaded: target_core_mod]
[Thu Dec 17 14:27:59 2015] CPU: 0 PID: 3036 Comm: iscsi_trx Tainted: G
       W I     4.4.0-rc4-ede2 #1
[Thu Dec 17 14:27:59 2015] Hardware name: HP ProLiant DL360 G6, BIOS
P64 01/22/2015
[Thu Dec 17 14:27:59 2015]  ffffffffc02b2e37 ffff880c0289b958
ffffffff813ad644 0000000000000000
[Thu Dec 17 14:27:59 2015]  ffff880c0289b990 ffffffff81079702
ffff880c0289ba50 0000000846c21000
[Thu Dec 17 14:27:59 2015]  ffff880c009ea200 0000000000001000
ffffea00122ed700 ffff880c0289b9a0
[Thu Dec 17 14:27:59 2015] Call Trace:
[Thu Dec 17 14:27:59 2015]  [<ffffffff813ad644>] dump_stack+0x44/0x60
[Thu Dec 17 14:27:59 2015]  [<ffffffff81079702>] warn_slowpath_common+0x82/0xc0
[Thu Dec 17 14:27:59 2015]  [<ffffffff810797fa>] warn_slowpath_null+0x1a/0x20
[Thu Dec 17 14:27:59 2015]  [<ffffffffc028b41b>]
ceph_write_begin+0xfb/0x120 [ceph]
[Thu Dec 17 14:27:59 2015]  [<ffffffff8117c8df>]
generic_perform_write+0xbf/0x1a0
[Thu Dec 17 14:27:59 2015]  [<ffffffffc0285f9c>]
ceph_write_iter+0xf5c/0x1010 [ceph]
[Thu Dec 17 14:27:59 2015]  [<ffffffff817c3396>] ? __schedule+0x386/0x9c0
[Thu Dec 17 14:27:59 2015]  [<ffffffff817c3a05>] ? schedule+0x35/0x80
[Thu Dec 17 14:27:59 2015]  [<ffffffff811d7c65>] ? __slab_free+0xb5/0x290
[Thu Dec 17 14:27:59 2015]  [<ffffffff813c0003>] ?
iov_iter_get_pages+0x113/0x210
[Thu Dec 17 14:27:59 2015]  [<ffffffff811f6c93>] vfs_iter_write+0x63/0xa0
[Thu Dec 17 14:27:59 2015]  [<ffffffffc02d2f29>]
fd_do_rw.isra.5+0xc9/0x1b0 [target_core_file]
[Thu Dec 17 14:27:59 2015]  [<ffffffffc02d30d5>]
fd_execute_rw+0xc5/0x2a0 [target_core_file]
[Thu Dec 17 14:27:59 2015]  [<ffffffffc0430e72>]
sbc_execute_rw+0x22/0x30 [target_core_mod]
[Thu Dec 17 14:27:59 2015]  [<ffffffffc042f9cf>]
__target_execute_cmd+0x1f/0x70 [target_core_mod]
[Thu Dec 17 14:27:59 2015]  [<ffffffffc0430525>]
target_execute_cmd+0x195/0x2a0 [target_core_mod]
[Thu Dec 17 14:27:59 2015]  [<ffffffffc05b778a>]
iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
[Thu Dec 17 14:27:59 2015]  [<ffffffffc05c08da>]
iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
[Thu Dec 17 14:27:59 2015]  [<ffffffffc05c6c4d>]
iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
[Thu Dec 17 14:27:59 2015]  [<ffffffff8101463d>] ? __switch_to+0x1cd/0x570
[Thu Dec 17 14:27:59 2015]  [<ffffffffc05c6730>] ?
iscsi_target_tx_thread+0x1c0/0x1c0 [iscsi_target_mod]
[Thu Dec 17 14:27:59 2015]  [<ffffffff810974c9>] kthread+0xc9/0xe0
[Thu Dec 17 14:27:59 2015]  [<ffffffff81097400>] ?
kthread_create_on_node+0x180/0x180
[Thu Dec 17 14:27:59 2015]  [<ffffffff817c794f>] ret_from_fork+0x3f/0x70
[Thu Dec 17 14:27:59 2015]  [<ffffffff81097400>] ?
kthread_create_on_node+0x180/0x180
[Thu Dec 17 14:27:59 2015] ---[ end trace 8346192e3f29ed5d ]---

Each of the WARNING on line 1171 is followed by a WARNING on line 125.
The dmesg output is attached to the tracker ticket 14086

Regards,
Eric

On Thu, Dec 17, 2015 at 2:38 AM, Yan, Zheng <ukernel@gmail.com> wrote:
> On Thu, Dec 17, 2015 at 4:56 PM, Eric Eastman
> <eric.eastman@keepertech.com> wrote:
>> I patched the 4.4rc4 kernel source and restarted the test.  Shortly
>> after starting it, this showed up in dmesg:
>>
>> [Thu Dec 17 03:29:55 2015] WARNING: CPU: 0 PID: 2547 at
>> fs/ceph/addr.c:1162 ceph_write_begin+0xfb/0x120 [ceph]()
>> [Thu Dec 17 03:29:55 2015] Modules linked in: iscsi_target_mod
>> vhost_scsi tcm_qla2xxx ib_srpt tcm_fc tcm_usb_gadget tcm_loop
>> target_core_file target_core_iblock target_core_pscsi target_core_user
>> target_core_mod ipmi_devintf vhost qla2xxx ib_cm ib_sa ib_mad ib_core
>> ib_addr libfc scsi_transport_fc libcomposite udc_core uio configfs ttm
>> ipmi_ssif drm_kms_helper drm coretemp kvm gpio_ich i2c_algo_bit
>> i7core_edac fb_sys_fops syscopyarea edac_core sysfillrect sysimgblt
>> ipmi_si input_leds hpilo ipmi_msghandler shpchp acpi_power_meter
>> irqbypass serio_raw 8250_fintek lpc_ich mac_hid ceph bonding libceph
>> lp parport libcrc32c fscache mlx4_en vxlan ip6_udp_tunnel udp_tunnel
>> ptp pps_core hid_generic usbhid hid mlx4_core hpsa psmouse bnx2 fjes
>> scsi_transport_sas [last unloaded: target_core_mod]
>> [Thu Dec 17 03:29:55 2015] CPU: 0 PID: 2547 Comm: iscsi_trx Tainted: G
>>        W I     4.4.0-rc4-ede1 #1
>> [Thu Dec 17 03:29:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
>> P64 01/22/2015
>> [Thu Dec 17 03:29:55 2015]  ffffffffc020cd47 ffff8805f1e97958
>> ffffffff813ad644 0000000000000000
>> [Thu Dec 17 03:29:55 2015]  ffff8805f1e97990 ffffffff81079702
>> ffff8805f1e97a50 00000000015dd000
>> [Thu Dec 17 03:29:55 2015]  ffff880c034df800 0000000000000200
>> ffffea0000b26a80 ffff8805f1e979a0
>> [Thu Dec 17 03:29:55 2015] Call Trace:
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff813ad644>] dump_stack+0x44/0x60
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff81079702>] warn_slowpath_common+0x82/0xc0
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff810797fa>] warn_slowpath_null+0x1a/0x20
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc01e53bb>]
>> ceph_write_begin+0xfb/0x120 [ceph]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff8117c8df>]
>> generic_perform_write+0xbf/0x1a0
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc01dff9c>]
>> ceph_write_iter+0xf5c/0x1010 [ceph]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff810a888c>] ? __enqueue_entity+0x6c/0x70
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff813c0003>] ?
>> iov_iter_get_pages+0x113/0x210
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff816b6802>] ?
>> skb_copy_datagram_iter+0x122/0x250
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff811f6c93>] vfs_iter_write+0x63/0xa0
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc03c3f29>]
>> fd_do_rw.isra.5+0xc9/0x1b0 [target_core_file]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc03c40d5>]
>> fd_execute_rw+0xc5/0x2a0 [target_core_file]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc0445e72>]
>> sbc_execute_rw+0x22/0x30 [target_core_mod]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc04449cf>]
>> __target_execute_cmd+0x1f/0x70 [target_core_mod]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc0445525>]
>> target_execute_cmd+0x195/0x2a0 [target_core_mod]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05c978a>]
>> iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05d28da>]
>> iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05d8c4d>]
>> iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff8101463d>] ? __switch_to+0x1cd/0x570
>> [Thu Dec 17 03:29:55 2015]  [<ffffffffc05d8730>] ?
>> iscsi_target_tx_thread+0x1c0/0x1c0 [iscsi_target_mod]
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff810974c9>] kthread+0xc9/0xe0
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff81097400>] ?
>> kthread_create_on_node+0x180/0x180
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff817c794f>] ret_from_fork+0x3f/0x70
>> [Thu Dec 17 03:29:55 2015]  [<ffffffff81097400>] ?
>> kthread_create_on_node+0x180/0x180
>> [Thu Dec 17 03:29:55 2015] ---[ end trace 382a45986961da4e ]---
>
>
> Could you please try the apply the new incremental patch and try again.
>
>
> Regards
> Yan, Zheng
>
>
>>
>> There are WARNINGs on both line 125 and 1162. I will attached the
>> whole set of dmesg output to the tracker ticket 14086
>>
>> I wanted to note that file system snapshots are enabled and being used
>> on this file system.
>>
>> Thanks
>> Eric
>>
>> On Wed, Dec 16, 2015 at 8:15 AM, Eric Eastman
>> <eric.eastman@keepertech.com> wrote:
>>>>>
>>>> This warning is really strange. Could you try the attached debug patch.
>>>>
>>>> Regards
>>>> Yan, Zheng
>>>
>>> I will try the patch and get back to the list.
>>>
>>> Eric

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-17 19:49             ` Eric Eastman
@ 2015-12-18  4:05               ` Yan, Zheng
       [not found]                 ` <00b601d13954$5d4a94e0$17dfbea0$@samsung.com>
  0 siblings, 1 reply; 20+ messages in thread
From: Yan, Zheng @ 2015-12-18  4:05 UTC (permalink / raw)
  To: Eric Eastman; +Cc: Ceph Development

[-- Attachment #1: Type: text/plain, Size: 4183 bytes --]

On Fri, Dec 18, 2015 at 3:49 AM, Eric Eastman
<eric.eastman@keepertech.com> wrote:
> With cephfs.patch and cephfs1.patch applied and I am now seeing:
>
> [Thu Dec 17 14:27:59 2015] ------------[ cut here ]------------
> [Thu Dec 17 14:27:59 2015] WARNING: CPU: 0 PID: 3036 at
> fs/ceph/addr.c:1171 ceph_write_begin+0xfb/0x120 [ceph]()
> [Thu Dec 17 14:27:59 2015] Modules linked in: iscsi_target_mod
> vhost_scsi tcm_qla2xxx ib_srpt tcm_fc tcm_usb_gadget tcm_loop
> target_core_file target_core_iblock target_core_pscsi target_core_user
> target_core_mod ipmi_devintf vhost qla2xxx ib_cm ib_sa ib_mad ib_core
> ib_addr libfc scsi_transport_fc libcomposite udc_core uio configfs ttm
> drm_kms_helper drm ipmi_ssif coretemp gpio_ich i2c_algo_bit kvm
> fb_sys_fops syscopyarea sysfillrect sysimgblt shpchp input_leds ceph
> irqbypass i7core_edac serio_raw hpilo edac_core ipmi_si
> ipmi_msghandler 8250_fintek lpc_ich acpi_power_meter libceph mac_hid
> libcrc32c fscache bonding lp parport mlx4_en vxlan ip6_udp_tunnel
> udp_tunnel ptp pps_core hid_generic usbhid hid mlx4_core hpsa psmouse
> bnx2 fjes scsi_transport_sas [last unloaded: target_core_mod]
> [Thu Dec 17 14:27:59 2015] CPU: 0 PID: 3036 Comm: iscsi_trx Tainted: G
>        W I     4.4.0-rc4-ede2 #1
> [Thu Dec 17 14:27:59 2015] Hardware name: HP ProLiant DL360 G6, BIOS
> P64 01/22/2015
> [Thu Dec 17 14:27:59 2015]  ffffffffc02b2e37 ffff880c0289b958
> ffffffff813ad644 0000000000000000
> [Thu Dec 17 14:27:59 2015]  ffff880c0289b990 ffffffff81079702
> ffff880c0289ba50 0000000846c21000
> [Thu Dec 17 14:27:59 2015]  ffff880c009ea200 0000000000001000
> ffffea00122ed700 ffff880c0289b9a0
> [Thu Dec 17 14:27:59 2015] Call Trace:
> [Thu Dec 17 14:27:59 2015]  [<ffffffff813ad644>] dump_stack+0x44/0x60
> [Thu Dec 17 14:27:59 2015]  [<ffffffff81079702>] warn_slowpath_common+0x82/0xc0
> [Thu Dec 17 14:27:59 2015]  [<ffffffff810797fa>] warn_slowpath_null+0x1a/0x20
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc028b41b>]
> ceph_write_begin+0xfb/0x120 [ceph]
> [Thu Dec 17 14:27:59 2015]  [<ffffffff8117c8df>]
> generic_perform_write+0xbf/0x1a0
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc0285f9c>]
> ceph_write_iter+0xf5c/0x1010 [ceph]
> [Thu Dec 17 14:27:59 2015]  [<ffffffff817c3396>] ? __schedule+0x386/0x9c0
> [Thu Dec 17 14:27:59 2015]  [<ffffffff817c3a05>] ? schedule+0x35/0x80
> [Thu Dec 17 14:27:59 2015]  [<ffffffff811d7c65>] ? __slab_free+0xb5/0x290
> [Thu Dec 17 14:27:59 2015]  [<ffffffff813c0003>] ?
> iov_iter_get_pages+0x113/0x210
> [Thu Dec 17 14:27:59 2015]  [<ffffffff811f6c93>] vfs_iter_write+0x63/0xa0
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc02d2f29>]
> fd_do_rw.isra.5+0xc9/0x1b0 [target_core_file]
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc02d30d5>]
> fd_execute_rw+0xc5/0x2a0 [target_core_file]
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc0430e72>]
> sbc_execute_rw+0x22/0x30 [target_core_mod]
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc042f9cf>]
> __target_execute_cmd+0x1f/0x70 [target_core_mod]
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc0430525>]
> target_execute_cmd+0x195/0x2a0 [target_core_mod]
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc05b778a>]
> iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc05c08da>]
> iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc05c6c4d>]
> iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
> [Thu Dec 17 14:27:59 2015]  [<ffffffff8101463d>] ? __switch_to+0x1cd/0x570
> [Thu Dec 17 14:27:59 2015]  [<ffffffffc05c6730>] ?
> iscsi_target_tx_thread+0x1c0/0x1c0 [iscsi_target_mod]
> [Thu Dec 17 14:27:59 2015]  [<ffffffff810974c9>] kthread+0xc9/0xe0
> [Thu Dec 17 14:27:59 2015]  [<ffffffff81097400>] ?
> kthread_create_on_node+0x180/0x180
> [Thu Dec 17 14:27:59 2015]  [<ffffffff817c794f>] ret_from_fork+0x3f/0x70
> [Thu Dec 17 14:27:59 2015]  [<ffffffff81097400>] ?
> kthread_create_on_node+0x180/0x180
> [Thu Dec 17 14:27:59 2015] ---[ end trace 8346192e3f29ed5d ]---
>

The page gets unlocked mystically. I still don't find any clue. Could
you please try the new patch (not incremental patch). Besides, please
enable CONFIG_DEBUG_VM when compiling the kernel.

Thanks you very much
Yan, Zheng

[-- Attachment #2: cephfs_new.patch --]
[-- Type: application/octet-stream, Size: 4005 bytes --]

diff --git a/fs/ceph/addr.c b/fs/ceph/addr.c
index b7d218a..3b69f8f 100644
--- a/fs/ceph/addr.c
+++ b/fs/ceph/addr.c
@@ -199,6 +199,8 @@ static int readpage_nounlock(struct file *filp, struct page *page)
 	u64 off = page_offset(page);
 	u64 len = PAGE_CACHE_SIZE;
 
+	WARN_ON(!PageLocked(page));
+
 	if (off >= i_size_read(inode)) {
 		zero_user_segment(page, 0, PAGE_CACHE_SIZE);
 		SetPageUptodate(page);
@@ -487,6 +489,7 @@ static int writepage_nounlock(struct page *page, struct writeback_control *wbc)
 	int err = 0, len = PAGE_CACHE_SIZE;
 
 	dout("writepage %p idx %lu\n", page, page->index);
+	WARN_ON(!PageLocked(page));
 
 	if (!page->mapping || !page->mapping->host) {
 		dout("writepage %p - no mapping\n", page);
@@ -1055,9 +1058,15 @@ static int ceph_update_writeable_page(struct file *file,
 	int r;
 	struct ceph_snap_context *snapc, *oldest;
 
+	int last_op = 0;
 retry_locked:
 	/* writepages currently holds page lock, but if we change that later, */
+	WARN_ON(PageWriteback(page));
+	if (!PageLocked(page))
+		pr_err("page is unlocked before wait writeback %p, last op %d\n", page, last_op);
 	wait_on_page_writeback(page);
+	if (!PageLocked(page))
+		pr_err("page is unlocked after wait writeback %p last op %d\n", page, last_op);
 
 	snapc = page_snap_context(page);
 	if (snapc && snapc != ci->i_head_snapc) {
@@ -1081,18 +1090,26 @@ retry_locked:
 			r = wait_event_interruptible(ci->i_cap_wq,
 			       context_is_writeable_or_written(inode, snapc));
 			ceph_put_snap_context(snapc);
-			if (r == -ERESTARTSYS)
-				return r;
-			return -EAGAIN;
+			if (r != -ERESTARTSYS)
+				r = -EAGAIN;
+			pr_err("update_writeable_page no snapc %p %d\n", page, r);
+			return r;
 		}
 		ceph_put_snap_context(oldest);
 
 		/* yay, writeable, do it now (without dropping page lock) */
 		dout(" page %p snapc %p not current, but oldest\n",
 		     page, snapc);
-		if (!clear_page_dirty_for_io(page))
+		if (!clear_page_dirty_for_io(page)) {
+			last_op = 1;
+			if (!PageLocked(page))
+				pr_err("clear_page_dirty_for_io unlocks %p\n", page);
 			goto retry_locked;
+		}
 		r = writepage_nounlock(page, NULL);
+		last_op = 2;
+		if (!PageLocked(page))
+			pr_err("writepage_nounlock unlocks %p %d\n", page, r);
 		if (r < 0)
 			goto fail_nosnap;
 		goto retry_locked;
@@ -1123,11 +1140,15 @@ retry_locked:
 
 	/* we need to read it. */
 	r = readpage_nounlock(file, page);
+	last_op = 3;
+	if (!PageLocked(page))
+		pr_err("readpage_nounlock unlocks %p %d\n", page, r);
 	if (r < 0)
 		goto fail_nosnap;
 	goto retry_locked;
 fail_nosnap:
 	unlock_page(page);
+	pr_err("update_writeable_page failed %p %d\n", page, r);
 	return r;
 }
 
@@ -1149,16 +1170,19 @@ static int ceph_write_begin(struct file *file, struct address_space *mapping,
 		page = grab_cache_page_write_begin(mapping, index, 0);
 		if (!page)
 			return -ENOMEM;
+		WARN_ON(!PageLocked(page));
 		*pagep = page;
 
 		dout("write_begin file %p inode %p page %p %d~%d\n", file,
 		     inode, page, (int)pos, (int)len);
 
 		r = ceph_update_writeable_page(file, pos, len, page);
-		if (r < 0)
+		if (r < 0) {
 			page_cache_release(page);
-		else
+		} else {
+			WARN_ON(!PageLocked(page));
 			*pagep = page;
+		}
 	} while (r == -EAGAIN);
 
 	return r;
@@ -1378,11 +1402,13 @@ static int ceph_page_mkwrite(struct vm_area_struct *vma, struct vm_fault *vmf)
 
 	ret = VM_FAULT_NOPAGE;
 	if ((off > size) ||
-	    (page->mapping != inode->i_mapping))
+	    (page->mapping != inode->i_mapping)) {
+		unlock_page(page);
 		goto out;
+	}
 
 	ret = ceph_update_writeable_page(vma->vm_file, off, len, page);
-	if (ret == 0) {
+	if (ret >= 0) {
 		/* success.  we'll keep the page locked. */
 		set_page_dirty(page);
 		ret = VM_FAULT_LOCKED;
@@ -1393,8 +1419,6 @@ static int ceph_page_mkwrite(struct vm_area_struct *vma, struct vm_fault *vmf)
 			ret = VM_FAULT_SIGBUS;
 	}
 out:
-	if (ret != VM_FAULT_LOCKED)
-		unlock_page(page);
 	if (ret == VM_FAULT_LOCKED ||
 	    ci->i_inline_version != CEPH_INLINE_NONE) {
 		int dirty;

^ permalink raw reply related	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
       [not found]                 ` <00b601d13954$5d4a94e0$17dfbea0$@samsung.com>
@ 2015-12-18  6:23                   ` Eric Eastman
  2015-12-18  7:18                     ` Yan, Zheng
  0 siblings, 1 reply; 20+ messages in thread
From: Eric Eastman @ 2015-12-18  6:23 UTC (permalink / raw)
  To: Chao Yu, Yan, Zheng, Ceph Development

> Hi Yan Zheng, Eric Eastman
>
> Similar bug was reported in f2fs, btrfs, it does affect 4.4-rc4, the fixing
> patch was merged into 4.4-rc5, dfd01f026058 ("sched/wait: Fix the signal
> handling fix").
>
> Related report & discussion was here:
> https://lkml.org/lkml/2015/12/12/149
>
> I'm not sure the current reported issue of ceph was related to that though,
> but at least try testing with an upgraded or patched kernel could verify it.
> :)
>
> Thanks,
>
>> -----Original Message-----
>> From: ceph-devel-owner@vger.kernel.org [mailto:ceph-devel-owner@vger.kernel.org] On Behalf Of
>> Yan, Zheng
>> Sent: Friday, December 18, 2015 12:05 PM
>> To: Eric Eastman
>> Cc: Ceph Development
>> Subject: Re: Issue with Ceph File System and LIO
>>
>> On Fri, Dec 18, 2015 at 3:49 AM, Eric Eastman
>> <eric.eastman@keepertech.com> wrote:
>> > With cephfs.patch and cephfs1.patch applied and I am now seeing:
>> >
>> > [Thu Dec 17 14:27:59 2015] ------------[ cut here ]------------
>> > [Thu Dec 17 14:27:59 2015] WARNING: CPU: 0 PID: 3036 at
>> > fs/ceph/addr.c:1171 ceph_write_begin+0xfb/0x120 [ceph]()
>> > [Thu Dec 17 14:27:59 2015] Modules linked in: iscsi_target_mod
...
>> >
>>
>> The page gets unlocked mystically. I still don't find any clue. Could
>> you please try the new patch (not incremental patch). Besides, please
>> enable CONFIG_DEBUG_VM when compiling the kernel.
>>
>> Thanks you very much
>> Yan, Zheng
>
I have just installed the cephfs_new.patch and have set
CONFIG_DEBUG_VM=y on a new 4.4rc4 kernel and restarted the ESXi iSCSI
test to my Ceph File System gateway.  I plan to let it run overnight
and report the status tomorrow.

Let me know if I should move on to 4.4rc5 with or without patches and
with or without  CONFIG_DEBUG_VM=y

Looking at the network traffic stats on my iSCSI gateway, with
CONFIG_DEBUG_VM=y, throughput seems to be down by a factor of at least
10 compared to my last test without setting CONFIG_DEBUG_VM=y

Regards,
Eric

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-18  6:23                   ` Eric Eastman
@ 2015-12-18  7:18                     ` Yan, Zheng
  2015-12-18 15:53                       ` Eric Eastman
  2015-12-21  2:38                       ` Eric Eastman
  0 siblings, 2 replies; 20+ messages in thread
From: Yan, Zheng @ 2015-12-18  7:18 UTC (permalink / raw)
  To: Eric Eastman; +Cc: Chao Yu, Ceph Development

On Fri, Dec 18, 2015 at 2:23 PM, Eric Eastman
<eric.eastman@keepertech.com> wrote:
>> Hi Yan Zheng, Eric Eastman
>>
>> Similar bug was reported in f2fs, btrfs, it does affect 4.4-rc4, the fixing
>> patch was merged into 4.4-rc5, dfd01f026058 ("sched/wait: Fix the signal
>> handling fix").
>>
>> Related report & discussion was here:
>> https://lkml.org/lkml/2015/12/12/149
>>
>> I'm not sure the current reported issue of ceph was related to that though,
>> but at least try testing with an upgraded or patched kernel could verify it.
>> :)
>>
>> Thanks,
>>
>>> -----Original Message-----
>>> From: ceph-devel-owner@vger.kernel.org [mailto:ceph-devel-owner@vger.kernel.org] On Behalf Of
>>> Yan, Zheng
>>> Sent: Friday, December 18, 2015 12:05 PM
>>> To: Eric Eastman
>>> Cc: Ceph Development
>>> Subject: Re: Issue with Ceph File System and LIO
>>>
>>> On Fri, Dec 18, 2015 at 3:49 AM, Eric Eastman
>>> <eric.eastman@keepertech.com> wrote:
>>> > With cephfs.patch and cephfs1.patch applied and I am now seeing:
>>> >
>>> > [Thu Dec 17 14:27:59 2015] ------------[ cut here ]------------
>>> > [Thu Dec 17 14:27:59 2015] WARNING: CPU: 0 PID: 3036 at
>>> > fs/ceph/addr.c:1171 ceph_write_begin+0xfb/0x120 [ceph]()
>>> > [Thu Dec 17 14:27:59 2015] Modules linked in: iscsi_target_mod
> ...
>>> >
>>>
>>> The page gets unlocked mystically. I still don't find any clue. Could
>>> you please try the new patch (not incremental patch). Besides, please
>>> enable CONFIG_DEBUG_VM when compiling the kernel.
>>>
>>> Thanks you very much
>>> Yan, Zheng
>>
> I have just installed the cephfs_new.patch and have set
> CONFIG_DEBUG_VM=y on a new 4.4rc4 kernel and restarted the ESXi iSCSI
> test to my Ceph File System gateway.  I plan to let it run overnight
> and report the status tomorrow.
>
> Let me know if I should move on to 4.4rc5 with or without patches and
> with or without  CONFIG_DEBUG_VM=y
>

please try rc5 kernel without patches and DEBUG_VM=y

Regards
Yan, Zheng


> Looking at the network traffic stats on my iSCSI gateway, with
> CONFIG_DEBUG_VM=y, throughput seems to be down by a factor of at least
> 10 compared to my last test without setting CONFIG_DEBUG_VM=y
>
> Regards,
> Eric

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-18  7:18                     ` Yan, Zheng
@ 2015-12-18 15:53                       ` Eric Eastman
  2015-12-21  2:38                       ` Eric Eastman
  1 sibling, 0 replies; 20+ messages in thread
From: Eric Eastman @ 2015-12-18 15:53 UTC (permalink / raw)
  To: Yan, Zheng; +Cc: Chao Yu, Ceph Development

On Fri, Dec 18, 2015 at 12:18 AM, Yan, Zheng <ukernel@gmail.com> wrote:
> On Fri, Dec 18, 2015 at 2:23 PM, Eric Eastman
> <eric.eastman@keepertech.com> wrote:
>>> Hi Yan Zheng, Eric Eastman
>>>
>>> Similar bug was reported in f2fs, btrfs, it does affect 4.4-rc4, the fixing
>>> patch was merged into 4.4-rc5, dfd01f026058 ("sched/wait: Fix the signal
>>> handling fix").
>>>
>>> Related report & discussion was here:
>>> https://lkml.org/lkml/2015/12/12/149
>>>
>>> I'm not sure the current reported issue of ceph was related to that though,
>>> but at least try testing with an upgraded or patched kernel could verify it.
>>> :)
>>>
>>> Thanks,
>>>
>>>> -----Original Message-----
>>>> From: ceph-devel-owner@vger.kernel.org [mailto:ceph-devel-owner@vger.kernel.org] On Behalf Of
>>>> Yan, Zheng
>>>> Sent: Friday, December 18, 2015 12:05 PM
>>>> To: Eric Eastman
>>>>
>>>> The page gets unlocked mystically. I still don't find any clue. Could
>>>> you please try the new patch (not incremental patch). Besides, please
>>>> enable CONFIG_DEBUG_VM when compiling the kernel.
>>>>
>>>> Thanks you very much
>>>> Yan, Zheng
>>>
>> I have just installed the cephfs_new.patch and have set
>> CONFIG_DEBUG_VM=y on a new 4.4rc4 kernel and restarted the ESXi iSCSI
>> test to my Ceph File System gateway.  I plan to let it run overnight
>> and report the status tomorrow.
>>
>> Let me know if I should move on to 4.4rc5 with or without patches and
>> with or without  CONFIG_DEBUG_VM=y
>>
>
> please try rc5 kernel without patches and DEBUG_VM=y
>
> Regards
> Yan, Zheng
>

With the 4.4rc4 kernel and the cephfs_new.patch and CONFIG_DEBUG_VM=y
I hit a BUG in mm/filemap.c. I will start a test with the 4.4rc5
kernel and get back to the list.  I put the whole dmesg -T output
showing this BUG and some Ceph Warnings in the tracker ticket #14086

Fri Dec 18 01:14:39 2015] kernel BUG at mm/filemap.c:812!
[Fri Dec 18 01:14:39 2015] invalid opcode: 0000 [#1] SMP
[Fri Dec 18 01:14:39 2015] Modules linked in: iscsi_target_mod
vhost_scsi tcm_qla2xxx ib_srpt tcm_fc tcm_usb_gadget tcm_loop
target_core_file target_core_iblock target_core_pscsi target_core_user
target_core_mod ipmi_devintf vhost qla2xxx ib_cm ib_sa ib_mad ib_core
ib_addr libfc scsi_transport_fc libcomposite udc_core uio configfs ttm
drm_kms_helper coretemp drm kvm ipmi_ssif gpio_ich ceph i2c_algo_bit
fb_sys_fops syscopyarea input_leds sysfillrect sysimgblt irqbypass
shpchp hpilo serio_raw acpi_power_meter ipmi_si lpc_ich i7core_edac
ipmi_msghandler edac_core 8250_fintek libceph mac_hid libcrc32c
fscache bonding lp parport mlx4_en vxlan ip6_udp_tunnel udp_tunnel ptp
pps_core hid_generic usbhid hid mlx4_core psmouse hpsa bnx2 fjes
scsi_transport_sas [last unloaded: target_core_mod]
[Fri Dec 18 01:14:39 2015] CPU: 0 PID: 2147 Comm: iscsi_trx Tainted: G
       W I     4.4.0-rc4-ede3-DEBUG_VM #1
[Fri Dec 18 01:14:39 2015] Hardware name: HP ProLiant DL360 G6, BIOS
P64 01/22/2015
[Fri Dec 18 01:14:39 2015] task: ffff880c02077080 ti: ffff880bfce5c000
task.ti: ffff880bfce5c000
[Fri Dec 18 01:14:39 2015] RIP: 0010:[<ffffffff8117d041>]
[<ffffffff8117d041>] unlock_page+0x81/0x90
[Fri Dec 18 01:14:39 2015] RSP: 0018:ffff880bfce5f9b8  EFLAGS: 00010282
[Fri Dec 18 01:14:39 2015] RAX: 0000000000000021 RBX: ffffea0015d36ac0
RCX: 0000000000000000
[Fri Dec 18 01:14:40 2015] RDX: 0000000000000021 RSI: ffff880607a0dc78
RDI: ffff880607a0dc78
[Fri Dec 18 01:14:40 2015] RBP: ffff880bfce5f9b8 R08: 0000000000000000
R09: 000000000000041e
[Fri Dec 18 01:14:40 2015] R10: 0000000000000246 R11: 000000000000041e
R12: 0000000000000000
[Fri Dec 18 01:14:40 2015] R13: 0000000000001000 R14: ffff8800dad39f88
R15: 0000000000001000
[Fri Dec 18 01:14:40 2015] FS:  0000000000000000(0000)
GS:ffff880607a00000(0000) knlGS:0000000000000000
[Fri Dec 18 01:14:40 2015] CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
[Fri Dec 18 01:14:40 2015] CR2: 00000000008e6f90 CR3: 0000000001c0a000
CR4: 00000000000006f0
[Fri Dec 18 01:14:40 2015] Stack:
[Fri Dec 18 01:14:40 2015]  ffff880bfce5fa00 ffffffffc02d4ac6
ffff880bfce5fa00 ffffffff813c5456
[Fri Dec 18 01:14:40 2015]  0000000000001000 000000081e838000
ffff880bfce5fc80 0000000000000000
[Fri Dec 18 01:14:40 2015]  ffff8800dad3a0f0 ffff880bfce5fa88
ffffffff8117cb95 ffff88002cd5ec80
[Fri Dec 18 01:14:40 2015] Call Trace:
[Fri Dec 18 01:14:40 2015]  [<ffffffffc02d4ac6>]
ceph_write_end+0x66/0x180 [ceph]
[Fri Dec 18 01:14:40 2015]  [<ffffffff813c5456>] ?
iov_iter_copy_from_user_atomic+0x156/0x220
[Fri Dec 18 01:14:40 2015]  [<ffffffff8117cb95>]
generic_perform_write+0x105/0x1a0
[Fri Dec 18 01:14:40 2015]  [<ffffffffc02cff9c>]
ceph_write_iter+0xf5c/0x1010 [ceph]
[Fri Dec 18 01:14:40 2015]  [<ffffffff817c8af6>] ? __schedule+0x386/0x9c0
[Fri Dec 18 01:14:40 2015]  [<ffffffff817c9165>] ? schedule+0x35/0x80
[Fri Dec 18 01:14:40 2015]  [<ffffffff813c0003>] ?
insn_get_immediate.part.8+0x293/0x300
[Fri Dec 18 01:14:40 2015]  [<ffffffff816bbeb2>] ?
skb_copy_datagram_iter+0x122/0x250
[Fri Dec 18 01:14:40 2015]  [<ffffffff811fbe23>] vfs_iter_write+0x63/0xa0
[Fri Dec 18 01:14:40 2015]  [<ffffffffc0252f29>]
fd_do_rw.isra.5+0xc9/0x1b0 [target_core_file]
[Fri Dec 18 01:14:40 2015]  [<ffffffffc02530d5>]
fd_execute_rw+0xc5/0x2a0 [target_core_file]
[Fri Dec 18 01:14:40 2015]  [<ffffffffc0352e72>]
sbc_execute_rw+0x22/0x30 [target_core_mod]
[Fri Dec 18 01:14:40 2015]  [<ffffffffc03519cf>]
__target_execute_cmd+0x1f/0x70 [target_core_mod]
[Fri Dec 18 01:14:40 2015]  [<ffffffffc0352525>]
target_execute_cmd+0x195/0x2a0 [target_core_mod]
[Fri Dec 18 01:14:40 2015]  [<ffffffffc05bf78a>]
iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
[Fri Dec 18 01:14:40 2015]  [<ffffffffc05c88da>]
iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
[Fri Dec 18 01:14:40 2015]  [<ffffffffc05cec4d>]
iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
[Fri Dec 18 01:14:40 2015]  [<ffffffff8101463d>] ? __switch_to+0x1cd/0x570
[Fri Dec 18 01:14:40 2015]  [<ffffffffc05ce730>] ?
iscsi_target_tx_thread+0x1c0/0x1c0 [iscsi_target_mod]
[Fri Dec 18 01:14:40 2015]  [<ffffffff81097859>] kthread+0xc9/0xe0
[Fri Dec 18 01:14:40 2015]  [<ffffffff81097790>] ?
kthread_create_on_node+0x180/0x180
[Fri Dec 18 01:14:40 2015]  [<ffffffff817cd0cf>] ret_from_fork+0x3f/0x70
[Fri Dec 18 01:14:40 2015]  [<ffffffff81097790>] ?
kthread_create_on_node+0x180/0x180
[Fri Dec 18 01:14:40 2015] Code: b8 00 00 00 48 8b 80 a8 00 00 00 48
d3 ea 48 8d 14 52 48 8d 3c d0 31 d2 e8 2d cb f3 ff 5d c3 48 c7 c6 a0
12 ad 81 e8 3f c5 02 00 <0f> 0b 66 66 66 66 2e 0f 1f 84 00 00 00 00 00
66 66 66 66 90 55
[Fri Dec 18 01:14:40 2015] RIP  [<ffffffff8117d041>] unlock_page+0x81/0x90
[Fri Dec 18 01:14:40 2015]  RSP <ffff880bfce5f9b8>
[Fri Dec 18 01:14:40 2015] ---[ end trace d2dd732cc24afbf8 ]---

Regards,
Eric Eastman

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-15  6:08 Issue with Ceph File System and LIO Eric Eastman
  2015-12-15  9:26 ` Mike Christie
  2015-12-15 12:05 ` Yan, Zheng
@ 2015-12-18 21:54 ` Mike Christie
  2015-12-19  2:13   ` Eric Eastman
  2 siblings, 1 reply; 20+ messages in thread
From: Mike Christie @ 2015-12-18 21:54 UTC (permalink / raw)
  To: Eric Eastman, Ceph Development

Eric,

Do you have iSCSI data digests on?

On 12/15/2015 12:08 AM, Eric Eastman wrote:
> I am testing Linux Target SCSI, LIO, with a Ceph File System backstore
> and I am seeing this error on my LIO gateway.  I am using Ceph v9.2.0
> on a 4.4rc4 Kernel, on Trusty, using a kernel mounted Ceph File
> System.  A file on the Ceph File System is exported via iSCSI to a
> VMware ESXi 5.0 server, and I am seeing this error when doing a lot of
> I/O on the ESXi server.   Is this a LIO or a Ceph issue?
> 
> [Tue Dec 15 00:46:55 2015] ------------[ cut here ]------------
> [Tue Dec 15 00:46:55 2015] WARNING: CPU: 0 PID: 1123421 at
> /home/kernel/COD/linux/fs/ceph/addr.c:125
> ceph_set_page_dirty+0x230/0x240 [ceph]()
> [Tue Dec 15 00:46:55 2015] Modules linked in: iptable_filter ip_tables
> x_tables xfs rbd iscsi_target_mod vhost_scsi tcm_qla2xxx ib_srpt
> tcm_fc tcm_usb_gadget tcm_loop target_core_file target_core_iblock
> target_core_pscsi target_core_user target_core_mod ipmi_devintf vhost
> qla2xxx ib_cm ib_sa ib_mad ib_core ib_addr libfc scsi_transport_fc
> libcomposite udc_core uio configfs ipmi_ssif ttm drm_kms_helper
> gpio_ich drm i2c_algo_bit fb_sys_fops coretemp syscopyarea ipmi_si
> sysfillrect ipmi_msghandler sysimgblt kvm acpi_power_meter 8250_fintek
> irqbypass hpilo shpchp input_leds serio_raw lpc_ich i7core_edac
> edac_core mac_hid ceph libceph libcrc32c fscache bonding lp parport
> mlx4_en vxlan ip6_udp_tunnel udp_tunnel ptp pps_core hid_generic
> usbhid hid hpsa mlx4_core psmouse bnx2 scsi_transport_sas fjes [last
> unloaded: target_core_mod]
> [Tue Dec 15 00:46:55 2015] CPU: 0 PID: 1123421 Comm: iscsi_trx
> Tainted: G        W I     4.4.0-040400rc4-generic #201512061930
> [Tue Dec 15 00:46:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
> P64 01/22/2015
> [Tue Dec 15 00:46:55 2015]  0000000000000000 00000000fdc0ce43
> ffff880bf38c38c0 ffffffff813c8ab4
> [Tue Dec 15 00:46:55 2015]  0000000000000000 ffff880bf38c38f8
> ffffffff8107d772 ffffea00127a8680
> [Tue Dec 15 00:46:55 2015]  ffff8804e52c1448 ffff8804e52c15b0
> ffff8804e52c10f0 0000000000000200
> [Tue Dec 15 00:46:55 2015] Call Trace:
> [Tue Dec 15 00:46:55 2015]  [<ffffffff813c8ab4>] dump_stack+0x44/0x60
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d772>] warn_slowpath_common+0x82/0xc0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d8ba>] warn_slowpath_null+0x1a/0x20
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fadb0>]
> ceph_set_page_dirty+0x230/0x240 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff81188770>] ?
> pagecache_get_page+0x150/0x1c0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fe338>] ?
> ceph_pool_perm_check+0x48/0x700 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8119301d>] set_page_dirty+0x3d/0x70
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fcd7e>]
> ceph_write_end+0x5e/0x180 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dc006>] ?
> iov_iter_copy_from_user_atomic+0x156/0x220
> [Tue Dec 15 00:46:55 2015]  [<ffffffff81187bc4>]
> generic_perform_write+0x114/0x1c0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01f818a>]
> ceph_write_iter+0xf8a/0x1050 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0205983>] ?
> ceph_put_cap_refs+0x143/0x320 [ceph]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff810b10ba>] ?
> check_preempt_wakeup+0xfa/0x220
> [Tue Dec 15 00:46:55 2015]  [<ffffffff811a7eec>] ? zone_statistics+0x7c/0xa0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dd2ee>] ? copy_page_to_iter+0x5e/0xa0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff816e5d22>] ?
> skb_copy_datagram_iter+0x122/0x250
> [Tue Dec 15 00:46:55 2015]  [<ffffffff812053f6>] vfs_iter_write+0x76/0xc0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cbf88>]
> fd_do_rw.isra.5+0xd8/0x1e0 [target_core_file]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cc155>]
> fd_execute_rw+0xc5/0x2a0 [target_core_file]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04696f2>]
> sbc_execute_rw+0x22/0x30 [target_core_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04681ef>]
> __target_execute_cmd+0x1f/0x70 [target_core_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0468da5>]
> target_execute_cmd+0x195/0x2a0 [target_core_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05db89a>]
> iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05e4aea>]
> iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eafbd>]
> iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8101566c>] ? __switch_to+0x1dc/0x5a0
> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eaaa0>] ?
> iscsi_target_tx_thread+0x1e0/0x1e0 [iscsi_target_mod]
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c8b8>] kthread+0xd8/0xf0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
> kthread_create_on_node+0x1a0/0x1a0
> [Tue Dec 15 00:46:55 2015]  [<ffffffff817fc58f>] ret_from_fork+0x3f/0x70
> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
> kthread_create_on_node+0x1a0/0x1a0
> [Tue Dec 15 00:46:55 2015] ---[ end trace 4079437668c77cbb ]---
> [Tue Dec 15 00:47:45 2015] ABORT_TASK: Found referenced iSCSI task_tag: 95784927
> [Tue Dec 15 00:47:45 2015] ABORT_TASK: ref_tag: 95784927 already
> complete, skipping
> 
> If it is a Ceph File System issue, let me know and I will open a bug.
> 
> Thanks
> 
> Eric
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 


^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-18 21:54 ` Mike Christie
@ 2015-12-19  2:13   ` Eric Eastman
  0 siblings, 0 replies; 20+ messages in thread
From: Eric Eastman @ 2015-12-19  2:13 UTC (permalink / raw)
  To: Mike Christie; +Cc: Ceph Development

Hi Mike,

On the EXSi server both Header Digest and Data Digest are set to Prohibited.

Eric

On Fri, Dec 18, 2015 at 2:54 PM, Mike Christie <mchristi@redhat.com> wrote:
> Eric,
>
> Do you have iSCSI data digests on?
>
> On 12/15/2015 12:08 AM, Eric Eastman wrote:
>> I am testing Linux Target SCSI, LIO, with a Ceph File System backstore
>> and I am seeing this error on my LIO gateway.  I am using Ceph v9.2.0
>> on a 4.4rc4 Kernel, on Trusty, using a kernel mounted Ceph File
>> System.  A file on the Ceph File System is exported via iSCSI to a
>> VMware ESXi 5.0 server, and I am seeing this error when doing a lot of
>> I/O on the ESXi server.   Is this a LIO or a Ceph issue?
>>
>> [Tue Dec 15 00:46:55 2015] ------------[ cut here ]------------
>> [Tue Dec 15 00:46:55 2015] WARNING: CPU: 0 PID: 1123421 at
>> /home/kernel/COD/linux/fs/ceph/addr.c:125
>> ceph_set_page_dirty+0x230/0x240 [ceph]()
>> [Tue Dec 15 00:46:55 2015] Modules linked in: iptable_filter ip_tables
>> x_tables xfs rbd iscsi_target_mod vhost_scsi tcm_qla2xxx ib_srpt
>> tcm_fc tcm_usb_gadget tcm_loop target_core_file target_core_iblock
>> target_core_pscsi target_core_user target_core_mod ipmi_devintf vhost
>> qla2xxx ib_cm ib_sa ib_mad ib_core ib_addr libfc scsi_transport_fc
>> libcomposite udc_core uio configfs ipmi_ssif ttm drm_kms_helper
>> gpio_ich drm i2c_algo_bit fb_sys_fops coretemp syscopyarea ipmi_si
>> sysfillrect ipmi_msghandler sysimgblt kvm acpi_power_meter 8250_fintek
>> irqbypass hpilo shpchp input_leds serio_raw lpc_ich i7core_edac
>> edac_core mac_hid ceph libceph libcrc32c fscache bonding lp parport
>> mlx4_en vxlan ip6_udp_tunnel udp_tunnel ptp pps_core hid_generic
>> usbhid hid hpsa mlx4_core psmouse bnx2 scsi_transport_sas fjes [last
>> unloaded: target_core_mod]
>> [Tue Dec 15 00:46:55 2015] CPU: 0 PID: 1123421 Comm: iscsi_trx
>> Tainted: G        W I     4.4.0-040400rc4-generic #201512061930
>> [Tue Dec 15 00:46:55 2015] Hardware name: HP ProLiant DL360 G6, BIOS
>> P64 01/22/2015
>> [Tue Dec 15 00:46:55 2015]  0000000000000000 00000000fdc0ce43
>> ffff880bf38c38c0 ffffffff813c8ab4
>> [Tue Dec 15 00:46:55 2015]  0000000000000000 ffff880bf38c38f8
>> ffffffff8107d772 ffffea00127a8680
>> [Tue Dec 15 00:46:55 2015]  ffff8804e52c1448 ffff8804e52c15b0
>> ffff8804e52c10f0 0000000000000200
>> [Tue Dec 15 00:46:55 2015] Call Trace:
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff813c8ab4>] dump_stack+0x44/0x60
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d772>] warn_slowpath_common+0x82/0xc0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8107d8ba>] warn_slowpath_null+0x1a/0x20
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fadb0>]
>> ceph_set_page_dirty+0x230/0x240 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff81188770>] ?
>> pagecache_get_page+0x150/0x1c0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fe338>] ?
>> ceph_pool_perm_check+0x48/0x700 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8119301d>] set_page_dirty+0x3d/0x70
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01fcd7e>]
>> ceph_write_end+0x5e/0x180 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dc006>] ?
>> iov_iter_copy_from_user_atomic+0x156/0x220
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff81187bc4>]
>> generic_perform_write+0x114/0x1c0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc01f818a>]
>> ceph_write_iter+0xf8a/0x1050 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0205983>] ?
>> ceph_put_cap_refs+0x143/0x320 [ceph]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff810b10ba>] ?
>> check_preempt_wakeup+0xfa/0x220
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff811a7eec>] ? zone_statistics+0x7c/0xa0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff813dd2ee>] ? copy_page_to_iter+0x5e/0xa0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff816e5d22>] ?
>> skb_copy_datagram_iter+0x122/0x250
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff812053f6>] vfs_iter_write+0x76/0xc0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cbf88>]
>> fd_do_rw.isra.5+0xd8/0x1e0 [target_core_file]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc02cc155>]
>> fd_execute_rw+0xc5/0x2a0 [target_core_file]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04696f2>]
>> sbc_execute_rw+0x22/0x30 [target_core_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc04681ef>]
>> __target_execute_cmd+0x1f/0x70 [target_core_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc0468da5>]
>> target_execute_cmd+0x195/0x2a0 [target_core_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05db89a>]
>> iscsit_execute_cmd+0x20a/0x270 [iscsi_target_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05e4aea>]
>> iscsit_sequence_cmd+0xda/0x190 [iscsi_target_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eafbd>]
>> iscsi_target_rx_thread+0x51d/0xe30 [iscsi_target_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8101566c>] ? __switch_to+0x1dc/0x5a0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffffc05eaaa0>] ?
>> iscsi_target_tx_thread+0x1e0/0x1e0 [iscsi_target_mod]
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c8b8>] kthread+0xd8/0xf0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
>> kthread_create_on_node+0x1a0/0x1a0
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff817fc58f>] ret_from_fork+0x3f/0x70
>> [Tue Dec 15 00:46:55 2015]  [<ffffffff8109c7e0>] ?
>> kthread_create_on_node+0x1a0/0x1a0
>> [Tue Dec 15 00:46:55 2015] ---[ end trace 4079437668c77cbb ]---
>> [Tue Dec 15 00:47:45 2015] ABORT_TASK: Found referenced iSCSI task_tag: 95784927
>> [Tue Dec 15 00:47:45 2015] ABORT_TASK: ref_tag: 95784927 already
>> complete, skipping
>>
>> If it is a Ceph File System issue, let me know and I will open a bug.
>>
>> Thanks
>>
>> Eric
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-18  7:18                     ` Yan, Zheng
  2015-12-18 15:53                       ` Eric Eastman
@ 2015-12-21  2:38                       ` Eric Eastman
  2015-12-21 14:51                         ` Gregory Farnum
  2015-12-23  5:26                         ` Eric Eastman
  1 sibling, 2 replies; 20+ messages in thread
From: Eric Eastman @ 2015-12-21  2:38 UTC (permalink / raw)
  To: Yan, Zheng; +Cc: Chao Yu, Ceph Development

On Fri, Dec 18, 2015 at 12:18 AM, Yan, Zheng <ukernel@gmail.com> wrote:
> On Fri, Dec 18, 2015 at 2:23 PM, Eric Eastman
> <eric.eastman@keepertech.com> wrote:
>>> Hi Yan Zheng, Eric Eastman
>>>
>>> Similar bug was reported in f2fs, btrfs, it does affect 4.4-rc4, the fixing
>>> patch was merged into 4.4-rc5, dfd01f026058 ("sched/wait: Fix the signal
>>> handling fix").
>>>
>>> Related report & discussion was here:
>>> https://lkml.org/lkml/2015/12/12/149
>>>
>>> I'm not sure the current reported issue of ceph was related to that though,
>>> but at least try testing with an upgraded or patched kernel could verify it.
>>> :)
>>>
>>> Thanks,

>
> please try rc5 kernel without patches and DEBUG_VM=y
>
> Regards
> Yan, Zheng


The latest test with 4.4rc5 with CONFIG_DEBUG_VM=y has ran for over 36
hours with no ERRORS or WARNINGS.  My plan is to install the 4.4rc6
kernel from the Ubuntu kernel-ppa site once it is available, and rerun
the tests.

Before running this test I had to rebuild the Ceph File System as
after the last logged errors on Friday using the 4.4rc4 kernel, the
Ceph File system hung accessing the exported image file.  After
rebooting my iSCSI gateway using the Ceph File System, from / using
command: strace du -a cephfs, the mount point, the hang happened on
the newfsstatat call on my image file:

write(1, "0\tcephfs/ctdb/.ctdb.lock\n", 250 cephfs/ctdb/.ctdb.lock
) = 25
close(5)                                = 0
write(1, "0\tcephfs/ctdb\n", 140 cephfs/ctdb
)        = 14
newfstatat(4, "iscsi", {st_mode=S_IFDIR|0755, st_size=993814480896,
...}, AT_SYMLINK_NOFOLLOW) = 0
openat(4, "iscsi", O_RDONLY|O_NOCTTY|O_NONBLOCK|O_DIRECTORY|O_NOFOLLOW) = 3
fcntl(3, F_GETFD)                       = 0
fcntl(3, F_SETFD, FD_CLOEXEC)           = 0
fstat(3, {st_mode=S_IFDIR|0755, st_size=993814480896, ...}) = 0
fcntl(3, F_GETFL)                       = 0x38800 (flags
O_RDONLY|O_NONBLOCK|O_LARGEFILE|O_DIRECTORY|O_NOFOLLOW)
fcntl(3, F_SETFD, FD_CLOEXEC)           = 0
newfstatat(4, "iscsi", {st_mode=S_IFDIR|0755, st_size=993814480896,
...}, AT_SYMLINK_NOFOLLOW) = 0
fcntl(3, F_DUPFD, 3)                    = 5
fcntl(5, F_GETFD)                       = 0
fcntl(5, F_SETFD, FD_CLOEXEC)           = 0
getdents(3, /* 8 entries */, 65536)     = 288
getdents(3, /* 0 entries */, 65536)     = 0
close(3)                                = 0
newfstatat(5, "iscsi900g.img", ^C
^C^C^C
^Z
I could not break out with a ^C, and had to background the process to
get my prompt back. The process would not die so I had to hard reset
the system.

This same hang happened on 2 other kernel mounted systems using a 4.3.0 kernel.

On a separate system, I fuse mounted the file system and a du -a
cephfs hung at the same point. Once again I could not break out of the
hang, and had to hard reset the system.

Restarting the MDS and Monitors did not clear the issue. Taking a
quick look at the dumpcache showed it was large

# ceph mds tell 0 dumpcache /tmp/dump.txt
ok
# wc /tmp/dump.txt
  370556  5002449 59211054 /tmp/dump.txt
# tail /tmp/dump.txt
[inode 10000259276 [...c4,head] ~mds0/stray0/10000259276/ auth v977593
snaprealm=0x561339e3fb00 f(v0 m2015-12-12 00:51:04.345614) n(v0
rc2015-12-12 00:51:04.345614 1=0+1) (iversion lock) 0x561339c66228]
[inode 1000020c1ba [...a6,head] ~mds0/stray0/1000020c1ba/ auth v742016
snaprealm=0x56133ad19600 f(v0 m2015-12-10 18:25:55.880167) n(v0
rc2015-12-10 18:25:55.880167 1=0+1) (iversion lock) 0x56133a5e0d88]
[inode 100000d0088 [...77,head] ~mds0/stray6/100000d0088/ auth v292336
snaprealm=0x5613537673c0 f(v0 m2015-12-08 19:23:20.269283) n(v0
rc2015-12-08 19:23:20.269283 1=0+1) (iversion lock) 0x56134c2f7378]

I tried one more thing:

ceph daemon mds.0 flush journal

and restarted the MDS. Accessing the file system still locked up, but
a du -a cephfs did not even get to the iscsi900g.img file. As I was
running on a broken rc kernel, with snapshots turned on, when this
corruption happened, I decided to recreated the file system and
restarted the ESXi iSCSI test.

Regards,
Eric

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-21  2:38                       ` Eric Eastman
@ 2015-12-21 14:51                         ` Gregory Farnum
  2015-12-23  5:26                         ` Eric Eastman
  1 sibling, 0 replies; 20+ messages in thread
From: Gregory Farnum @ 2015-12-21 14:51 UTC (permalink / raw)
  To: Eric Eastman; +Cc: Yan, Zheng, Chao Yu, Ceph Development

On Sun, Dec 20, 2015 at 6:38 PM, Eric Eastman
<eric.eastman@keepertech.com> wrote:
> On Fri, Dec 18, 2015 at 12:18 AM, Yan, Zheng <ukernel@gmail.com> wrote:
>> On Fri, Dec 18, 2015 at 2:23 PM, Eric Eastman
>> <eric.eastman@keepertech.com> wrote:
>>>> Hi Yan Zheng, Eric Eastman
>>>>
>>>> Similar bug was reported in f2fs, btrfs, it does affect 4.4-rc4, the fixing
>>>> patch was merged into 4.4-rc5, dfd01f026058 ("sched/wait: Fix the signal
>>>> handling fix").
>>>>
>>>> Related report & discussion was here:
>>>> https://lkml.org/lkml/2015/12/12/149
>>>>
>>>> I'm not sure the current reported issue of ceph was related to that though,
>>>> but at least try testing with an upgraded or patched kernel could verify it.
>>>> :)
>>>>
>>>> Thanks,
>
>>
>> please try rc5 kernel without patches and DEBUG_VM=y
>>
>> Regards
>> Yan, Zheng
>
>
> The latest test with 4.4rc5 with CONFIG_DEBUG_VM=y has ran for over 36
> hours with no ERRORS or WARNINGS.  My plan is to install the 4.4rc6
> kernel from the Ubuntu kernel-ppa site once it is available, and rerun
> the tests.
>
> Before running this test I had to rebuild the Ceph File System as
> after the last logged errors on Friday using the 4.4rc4 kernel, the
> Ceph File system hung accessing the exported image file.  After
> rebooting my iSCSI gateway using the Ceph File System, from / using
> command: strace du -a cephfs, the mount point, the hang happened on
> the newfsstatat call on my image file:
>
> write(1, "0\tcephfs/ctdb/.ctdb.lock\n", 250 cephfs/ctdb/.ctdb.lock
> ) = 25
> close(5)                                = 0
> write(1, "0\tcephfs/ctdb\n", 140 cephfs/ctdb
> )        = 14
> newfstatat(4, "iscsi", {st_mode=S_IFDIR|0755, st_size=993814480896,
> ...}, AT_SYMLINK_NOFOLLOW) = 0
> openat(4, "iscsi", O_RDONLY|O_NOCTTY|O_NONBLOCK|O_DIRECTORY|O_NOFOLLOW) = 3
> fcntl(3, F_GETFD)                       = 0
> fcntl(3, F_SETFD, FD_CLOEXEC)           = 0
> fstat(3, {st_mode=S_IFDIR|0755, st_size=993814480896, ...}) = 0
> fcntl(3, F_GETFL)                       = 0x38800 (flags
> O_RDONLY|O_NONBLOCK|O_LARGEFILE|O_DIRECTORY|O_NOFOLLOW)
> fcntl(3, F_SETFD, FD_CLOEXEC)           = 0
> newfstatat(4, "iscsi", {st_mode=S_IFDIR|0755, st_size=993814480896,
> ...}, AT_SYMLINK_NOFOLLOW) = 0
> fcntl(3, F_DUPFD, 3)                    = 5
> fcntl(5, F_GETFD)                       = 0
> fcntl(5, F_SETFD, FD_CLOEXEC)           = 0
> getdents(3, /* 8 entries */, 65536)     = 288
> getdents(3, /* 0 entries */, 65536)     = 0
> close(3)                                = 0
> newfstatat(5, "iscsi900g.img", ^C
> ^C^C^C
> ^Z
> I could not break out with a ^C, and had to background the process to
> get my prompt back. The process would not die so I had to hard reset
> the system.
>
> This same hang happened on 2 other kernel mounted systems using a 4.3.0 kernel.
>
> On a separate system, I fuse mounted the file system and a du -a
> cephfs hung at the same point. Once again I could not break out of the
> hang, and had to hard reset the system.
>
> Restarting the MDS and Monitors did not clear the issue. Taking a
> quick look at the dumpcache showed it was large
>
> # ceph mds tell 0 dumpcache /tmp/dump.txt
> ok
> # wc /tmp/dump.txt
>   370556  5002449 59211054 /tmp/dump.txt
> # tail /tmp/dump.txt
> [inode 10000259276 [...c4,head] ~mds0/stray0/10000259276/ auth v977593
> snaprealm=0x561339e3fb00 f(v0 m2015-12-12 00:51:04.345614) n(v0
> rc2015-12-12 00:51:04.345614 1=0+1) (iversion lock) 0x561339c66228]
> [inode 1000020c1ba [...a6,head] ~mds0/stray0/1000020c1ba/ auth v742016
> snaprealm=0x56133ad19600 f(v0 m2015-12-10 18:25:55.880167) n(v0
> rc2015-12-10 18:25:55.880167 1=0+1) (iversion lock) 0x56133a5e0d88]
> [inode 100000d0088 [...77,head] ~mds0/stray6/100000d0088/ auth v292336
> snaprealm=0x5613537673c0 f(v0 m2015-12-08 19:23:20.269283) n(v0
> rc2015-12-08 19:23:20.269283 1=0+1) (iversion lock) 0x56134c2f7378]

These are deleted files that haven't been trimmed yet...

>
> I tried one more thing:
>
> ceph daemon mds.0 flush journal
>
> and restarted the MDS. Accessing the file system still locked up, but
> a du -a cephfs did not even get to the iscsi900g.img file. As I was
> running on a broken rc kernel, with snapshots turned on

...and I think we have some known issues in the tracker about snap
trimming and snapshotted inodes. So this is not entirely surprising.
:/
-Greg


>, when this
> corruption happened, I decided to recreated the file system and
> restarted the ESXi iSCSI test.
>
> Regards,
> Eric
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 20+ messages in thread

* Re: Issue with Ceph File System and LIO
  2015-12-21  2:38                       ` Eric Eastman
  2015-12-21 14:51                         ` Gregory Farnum
@ 2015-12-23  5:26                         ` Eric Eastman
  1 sibling, 0 replies; 20+ messages in thread
From: Eric Eastman @ 2015-12-23  5:26 UTC (permalink / raw)
  To: Yan, Zheng; +Cc: Chao Yu, Ceph Development

On Sun, Dec 20, 2015 at 7:38 PM, Eric Eastman
<eric.eastman@keepertech.com> wrote:
> On Fri, Dec 18, 2015 at 12:18 AM, Yan, Zheng <ukernel@gmail.com> wrote:
>> On Fri, Dec 18, 2015 at 2:23 PM, Eric Eastman
>> <eric.eastman@keepertech.com> wrote:
>>>> Hi Yan Zheng, Eric Eastman
>>>>
>>>> Similar bug was reported in f2fs, btrfs, it does affect 4.4-rc4, the fixing
>>>> patch was merged into 4.4-rc5, dfd01f026058 ("sched/wait: Fix the signal
>>>> handling fix").
>>>>
>>>> Related report & discussion was here:
>>>> https://lkml.org/lkml/2015/12/12/149
>>>>
>>>> I'm not sure the current reported issue of ceph was related to that though,
>>>> but at least try testing with an upgraded or patched kernel could verify it.
>>>> :)
>>>>
>>>> Thanks,
>
>>
>> please try rc5 kernel without patches and DEBUG_VM=y
>>
>> Regards
>> Yan, Zheng
>
>
> The latest test with 4.4rc5 with CONFIG_DEBUG_VM=y has ran for over 36
> hours with no ERRORS or WARNINGS.  My plan is to install the 4.4rc6
> kernel from the Ubuntu kernel-ppa site once it is available, and rerun
> the tests.
>

Test has run for 2 days using the 4.4rc6 kernel from the Ubuntu
kernel-ppa kernel site without error or warning.  Looks like it was a
4.4rc4 bug.

^ permalink raw reply	[flat|nested] 20+ messages in thread

end of thread, other threads:[~2015-12-23  5:26 UTC | newest]

Thread overview: 20+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2015-12-15  6:08 Issue with Ceph File System and LIO Eric Eastman
2015-12-15  9:26 ` Mike Christie
2015-12-15 10:12   ` John Spray
2015-12-15 12:05 ` Yan, Zheng
2015-12-15 16:51   ` Eric Eastman
2015-12-16 14:29     ` Yan, Zheng
2015-12-16 15:15       ` Eric Eastman
2015-12-17  8:56         ` Eric Eastman
2015-12-17  9:38           ` Yan, Zheng
2015-12-17 19:49             ` Eric Eastman
2015-12-18  4:05               ` Yan, Zheng
     [not found]                 ` <00b601d13954$5d4a94e0$17dfbea0$@samsung.com>
2015-12-18  6:23                   ` Eric Eastman
2015-12-18  7:18                     ` Yan, Zheng
2015-12-18 15:53                       ` Eric Eastman
2015-12-21  2:38                       ` Eric Eastman
2015-12-21 14:51                         ` Gregory Farnum
2015-12-23  5:26                         ` Eric Eastman
2015-12-17 16:14           ` Minfei Huang
2015-12-18 21:54 ` Mike Christie
2015-12-19  2:13   ` Eric Eastman

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.