All of lore.kernel.org
 help / color / mirror / Atom feed
* [Bug 214147] New: ISCSI broken in last release
@ 2021-08-23 16:08 bugzilla-daemon
  2021-08-23 19:19 ` [Bug 214147] " bugzilla-daemon
                   ` (5 more replies)
  0 siblings, 6 replies; 7+ messages in thread
From: bugzilla-daemon @ 2021-08-23 16:08 UTC (permalink / raw)
  To: linux-scsi

https://bugzilla.kernel.org/show_bug.cgi?id=214147

            Bug ID: 214147
           Summary: ISCSI broken in last release
           Product: IO/Storage
           Version: 2.5
    Kernel Version: 5.13.12
          Hardware: All
                OS: Linux
              Tree: Mainline
            Status: NEW
          Severity: normal
          Priority: P1
         Component: SCSI
          Assignee: linux-scsi@vger.kernel.org
          Reporter: slavon.net@gmail.com
        Regression: Yes

Created attachment 298441
  --> https://bugzilla.kernel.org/attachment.cgi?id=298441&action=edit
dmesg log

After some time iscsi go to broke and help only reboot

-- 
You may reply to this email to add a comment.

You are receiving this mail because:
You are the assignee for the bug.

^ permalink raw reply	[flat|nested] 7+ messages in thread

* [Bug 214147] ISCSI broken in last release
  2021-08-23 16:08 [Bug 214147] New: ISCSI broken in last release bugzilla-daemon
@ 2021-08-23 19:19 ` bugzilla-daemon
  2021-09-01 23:48 ` [Bug 214147] New: " michael.christie
                   ` (4 subsequent siblings)
  5 siblings, 0 replies; 7+ messages in thread
From: bugzilla-daemon @ 2021-08-23 19:19 UTC (permalink / raw)
  To: linux-scsi

https://bugzilla.kernel.org/show_bug.cgi?id=214147

--- Comment #1 from Badalian Slava (slavon.net@gmail.com) ---
5.13.4-1.el8.x86_64 also have this bug.
5.12 work ok

-- 
You may reply to this email to add a comment.

You are receiving this mail because:
You are the assignee for the bug.

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [Bug 214147] New: ISCSI broken in last release
  2021-08-23 16:08 [Bug 214147] New: ISCSI broken in last release bugzilla-daemon
  2021-08-23 19:19 ` [Bug 214147] " bugzilla-daemon
@ 2021-09-01 23:48 ` michael.christie
  2021-09-01 23:48 ` [Bug 214147] " bugzilla-daemon
                   ` (3 subsequent siblings)
  5 siblings, 0 replies; 7+ messages in thread
From: michael.christie @ 2021-09-01 23:48 UTC (permalink / raw)
  To: bugzilla-daemon, linux-scsi

On 8/23/21 6:08 AM, bugzilla-daemon@bugzilla.kernel.org wrote:
> https://bugzilla.kernel.org/show_bug.cgi?id=214147
> 
>             Bug ID: 214147
>            Summary: ISCSI broken in last release
>            Product: IO/Storage
>            Version: 2.5
>     Kernel Version: 5.13.12
>           Hardware: All
>                 OS: Linux
>               Tree: Mainline
>             Status: NEW
>           Severity: normal
>           Priority: P1
>          Component: SCSI
>           Assignee: linux-scsi@vger.kernel.org
>           Reporter: slavon.net@gmail.com
>         Regression: Yes
> 
> Created attachment 298441
>   --> https://bugzilla.kernel.org/attachment.cgi?id=298441&action=edit
> dmesg log
> 
> After some time iscsi go to broke and help only reboot
> 
What are you doing when you hit the issue?

What does your target setup look like? What are you using for the
backing store?

Are you able to build your own kernels?

The only major changes between 5.12 and 5.13 is some target patches
to batch cmds. However, it looks like you start to hit a problem
earlier than when that code comes into play. We first see you hit
a data out timeout, so we don't even have all the data for the
cmd, so the target changes in 5.13 don't come into play yet.

[10931.107057] Unable to recover from DataOut timeout while in ERL=0, closing iSCSI connection for I_T Nexus iqn.1991-05.com.microsoft:vhost11.dev.obs.group,i,0x400001370002,iqn.2003-01.org.linux-iscsi.vm2.x8664:sn.b07943625401,t,0x01


However, we do see some cmds have made it to the core target layer
because we can see the target layer is waiting on cmds to complete
for part of the lun reset handling:

[19906.593285] INFO: task kworker/4:1:3770999 blocked for more than 122 seconds.
[19906.603670]       Tainted: P           O      5.13.12-1.el8.elrepo.x86_64 #1
[19906.613975] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[19906.624208] task:kworker/4:1     state:D stack:    0 pid:3770999 ppid:     2 flags:0x00004000
[19906.624212] Workqueue: events target_tmr_work [target_core_mod]
[19906.624247] Call Trace:
[19906.624249]  __schedule+0x396/0x8a0
[19906.624252]  schedule+0x3c/0xa0
[19906.624255]  schedule_timeout+0x215/0x2b0
[19906.624258]  ? kasprintf+0x4e/0x70
[19906.624261]  wait_for_completion+0x9e/0x100
[19906.624264]  target_put_cmd_and_wait+0x55/0x80 [target_core_mod]
[19906.624279]  core_tmr_lun_reset+0x38b/0x660 [target_core_mod]
[19906.624294]  target_tmr_work+0xb4/0x110 [target_core_mod]
[19906.624309]  process_one_work+0x230/0x3d0
[19906.624312]  worker_thread+0x2d/0x3e0
[19906.624314]  ? process_one_work+0x3d0/0x3d0
[19906.624316]  kthread+0x118/0x140
[19906.624318]  ? set_kthread_struct+0x40/0x40
[19906.624320]  ret_from_fork+0x1f/0x30

and we can see iscsi layer not able to relogin because of outstanding
cmds/tmfs.

I can send you a patch that reverts the core target patches. If we can
rule them out then it would help narrow things down.

Or, because it sounds like this is easy to reproduce we can turn on some
extra lio debugging.

^ permalink raw reply	[flat|nested] 7+ messages in thread

* [Bug 214147] ISCSI broken in last release
  2021-08-23 16:08 [Bug 214147] New: ISCSI broken in last release bugzilla-daemon
  2021-08-23 19:19 ` [Bug 214147] " bugzilla-daemon
  2021-09-01 23:48 ` [Bug 214147] New: " michael.christie
@ 2021-09-01 23:48 ` bugzilla-daemon
  2021-09-03  6:22 ` bugzilla-daemon
                   ` (2 subsequent siblings)
  5 siblings, 0 replies; 7+ messages in thread
From: bugzilla-daemon @ 2021-09-01 23:48 UTC (permalink / raw)
  To: linux-scsi

https://bugzilla.kernel.org/show_bug.cgi?id=214147

--- Comment #2 from michael.christie@oracle.com ---
On 8/23/21 6:08 AM, bugzilla-daemon@bugzilla.kernel.org wrote:
> https://bugzilla.kernel.org/show_bug.cgi?id=214147
> 
>             Bug ID: 214147
>            Summary: ISCSI broken in last release
>            Product: IO/Storage
>            Version: 2.5
>     Kernel Version: 5.13.12
>           Hardware: All
>                 OS: Linux
>               Tree: Mainline
>             Status: NEW
>           Severity: normal
>           Priority: P1
>          Component: SCSI
>           Assignee: linux-scsi@vger.kernel.org
>           Reporter: slavon.net@gmail.com
>         Regression: Yes
> 
> Created attachment 298441
>   --> https://bugzilla.kernel.org/attachment.cgi?id=298441&action=edit
> dmesg log
> 
> After some time iscsi go to broke and help only reboot
> 
What are you doing when you hit the issue?

What does your target setup look like? What are you using for the
backing store?

Are you able to build your own kernels?

The only major changes between 5.12 and 5.13 is some target patches
to batch cmds. However, it looks like you start to hit a problem
earlier than when that code comes into play. We first see you hit
a data out timeout, so we don't even have all the data for the
cmd, so the target changes in 5.13 don't come into play yet.

[10931.107057] Unable to recover from DataOut timeout while in ERL=0, closing
iSCSI connection for I_T Nexus
iqn.1991-05.com.microsoft:vhost11.dev.obs.group,i,0x400001370002,iqn.2003-01.org.linux-iscsi.vm2.x8664:sn.b07943625401,t,0x01


However, we do see some cmds have made it to the core target layer
because we can see the target layer is waiting on cmds to complete
for part of the lun reset handling:

[19906.593285] INFO: task kworker/4:1:3770999 blocked for more than 122
seconds.
[19906.603670]       Tainted: P           O      5.13.12-1.el8.elrepo.x86_64 #1
[19906.613975] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this
message.
[19906.624208] task:kworker/4:1     state:D stack:    0 pid:3770999 ppid:     2
flags:0x00004000
[19906.624212] Workqueue: events target_tmr_work [target_core_mod]
[19906.624247] Call Trace:
[19906.624249]  __schedule+0x396/0x8a0
[19906.624252]  schedule+0x3c/0xa0
[19906.624255]  schedule_timeout+0x215/0x2b0
[19906.624258]  ? kasprintf+0x4e/0x70
[19906.624261]  wait_for_completion+0x9e/0x100
[19906.624264]  target_put_cmd_and_wait+0x55/0x80 [target_core_mod]
[19906.624279]  core_tmr_lun_reset+0x38b/0x660 [target_core_mod]
[19906.624294]  target_tmr_work+0xb4/0x110 [target_core_mod]
[19906.624309]  process_one_work+0x230/0x3d0
[19906.624312]  worker_thread+0x2d/0x3e0
[19906.624314]  ? process_one_work+0x3d0/0x3d0
[19906.624316]  kthread+0x118/0x140
[19906.624318]  ? set_kthread_struct+0x40/0x40
[19906.624320]  ret_from_fork+0x1f/0x30

and we can see iscsi layer not able to relogin because of outstanding
cmds/tmfs.

I can send you a patch that reverts the core target patches. If we can
rule them out then it would help narrow things down.

Or, because it sounds like this is easy to reproduce we can turn on some
extra lio debugging.

-- 
You may reply to this email to add a comment.

You are receiving this mail because:
You are the assignee for the bug.

^ permalink raw reply	[flat|nested] 7+ messages in thread

* [Bug 214147] ISCSI broken in last release
  2021-08-23 16:08 [Bug 214147] New: ISCSI broken in last release bugzilla-daemon
                   ` (2 preceding siblings ...)
  2021-09-01 23:48 ` [Bug 214147] " bugzilla-daemon
@ 2021-09-03  6:22 ` bugzilla-daemon
  2021-09-16  5:06 ` bugzilla-daemon
  2021-09-16  9:04 ` bugzilla-daemon
  5 siblings, 0 replies; 7+ messages in thread
From: bugzilla-daemon @ 2021-09-03  6:22 UTC (permalink / raw)
  To: linux-scsi

https://bugzilla.kernel.org/show_bug.cgi?id=214147

--- Comment #3 from Badalian Slava (slavon.net@gmail.com) ---
Comment on attachment 298441
  --> https://bugzilla.kernel.org/attachment.cgi?id=298441
dmesg log

Ok. Send me patch, i can test it.



we use ZFS (last git(. Have 2 pools - DRAID from 6 HDD amd stripe for 10 SSD
850 EVI/

Windows Cluser connect ZFS blocked devices as Shared clustered disk on iSCSI in
10GB mellonox channel.  (Noo RDMA support in windows side in iSCSI). 

Linux connect in NFS over RDMA on infiniband 40 g network.

We wont recreate SSD pool and move all VMs to HDD. Thern try move next VM. In
Windows host start migrate HV disk from, one clustered disk to second.. 10-15
min and iSCSI broke.

5.12 reconnect after some time. 5.13 system is zrease. Linux can't reboot. Only
ipmi power reset.



SATA is LSI3008. 

[root@vm2 ~]# lsscsi
[0:0:0:0]    disk    ATA      Samsung SSD 850  3B6Q  /dev/sdb
[0:0:1:0]    disk    ATA      Samsung SSD 850  3B6Q  /dev/sdc
[0:0:2:0]    disk    ATA      Samsung SSD 850  3B6Q  /dev/sdd
[0:0:3:0]    disk    ATA      Samsung SSD 850  3B6Q  /dev/sde
[0:0:4:0]    disk    ATA      Samsung SSD 850  2B6Q  /dev/sdf
[0:0:5:0]    disk    ATA      Samsung SSD 850  2B6Q  /dev/sdg
[0:0:6:0]    disk    ATA      Samsung SSD 850  3B6Q  /dev/sds
[0:0:7:0]    disk    ATA      Samsung SSD 850  3B6Q  /dev/sdt
[4:0:0:0]    disk    ATA      KINGSTON SV300S3 BBF0  /dev/sda
[9:0:0:0]    disk    ATA      Samsung SSD 850  3B6Q  /dev/sdh
[9:0:1:0]    disk    ATA      Samsung SSD 850  3B6Q  /dev/sdi
[9:0:2:0]    disk    ATA      ST2000DM001-1ER1 CC26  /dev/sdj
[9:0:3:0]    disk    ATA      ST2000DM001-1ER1 CC26  /dev/sdk
[9:0:4:0]    disk    ATA      INTEL SSDSC2CW24 400i  /dev/sdl
[9:0:5:0]    disk    ATA      ST2000DM001-1CH1 CC29  /dev/sdm
[9:0:6:0]    disk    ATA      ST2000DM001-1CH1 CC29  /dev/sdn
[9:0:7:0]    disk    ATA      WDC WD20EZRZ-00Z 0A80  /dev/sdo
[9:0:8:0]    disk    ATA      INTEL SSDSC2CW24 400i  /dev/sdp
[9:0:9:0]    disk    ATA      ST3000DM001-1CH1 CC27  /dev/sdq
[9:0:10:0]   disk    ATA      ST3000DM001-1CH1 CC27  /dev/sdr
[9:0:11:0]   enclosu LSI      SAS2X28          0e12  -
[N:0:1:1]    disk    Samsung SSD 950 PRO 256GB__1               /dev/nvme0n1
[N:1:1:1]    disk    Samsung SSD 950 PRO 256GB__1               /dev/nvme1n1
[root@vm2 ~]#

[root@vm2 ~]# zfs pool -^C
[root@vm2 ~]# zpool list -v
NAME                                            SIZE  ALLOC   FREE  CKPOINT 
EXPANDSZ   FRAG    CAP  DEDUP    HEALTH  ALTROOT
s                                              15.4T  11.5T  3.86T        -    
    -    61%    74%  1.20x    ONLINE  -
  sdf                                          3.62T  2.08T  1.54T        -    
    -    40%  57.5%      -    ONLINE
  sdg                                          3.62T  2.07T  1.56T        -    
    -    41%  57.0%      -    ONLINE
  ata-Samsung_SSD_850_EVO_1TB_S3NZNF0JC02715K   928G   812G   116G        -    
    -    77%  87.5%      -    ONLINE
  ata-Samsung_SSD_850_EVO_1TB_S3NZNF0JC02818K   928G   819G   109G        -    
    -    77%  88.3%      -    ONLINE
  ata-Samsung_SSD_850_EVO_1TB_S3NZNF0JC02714Y   928G   815G   113G        -    
    -    77%  87.8%      -    ONLINE
  ata-Samsung_SSD_850_EVO_1TB_S3NZNF0JC02708N   928G   827G   101G        -    
    -    79%  89.2%      -    ONLINE
  ata-Samsung_SSD_850_EVO_1TB_S3NZNF0JC02720L   928G   821G   107G        -    
    -    72%  88.4%      -    ONLINE
  ata-Samsung_SSD_850_EVO_1TB_S3NZNF0JC02717V   928G   834G  93.8G        -    
    -    76%  89.9%      -    ONLINE
  ata-Samsung_SSD_850_EVO_1TB_S3NZNF0JC02707F   928G   870G  58.1G        -    
    -    84%  93.7%      -    ONLINE
  ata-Samsung_SSD_850_EVO_1TB_S3NZNF0JC02719H   928G   907G  21.2G        -    
    -    94%  97.7%      -    ONLINE
  nvme1n1                                       238G   211G  27.1G        -    
    -    87%  88.6%      -    ONLINE
  nvme0n1                                       238G   210G  27.8G        -    
    -    87%  88.3%      -    ONLINE
  ata-INTEL_SSDSC2CW240A3_CVCV316506PA240FGN    222G   219G  2.55G        -    
    -    87%  98.8%      -    ONLINE
  ata-INTEL_SSDSC2CW240A3_CVCV316500UB240FGN    222G   220G  2.46G        -    
    -    87%  98.9%      -    ONLINE
z                                              10.9T  4.22T  6.68T        -    
    -    28%    38%  1.11x    ONLINE  -
  draid2:4d:7c:1s                              10.9T  4.22T  6.68T        -    
    -    28%  38.7%      -    ONLINE
    ata-ST2000DM001-1ER164_Z5606648                -      -      -        -    
    -      -      -      -    ONLINE
    ata-ST2000DM001-1ER164_Z4Z5X09J                -      -      -        -    
    -      -      -      -    ONLINE
    ata-ST2000DM001-1CH164_Z340KWSB                -      -      -        -    
    -      -      -      -    ONLINE
    ata-ST2000DM001-1CH164_Z1E668SP                -      -      -        -    
    -      -      -      -    ONLINE
    ata-ST3000DM001-1CH166_Z1F3WA6S                -      -      -        -    
    -      -      -      -    ONLINE
    ata-ST3000DM001-1CH166_Z1F3RQCR                -      -      -        -    
    -      -      -      -    ONLINE
    ata-WDC_WD20EZRZ-00Z5HB0_WD-WCC4N2KHEF9K       -      -      -        -    
    -      -      -      -    ONLINE
spare                                              -      -      -        -    
    -      -      -      -  -
  draid2-0-0                                       -      -      -        -    
    -      -      -      -     AVAIL

-- 
You may reply to this email to add a comment.

You are receiving this mail because:
You are the assignee for the bug.

^ permalink raw reply	[flat|nested] 7+ messages in thread

* [Bug 214147] ISCSI broken in last release
  2021-08-23 16:08 [Bug 214147] New: ISCSI broken in last release bugzilla-daemon
                   ` (3 preceding siblings ...)
  2021-09-03  6:22 ` bugzilla-daemon
@ 2021-09-16  5:06 ` bugzilla-daemon
  2021-09-16  9:04 ` bugzilla-daemon
  5 siblings, 0 replies; 7+ messages in thread
From: bugzilla-daemon @ 2021-09-16  5:06 UTC (permalink / raw)
  To: linux-scsi

https://bugzilla.kernel.org/show_bug.cgi?id=214147

--- Comment #4 from Badalian Slava (slavon.net@gmail.com) ---
I get some info .


Test case
have 2 zfs pool. HDD is zdev (raid6) SDD is stype.
simple copy from SSD to HDD

Specific^

1. I use NFS with RDMA and 16 threads (for linux hosts)
2/ i use targetcli and nvmet for share devs to WinClusteredHosts


When se get deadlock i see in perf

amples: 1M of event 'cycles', Event count (approx.): 711199023386
  Children      Self  Command          Shared Object             Symbol
-   30,57%     0,00%  z_wr_iss         [kernel.kallsyms]         [k]
ret_from_fork                                                                  
                                        ◆
     ret_from_fork                                                             
                                                                               
                             ▒
   - kthread                                                                   
                                                                               
                             ▒
      - 30,53% taskq_create                                                    
                                                                               
                             ▒
         - 16,50% zio_inject_init                                              
                                                                               
                             ▒
            - 4,77% 0xffffffffc0c7285f                                         
                                                                               
                             ▒
               - 4,73% raidz_rec_pr_coeff.isra.5                               
                                                                               
                             ▒
                  - 3,86% secpolicy_vnode_setid_retain                         
                                                                               
                             ▒
                     - 2,10% zthr_resume                                       
                                                                               
                             ▒
                        - 1,85% vdev_label_init                                
                                                                               
                             ▒
                           - 1,03% secpolicy_vnode_setid_retain                
                                                                               
                             ▒
                              - 0,73% zthr_cancel                              
                                                                               
                             ▒
                                   0,72% vdev_rebuild_update_sync              
                                                                               
                             ▒
                     - 1,57% zthr_cancel                                       
                                                                               
                             ▒
                        - 1,55% vdev_rebuild_update_sync                       
                                                                               
                             ▒
                             0,53% __mutex_lock.isra.13                        
                                                                               
                             ▒
                             0,53% avl_remove                                  
                                                                               
                             ▒
                    0,71% abd_bio_map_off                                      
                                                                               
                             ▒
            - 3,70% vdev_disk_open                                             
                                                                               
                             ▒
               - 3,69% range_tree_remove_impl                                  
                                                                               
                             ▒
                  - 3,45% range_tree_add_impl                                  
                                                                               
                             ▒
                     - 2,66% mmp_thread                                        
                                                                               
                             ▒
                          1,24% metaslab_enable                                
                                                                               
                             ▒
                          0,74% __mutex_lock.isra.13                           
                                                                               
                             ▒
            - 2,71% zvol_set_volsize                                           
                                                                               
                             ▒
               - 2,51% zfs_aclset_common                                       
                                                                               
                             ▒
                  - 2,51% metaslab_load.part.29                                
                                                                               
                             ▒
                       1,31% metaslab_class_evict_old                          
                                                                               
                             ▒
                       0,96% metaslab_unload.part.28                           
                                                                               
                             ▒
            - 1,84% zvol_set_volmode_sync_cb                                   
                                                                               
                             ▒
               - 1,83% zio_handle_device_injection_impl.part.15                
                                                                               
                             ▒
                  - 1,57% spl_taskq_fini                                       
                                                                               
                             ▒
                     - 1,55% _raw_spin_lock_irqsave                            
                                                                               
                             ▒
                          native_queued_spin_lock_slowpath                     
                                                                               
                             ▒
            - 1,20% zvol_set_volmode_sync                                      
                                                                               
                             ▒
                 1,02% l2arc_read_done                                         
                                                                               
                             ▒
            - 0,63% zio_inject_list_next                                       
                                                                               
                             ▒
               - 0,62% zfs_mode_compute                                        
                                                                               
                             ▒
                  - 0,59% zfs_mode_compute                                     
                                                                               
                             ▒
                       0,53% abd_iterate_func2.part.16                         
                                                                               
                             ▒
         - 12,00% _raw_spin_lock_irqsave                                       
                                                                               
                             ▒
              native_queued_spin_lock_slowpath                                 
                                                                               
                             ▒
         - 0,93% schedule                                                      
                                                                               
                             ▒
              0,86% __sched_text_start                                         
                                                                               
                             ▒
-   30,57%     0,00%  z_wr_iss         [kernel.kallsyms]         [k] kthread   
                                                                               
                             ▒
   - 30,57% kthread                                                            
                                                                               
                             ▒
      - 30,53% taskq_create                                                    
                                                                               
                             ▒
         - 16,50% zio_inject_init                                              
                                                                               
                             ▒
            - 4,77% 0xffffffffc0c7285f                                         
                                                                               
                             ▒
               - 4,73% raidz_rec_pr_coeff.isra.5                                


and  i see in 


[root@vm2 ~]# cat /proc/locks
1: FLOCK  ADVISORY  WRITE 999321 00:18:3499 0 EOF
2: POSIX  ADVISORY  WRITE 949090 00:18:3395 0 EOF
3: POSIX  ADVISORY  WRITE 949090 00:18:3394 0 EOF
4: POSIX  ADVISORY  WRITE 948911 00:18:3322 0 EOF
5: POSIX  ADVISORY  WRITE 999305 00:18:3511 0 EOF

butt one.one/two in ACTIVE state., pid is nfs5 kernel p[process. deadlocked  in
nfs!

-- 
You may reply to this email to add a comment.

You are receiving this mail because:
You are the assignee for the bug.

^ permalink raw reply	[flat|nested] 7+ messages in thread

* [Bug 214147] ISCSI broken in last release
  2021-08-23 16:08 [Bug 214147] New: ISCSI broken in last release bugzilla-daemon
                   ` (4 preceding siblings ...)
  2021-09-16  5:06 ` bugzilla-daemon
@ 2021-09-16  9:04 ` bugzilla-daemon
  5 siblings, 0 replies; 7+ messages in thread
From: bugzilla-daemon @ 2021-09-16  9:04 UTC (permalink / raw)
  To: linux-scsi

https://bugzilla.kernel.org/show_bug.cgi?id=214147

--- Comment #5 from Badalian Slava (slavon.net@gmail.com) ---
WWith turned off ISCSI and NVMEt. only NFS active

.
```


[ 2164.273231] perf: interrupt took too long (3197 > 3148), lowering
kernel.perf_event_max_sample_rate to 62000
[ 3067.386864] perf: interrupt took too long (4023 > 3996), lowering
kernel.perf_event_max_sample_rate to 49000
[ 4508.876951] perf: interrupt took too long (5034 > 5028), lowering
kernel.perf_event_max_sample_rate to 39000
[ 7336.586251] perf: interrupt took too long (6305 > 6292), lowering
kernel.perf_event_max_sample_rate to 31000
[16343.647108] INFO: task txg_sync:897001 blocked for more than 122 seconds.
[16343.657122]       Tainted: P           O      5.13.12-1.el8.elrepo.x86_64 #1
[16343.667018] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this
message.
[16343.676854] task:txg_sync        state:D stack:    0 pid:897001 ppid:     2
flags:0x00004004
[16343.686665] Call Trace:
[16343.696132]  __schedule+0x396/0x8a0
[16343.705378]  schedule+0x3c/0xa0
[16343.714360]  schedule_timeout+0x197/0x2b0
[16343.723125]  ? timer_update_keys+0x30/0x30
[16343.731752]  io_schedule_timeout+0x19/0x40
[16343.740453]  __cv_timedwait_common+0x12b/0x160 [spl]
[16343.749335]  ? wait_woken+0x80/0x80
[16343.758275]  __cv_timedwait_io+0x15/0x20 [spl]
[16343.767344]  zio_wait+0x129/0x2b0 [zfs]
[16343.776599]  dsl_pool_sync+0xd0/0x470 [zfs]
[16343.785629]  ? spa_errlog_sync+0x25c/0x2b0 [zfs]
[16343.794490]  ? __kmalloc_node+0x181/0x330
[16343.803219]  spa_sync_iterate_to_convergence+0xf0/0x250 [zfs]
[16343.811995]  spa_sync+0x389/0x690 [zfs]
[16343.820544]  txg_sync_thread+0x270/0x2f0 [zfs]
[16343.829106]  ? txg_dispatch_callbacks+0x120/0x120 [zfs]
[16343.837545]  ? __thread_exit+0x20/0x20 [spl]
[16343.845678]  thread_generic_wrapper+0x6c/0x80 [spl]
[16343.853833]  kthread+0x118/0x140
[16343.861743]  ? set_kthread_struct+0x40/0x40
[16343.869501]  ret_from_fork+0x1f/0x30
[16589.391679] INFO: task txg_sync:897001 blocked for more than 122 seconds.
[16589.399411]       Tainted: P           O      5.13.12-1.el8.elrepo.x86_64 #1
[16589.407078] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this
message.
[16589.414718] task:txg_sync        state:D stack:    0 pid:897001 ppid:     2
flags:0x00004004
[16589.422347] Call Trace:
[16589.429638]  __schedule+0x396/0x8a0
[16589.436751]  schedule+0x3c/0xa0
[16589.443608]  schedule_timeout+0x197/0x2b0
[16589.450258]  ? timer_update_keys+0x30/0x30
[16589.456706]  io_schedule_timeout+0x19/0x40
[16589.462926]  __cv_timedwait_common+0x12b/0x160 [spl]
[16589.469088]  ? wait_woken+0x80/0x80
[16589.475108]  __cv_timedwait_io+0x15/0x20 [spl]
[16589.481039]  zio_wait+0x129/0x2b0 [zfs]
[16589.486879]  dsl_pool_sync+0xd0/0x470 [zfs]
[16589.492585]  ? spa_errlog_sync+0x25c/0x2b0 [zfs]
[16589.498376]  ? __kmalloc_node+0x181/0x330
[16589.504065]  spa_sync_iterate_to_convergence+0xf0/0x250 [zfs]
[16589.510015]  spa_sync+0x389/0x690 [zfs]
[16589.515917]  txg_sync_thread+0x270/0x2f0 [zfs]
[16589.521867]  ? txg_dispatch_callbacks+0x120/0x120 [zfs]
[16589.527853]  ? __thread_exit+0x20/0x20 [spl]
[16589.533715]  thread_generic_wrapper+0x6c/0x80 [spl]
[16589.539629]  kthread+0x118/0x140
[16589.545490]  ? set_kthread_struct+0x40/0x40
[16589.551393]  ret_from_fork+0x1f/0x30
[16712.262822] INFO: task txg_sync:897001 blocked for more than 245 seconds.
[16712.268937]       Tainted: P           O      5.13.12-1.el8.elrepo.x86_64 #1
[16712.275093] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this
message.
[16712.281379] task:txg_sync        state:D stack:    0 pid:897001 ppid:     2
flags:0x00004004
[16712.287872] Call Trace:
[16712.294197]  __schedule+0x396/0x8a0
[16712.300540]  schedule+0x3c/0xa0
[16712.306795]  schedule_timeout+0x197/0x2b0
[16712.313005]  ? timer_update_keys+0x30/0x30
[16712.319211]  io_schedule_timeout+0x19/0x40
[16712.325401]  __cv_timedwait_common+0x12b/0x160 [spl]
[16712.331551]  ? wait_woken+0x80/0x80
[16712.337545]  __cv_timedwait_io+0x15/0x20 [spl]
[16712.343428]  zio_wait+0x129/0x2b0 [zfs]
[16712.349237]  dsl_pool_sync+0xd0/0x470 [zfs]
[16712.354918]  ? spa_errlog_sync+0x25c/0x2b0 [zfs]
[16712.360658]  ? __kmalloc_node+0x181/0x330
[16712.366315]  spa_sync_iterate_to_convergence+0xf0/0x250 [zfs]
[16712.372219]  spa_sync+0x389/0x690 [zfs]
[16712.378091]  txg_sync_thread+0x270/0x2f0 [zfs]
[16712.384027]  ? txg_dispatch_callbacks+0x120/0x120 [zfs]
[16712.389992]  ? __thread_exit+0x20/0x20 [spl]
[16712.395841]  thread_generic_wrapper+0x6c/0x80 [spl]
[16712.401736]  kthread+0x118/0x140
[16712.407599]  ? set_kthread_struct+0x40/0x40
[16712.413528]  ret_from_fork+0x1f/0x30
[17080.879563] INFO: task txg_sync:897001 blocked for more than 122 seconds.
[17080.885652]       Tainted: P           O      5.13.12-1.el8.elrepo.x86_64 #1
[17080.891804] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this
message.
[17080.898104] task:txg_sync        state:D stack:    0 pid:897001 ppid:     2
flags:0x00004004
[17080.904598] Call Trace:
[17080.910927]  __schedule+0x396/0x8a0
[17080.917254]  schedule+0x3c/0xa0
[17080.923501]  schedule_timeout+0x197/0x2b0
[17080.929722]  ? timer_update_keys+0x30/0x30
[17080.935962]  io_schedule_timeout+0x19/0x40
[17080.942162]  __cv_timedwait_common+0x12b/0x160 [spl]
[17080.948331]  ? wait_woken+0x80/0x80
[17080.954333]  __cv_timedwait_io+0x15/0x20 [spl]
[17080.960221]  zio_wait+0x129/0x2b0 [zfs]
[17080.966039]  dsl_pool_sync+0xd0/0x470 [zfs]
[17080.971729]  ? spa_errlog_sync+0x25c/0x2b0 [zfs]
[17080.977490]  ? __kmalloc_node+0x181/0x330
[17080.983150]  spa_sync_iterate_to_convergence+0xf0/0x250 [zfs]
[17080.989056]  spa_sync+0x389/0x690 [zfs]
[17080.994918]  txg_sync_thread+0x270/0x2f0 [zfs]
[17081.000838]  ? txg_dispatch_callbacks+0x120/0x120 [zfs]
[17081.006814]  ? __thread_exit+0x20/0x20 [spl]
[17081.012668]  thread_generic_wrapper+0x6c/0x80 [spl]
[17081.018579]  kthread+0x118/0x140
[17081.024437]  ? set_kthread_struct+0x40/0x40
[17081.030328]  ret_from_fork+0x1f/0x30
[17449.496395] INFO: task txg_sync:897001 blocked for more than 122 seconds.
[17449.502486]       Tainted: P           O      5.13.12-1.el8.elrepo.x86_64 #1
[17449.508658] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this
message.
[17449.514971] task:txg_sync        state:D stack:    0 pid:897001 ppid:     2
flags:0x00004004
[17449.521482] Call Trace:
[17449.527816]  __schedule+0x396/0x8a0
[17449.534139]  schedule+0x3c/0xa0
[17449.540394]  schedule_timeout+0x197/0x2b0
[17449.546614]  ? timer_update_keys+0x30/0x30
[17449.552862]  io_schedule_timeout+0x19/0x40
[17449.559082]  __cv_timedwait_common+0x12b/0x160 [spl]
[17449.565219]  ? wait_woken+0x80/0x80
[17449.571213]  __cv_timedwait_io+0x15/0x20 [spl]
[17449.577088]  zio_wait+0x129/0x2b0 [zfs]
[17449.582897]  dsl_pool_sync+0xd0/0x470 [zfs]
[17449.588566]  ? spa_errlog_sync+0x25c/0x2b0 [zfs]
[17449.594304]  ? __kmalloc_node+0x181/0x330
[17449.599966]  spa_sync_iterate_to_convergence+0xf0/0x250 [zfs]
[17449.605877]  spa_sync+0x389/0x690 [zfs]
[17449.611760]  txg_sync_thread+0x270/0x2f0 [zfs]
[17449.617729]  ? txg_dispatch_callbacks+0x120/0x120 [zfs]
[17449.623713]  ? __thread_exit+0x20/0x20 [spl]
[17449.629578]  thread_generic_wrapper+0x6c/0x80 [spl]
[17449.635512]  kthread+0x118/0x140
[17449.641394]  ? set_kthread_struct+0x40/0x40
[17449.647310]  ret_from_fork+0x1f/0x30
[17940.988317] INFO: task txg_sync:897001 blocked for more than 122 seconds.
[17940.994463]       Tainted: P           O      5.13.12-1.el8.elrepo.x86_64 #1
[17941.000626] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this
message.
[17941.006941] task:txg_sync        state:D stack:    0 pid:897001 ppid:     2
flags:0x00004004
[17941.013462] Call Trace:
[17941.019812]  __schedule+0x396/0x8a0
[17941.026154]  schedule+0x3c/0xa0
[17941.032430]  schedule_timeout+0x197/0x2b0
[17941.038665]  ? timer_update_keys+0x30/0x30
[17941.044893]  io_schedule_timeout+0x19/0x40
[17941.051104]  __cv_timedwait_common+0x12b/0x160 [spl]
[17941.057261]  ? wait_woken+0x80/0x80
[17941.063275]  __cv_timedwait_io+0x15/0x20 [spl]
[17941.069174]  zio_wait+0x129/0x2b0 [zfs]
[17941.075005]  dsl_pool_sync+0xd0/0x470 [zfs]
[17941.080690]  ? spa_errlog_sync+0x25c/0x2b0 [zfs]
[17941.086450]  ? __kmalloc_node+0x181/0x330
[17941.092137]  spa_sync_iterate_to_convergence+0xf0/0x250 [zfs]
[17941.098102]  spa_sync+0x389/0x690 [zfs]
[17941.103993]  txg_sync_thread+0x270/0x2f0 [zfs]
[17941.109957]  ? txg_dispatch_callbacks+0x120/0x120 [zfs]
[17941.115964]  ? __thread_exit+0x20/0x20 [spl]
[17941.121841]  thread_generic_wrapper+0x6c/0x80 [spl]
[17941.127779]  kthread+0x118/0x140
[17941.133647]  ? set_kthread_struct+0x40/0x40
[17941.139561]  ret_from_fork+0x1f/0x30
[root@vm2 parameters]# cat /proc/locks
1: POSIX  ADVISORY  READ 999553 00:31:55 201 201
2: POSIX  ADVISORY  READ 999553 00:31:55 100 101
3: DELEG  ACTIVE    READ 999553 00:31:189 0 EOF
4: FLOCK  ADVISORY  WRITE 999321 00:18:3499 0 EOF
5: POSIX  ADVISORY  WRITE 949090 00:18:3395 0 EOF
6: POSIX  ADVISORY  WRITE 949090 00:18:3394 0 EOF
7: POSIX  ADVISORY  WRITE 948911 00:18:3322 0 EOF
8: DELEG  ACTIVE    READ 999553 00:31:55 0 EOF
9: POSIX  ADVISORY  WRITE 999305 00:18:3511 0 EOF



```

-- 
You may reply to this email to add a comment.

You are receiving this mail because:
You are the assignee for the bug.

^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2021-09-16  9:04 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-08-23 16:08 [Bug 214147] New: ISCSI broken in last release bugzilla-daemon
2021-08-23 19:19 ` [Bug 214147] " bugzilla-daemon
2021-09-01 23:48 ` [Bug 214147] New: " michael.christie
2021-09-01 23:48 ` [Bug 214147] " bugzilla-daemon
2021-09-03  6:22 ` bugzilla-daemon
2021-09-16  5:06 ` bugzilla-daemon
2021-09-16  9:04 ` bugzilla-daemon

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.