* suddenly slow writes on XFS Filesystem @ 2012-05-06 9:01 Stefan Priebe 2012-05-06 10:31 ` Martin Steigerwald ` (2 more replies) 0 siblings, 3 replies; 33+ messages in thread From: Stefan Priebe @ 2012-05-06 9:01 UTC (permalink / raw) To: xfs Hi, since a few days i've experienced a really slow fs on one of our backup systems. I'm not sure whether this is XFS related or related to the Controller / Disks. It is a raid 10 of 20 SATA Disks and i can only write to them with about 700kb/s while doing random i/o. I tried vanilla Kernel 3.0.30 and 3.3.4 - no difference. Writing to another partition on another xfs array works fine. Details: #~ df -h /dev/sdb1 4,6T 4,4T 207G 96% /mnt #~ df -i /dev/sdb1 4875737052 4659318044 216419008 96% /mnt #~ xfs_db -c frag -r /dev/sdb1 actual 83160469, ideal 82145389, fragmentation factor 1,22% #~ xfs_info /dev/sdb1 meta-data=/dev/sdb1 isize=256 agcount=5, agsize=268435392 blks = sectsz=512 attr=2 data = bsize=4096 blocks=1218967031, imaxpct=5 = sunit=64 swidth=1280 blks naming =version 2 bsize=4096 ascii-ci=0 log =internal bsize=4096 blocks=32768, version=2 = sectsz=512 sunit=64 blks, lazy-count=0 realtime =none extsz=4096 blocks=0, rtextents=0 #~ cat /proc/mounts /dev/sdb1 /mnt xfs rw,noatime,nodiratime,attr2,nobarrier,logbufs=8,logbsize=256k,sunit=512,swidth=10240,prjquota 0 0 Any ideas? Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-06 9:01 suddenly slow writes on XFS Filesystem Stefan Priebe @ 2012-05-06 10:31 ` Martin Steigerwald 2012-05-06 10:33 ` Martin Steigerwald 2012-05-07 1:34 ` Dave Chinner 2 siblings, 0 replies; 33+ messages in thread From: Martin Steigerwald @ 2012-05-06 10:31 UTC (permalink / raw) To: xfs; +Cc: Stefan Priebe Am Sonntag, 6. Mai 2012 schrieb Stefan Priebe: > Hi, Hi Stefan, > since a few days i've experienced a really slow fs on one of our > backup systems. > > I'm not sure whether this is XFS related or related to the Controller > / Disks. > > It is a raid 10 of 20 SATA Disks and i can only write to them with > about 700kb/s while doing random i/o. I tried vanilla Kernel 3.0.30 > and 3.3.4 - no difference. Writing to another partition on another xfs > array works fine. How do you measure the 700 KB/s? Anything in dmesg? I and likely others here would be interested in vmstat and iostat outputs during slowness. See: http://xfs.org/index.php/XFS_FAQ#Q:_What_information_should_I_include_when_reporting_a_problem.3F Ciao, -- Martin 'Helios' Steigerwald - http://www.Lichtvoll.de GPG: 03B0 0D6C 0040 0710 4AFA B82F 991B EAAC A599 84C7 _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-06 9:01 suddenly slow writes on XFS Filesystem Stefan Priebe 2012-05-06 10:31 ` Martin Steigerwald @ 2012-05-06 10:33 ` Martin Steigerwald 2012-05-06 15:45 ` Stan Hoeppner 2012-05-07 1:34 ` Dave Chinner 2 siblings, 1 reply; 33+ messages in thread From: Martin Steigerwald @ 2012-05-06 10:33 UTC (permalink / raw) To: xfs; +Cc: Stefan Priebe Am Sonntag, 6. Mai 2012 schrieb Stefan Priebe: > It is a raid 10 of 20 SATA Disks and i can only write to them with > about 700kb/s while doing random i/o. I tried vanilla Kernel 3.0.30 > and 3.3.4 - no difference. Writing to another partition on another xfs > array works fine. Additionally what RAID is this? SoftRAID or some - which one? - hardware RAID controller? And what disks are used, whats the rpm of these? -- Martin 'Helios' Steigerwald - http://www.Lichtvoll.de GPG: 03B0 0D6C 0040 0710 4AFA B82F 991B EAAC A599 84C7 _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-06 10:33 ` Martin Steigerwald @ 2012-05-06 15:45 ` Stan Hoeppner 2012-05-06 19:25 ` Stefan Priebe 2012-05-06 21:43 ` Martin Steigerwald 0 siblings, 2 replies; 33+ messages in thread From: Stan Hoeppner @ 2012-05-06 15:45 UTC (permalink / raw) To: Martin Steigerwald; +Cc: Stefan Priebe, xfs On 5/6/2012 5:33 AM, Martin Steigerwald wrote: > Am Sonntag, 6. Mai 2012 schrieb Stefan Priebe: >> It is a raid 10 of 20 SATA Disks and i can only write to them with >> about 700kb/s while doing random i/o. I tried vanilla Kernel 3.0.30 >> and 3.3.4 - no difference. Writing to another partition on another xfs >> array works fine. > > Additionally what RAID is this? SoftRAID or some - which one? - hardware > RAID controller? And what disks are used, whats the rpm of these? I doubt much of this stuff matters. Stefan's filesystem is 96% full, w/~200GB free. This free space is likely heavily fragmented. If he's doing allocation in this fragmented free space I'd think that would fully explain his write performance dropping off a cliff due to massive head seeking. -- Stan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-06 15:45 ` Stan Hoeppner @ 2012-05-06 19:25 ` Stefan Priebe 2012-05-07 1:39 ` Dave Chinner 2012-05-06 21:43 ` Martin Steigerwald 1 sibling, 1 reply; 33+ messages in thread From: Stefan Priebe @ 2012-05-06 19:25 UTC (permalink / raw) To: stan; +Cc: xfs Am 06.05.2012 17:45, schrieb Stan Hoeppner: > On 5/6/2012 5:33 AM, Martin Steigerwald wrote: >> Am Sonntag, 6. Mai 2012 schrieb Stefan Priebe: >>> It is a raid 10 of 20 SATA Disks and i can only write to them with >>> about 700kb/s while doing random i/o. I tried vanilla Kernel 3.0.30 >>> and 3.3.4 - no difference. Writing to another partition on another xfs >>> array works fine. >> >> Additionally what RAID is this? SoftRAID or some - which one? - hardware >> RAID controller? And what disks are used, whats the rpm of these? > > I doubt much of this stuff matters. Stefan's filesystem is 96% full, > w/~200GB free. This free space is likely heavily fragmented. If he's > doing allocation in this fragmented free space I'd think that would > fully explain his write performance dropping off a cliff due to massive > head seeking. > Thanks Stan that's it. After deleting 200GB-300GB it's running fine again. What is the general recommandation of free space? Greets Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-06 19:25 ` Stefan Priebe @ 2012-05-07 1:39 ` Dave Chinner 0 siblings, 0 replies; 33+ messages in thread From: Dave Chinner @ 2012-05-07 1:39 UTC (permalink / raw) To: Stefan Priebe; +Cc: stan, xfs On Sun, May 06, 2012 at 09:25:10PM +0200, Stefan Priebe wrote: > Am 06.05.2012 17:45, schrieb Stan Hoeppner: > >On 5/6/2012 5:33 AM, Martin Steigerwald wrote: > >>Am Sonntag, 6. Mai 2012 schrieb Stefan Priebe: > >>>It is a raid 10 of 20 SATA Disks and i can only write to them with > >>>about 700kb/s while doing random i/o. I tried vanilla Kernel 3.0.30 > >>>and 3.3.4 - no difference. Writing to another partition on another xfs > >>>array works fine. > >> > >>Additionally what RAID is this? SoftRAID or some - which one? - hardware > >>RAID controller? And what disks are used, whats the rpm of these? > > > >I doubt much of this stuff matters. Stefan's filesystem is 96% full, > >w/~200GB free. This free space is likely heavily fragmented. If he's > >doing allocation in this fragmented free space I'd think that would > >fully explain his write performance dropping off a cliff due to massive > >head seeking. > > > Thanks Stan that's it. After deleting 200GB-300GB it's running fine again. > > What is the general recommandation of free space? Depends on the size of the filesystem. If you've got a 500TB filesystem, then running at 98% full (10TB of free space) is not going to be a big deal. But running at 98% full on a 5TB volume is a big deal because there is relatively little freespace per AG and it will get rapidly fragmented. So for a 5TB volume, I' say don't run sustained operations at over 90% full. Going above this temporarily won't be a problem, but staying at >95% full will definitely cause accelerated aging of the filesystem. Cheers, Dave. -- Dave Chinner david@fromorbit.com _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-06 15:45 ` Stan Hoeppner 2012-05-06 19:25 ` Stefan Priebe @ 2012-05-06 21:43 ` Martin Steigerwald 2012-05-07 6:40 ` Stefan Priebe - Profihost AG 1 sibling, 1 reply; 33+ messages in thread From: Martin Steigerwald @ 2012-05-06 21:43 UTC (permalink / raw) To: stan; +Cc: Stefan Priebe, xfs Am Sonntag, 6. Mai 2012 schrieb Stan Hoeppner: > On 5/6/2012 5:33 AM, Martin Steigerwald wrote: > > Am Sonntag, 6. Mai 2012 schrieb Stefan Priebe: > >> It is a raid 10 of 20 SATA Disks and i can only write to them with > >> about 700kb/s while doing random i/o. I tried vanilla Kernel 3.0.30 > >> and 3.3.4 - no difference. Writing to another partition on another > >> xfs array works fine. > > > > Additionally what RAID is this? SoftRAID or some - which one? - > > hardware RAID controller? And what disks are used, whats the rpm of > > these? > > I doubt much of this stuff matters. Stefan's filesystem is 96% full, > w/~200GB free. This free space is likely heavily fragmented. If he's > doing allocation in this fragmented free space I'd think that would > fully explain his write performance dropping off a cliff due to massive > head seeking. Hmmm, I thought about that low free space percentage, too, but then 200 GB free still sounded like quite much to me. Nonetheless, since according to Stefan freeing some space solved the slowness issue, that seems to have been it. Can free space be defragmented in XFS? xfs_fsr only tries to defragment files as far as I know. Thanks, -- Martin 'Helios' Steigerwald - http://www.Lichtvoll.de GPG: 03B0 0D6C 0040 0710 4AFA B82F 991B EAAC A599 84C7 _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-06 21:43 ` Martin Steigerwald @ 2012-05-07 6:40 ` Stefan Priebe - Profihost AG 0 siblings, 0 replies; 33+ messages in thread From: Stefan Priebe - Profihost AG @ 2012-05-07 6:40 UTC (permalink / raw) To: Martin Steigerwald; +Cc: stan, xfs Am 06.05.2012 23:43, schrieb Martin Steigerwald: > Am Sonntag, 6. Mai 2012 schrieb Stan Hoeppner: >> On 5/6/2012 5:33 AM, Martin Steigerwald wrote: >>> Am Sonntag, 6. Mai 2012 schrieb Stefan Priebe: >>>> It is a raid 10 of 20 SATA Disks and i can only write to them with >>>> about 700kb/s while doing random i/o. I tried vanilla Kernel 3.0.30 >>>> and 3.3.4 - no difference. Writing to another partition on another >>>> xfs array works fine. >>> >>> Additionally what RAID is this? SoftRAID or some - which one? - >>> hardware RAID controller? And what disks are used, whats the rpm of >>> these? It is a HW Adaptec 5805 Controller with 20x 7200rpm SATA Disks. > Can free space be defragmented in XFS? xfs_fsr only tries to defragment > files as far as I know. That is a good question. Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-06 9:01 suddenly slow writes on XFS Filesystem Stefan Priebe 2012-05-06 10:31 ` Martin Steigerwald 2012-05-06 10:33 ` Martin Steigerwald @ 2012-05-07 1:34 ` Dave Chinner 2012-05-07 6:39 ` Stefan Priebe - Profihost AG 2 siblings, 1 reply; 33+ messages in thread From: Dave Chinner @ 2012-05-07 1:34 UTC (permalink / raw) To: Stefan Priebe; +Cc: xfs On Sun, May 06, 2012 at 11:01:14AM +0200, Stefan Priebe wrote: > Hi, > > since a few days i've experienced a really slow fs on one of our > backup systems. > > I'm not sure whether this is XFS related or related to the > Controller / Disks. > > It is a raid 10 of 20 SATA Disks and i can only write to them with > about 700kb/s while doing random i/o. What sort of random IO? size, read, write, direct or buffered, data or metadata, etc? iostat -x -d -m 5 and vmstat 5 traces would be useful to see if it is your array that is slow..... > I tried vanilla Kernel 3.0.30 > and 3.3.4 - no difference. Writing to another partition on another > xfs array works fine. > > Details: > #~ df -h > /dev/sdb1 4,6T 4,4T 207G 96% /mnt Your filesystem is near full - the allocation algorithms definitely slow down as you approach ENOSPC, and IO efficiency goes to hell because of a lack of contiguous free space to allocate from. > #~ df -i > /dev/sdb1 4875737052 4659318044 216419008 96% /mnt You have 4.6 *billion* inodes in your filesystem? > Any ideas? None until I understand your workload.... Cheers, Dave. -- Dave Chinner david@fromorbit.com _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 1:34 ` Dave Chinner @ 2012-05-07 6:39 ` Stefan Priebe - Profihost AG 2012-05-07 7:17 ` Dave Chinner ` (2 more replies) 0 siblings, 3 replies; 33+ messages in thread From: Stefan Priebe - Profihost AG @ 2012-05-07 6:39 UTC (permalink / raw) To: Dave Chinner; +Cc: stan, xfs Hi, after deleting 400GB it was faster. Now there are still 300GB free but it is slow as hell again ;-( Am 07.05.2012 03:34, schrieb Dave Chinner: > On Sun, May 06, 2012 at 11:01:14AM +0200, Stefan Priebe wrote: >> Hi, >> >> since a few days i've experienced a really slow fs on one of our >> backup systems. >> >> I'm not sure whether this is XFS related or related to the >> Controller / Disks. >> >> It is a raid 10 of 20 SATA Disks and i can only write to them with >> about 700kb/s while doing random i/o. > > What sort of random IO? size, read, write, direct or buffered, data > or metadata, etc? There are 4 rsync processes running and doing backups of other severs. > iostat -x -d -m 5 and vmstat 5 traces would be > useful to see if it is your array that is slow..... ~ # iostat -x -d -m 5 Linux 2.6.40.28intel (server844-han) 05/07/2012 _x86_64_ (8 CPU) Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util sdb 0,00 0,00 254,80 25,40 1,72 0,16 13,71 0,86 3,08 2,39 67,06 sda 0,00 0,20 0,00 1,20 0,00 0,00 6,50 0,00 0,00 0,00 0,00 Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util sdb 0,00 0,00 187,40 24,20 1,26 0,19 14,05 0,75 3,56 3,33 70,50 sda 0,00 0,00 0,00 0,40 0,00 0,00 4,50 0,00 0,00 0,00 0,00 Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util sdb 0,00 11,20 242,40 92,00 1,56 0,89 15,00 4,70 14,06 1,58 52,68 sda 0,00 0,20 0,00 2,60 0,00 0,02 12,00 0,00 0,00 0,00 0,00 Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util sdb 0,00 0,00 166,20 24,00 0,99 0,17 12,51 0,57 3,02 2,40 45,56 sda 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 qDevice: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util sdb 0,00 0,00 188,00 25,40 1,22 0,16 13,23 0,44 2,04 1,78 38,02 sda 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 # vmstat procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 7 0 0 788632 48 12189652 0 0 173 395 13 45 1 16 82 1 [root@server844-han /serverbackup (master)]# vmstat procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 4 0 0 778148 48 12189776 0 0 173 395 13 45 1 16 82 1 [root@server844-han /serverbackup (master)]# vmstat procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 2 0 0 774372 48 12189876 0 0 173 395 13 45 1 16 82 1 [root@server844-han /serverbackup (master)]# vmstat procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 5 0 0 771240 48 12189936 0 0 173 395 13 45 1 16 82 1 [root@server844-han /serverbackup (master)]# vmstat procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 6 0 0 768636 48 12190000 0 0 173 395 13 45 1 16 82 1 > >> I tried vanilla Kernel 3.0.30 >> and 3.3.4 - no difference. Writing to another partition on another >> xfs array works fine. >> >> Details: >> #~ df -h >> /dev/sdb1 4,6T 4,4T 207G 96% /mnt > > Your filesystem is near full - the allocation algorithms definitely > slow down as you approach ENOSPC, and IO efficiency goes to hell > because of a lack of contiguous free space to allocate from. I've now 94% used but it is still slow. It seems it was just getting fast with more than 450GB free space. /dev/sdb1 4,6T 4,3T 310G 94% /mnt >> #~ df -i >> /dev/sdb1 4875737052 4659318044 216419008 96% /mnt > You have 4.6 *billion* inodes in your filesystem? Yes - it backups around 100 servers with a lot of files. Greet Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 6:39 ` Stefan Priebe - Profihost AG @ 2012-05-07 7:17 ` Dave Chinner 2012-05-07 7:22 ` Stefan Priebe - Profihost AG 2012-05-07 8:21 ` Martin Steigerwald 2012-05-07 8:31 ` Martin Steigerwald 2 siblings, 1 reply; 33+ messages in thread From: Dave Chinner @ 2012-05-07 7:17 UTC (permalink / raw) To: Stefan Priebe - Profihost AG; +Cc: stan, xfs On Mon, May 07, 2012 at 08:39:13AM +0200, Stefan Priebe - Profihost AG wrote: > Hi, > > after deleting 400GB it was faster. Now there are still 300GB free but > it is slow as hell again ;-( > > Am 07.05.2012 03:34, schrieb Dave Chinner: > > On Sun, May 06, 2012 at 11:01:14AM +0200, Stefan Priebe wrote: > >> Hi, > >> > >> since a few days i've experienced a really slow fs on one of our > >> backup systems. > >> > >> I'm not sure whether this is XFS related or related to the > >> Controller / Disks. > >> > >> It is a raid 10 of 20 SATA Disks and i can only write to them with > >> about 700kb/s while doing random i/o. > > > > What sort of random IO? size, read, write, direct or buffered, data > > or metadata, etc? > There are 4 rsync processes running and doing backups of other severs. > > > iostat -x -d -m 5 and vmstat 5 traces would be > > useful to see if it is your array that is slow..... > > ~ # iostat -x -d -m 5 > Linux 2.6.40.28intel (server844-han) 05/07/2012 _x86_64_ > (8 CPU) > > Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util > sdb 0,00 0,00 254,80 25,40 1,72 0,16 13,71 0,86 3,08 2,39 67,06 > sda 0,00 0,20 0,00 1,20 0,00 0,00 6,50 0,00 0,00 0,00 0,00 > > Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util > sdb 0,00 0,00 187,40 24,20 1,26 0,19 14,05 0,75 3,56 3,33 70,50 > sda 0,00 0,00 0,00 0,40 0,00 0,00 4,50 0,00 0,00 0,00 0,00 > > Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util > sdb 0,00 11,20 242,40 92,00 1,56 0,89 15,00 4,70 14,06 1,58 52,68 > sda 0,00 0,20 0,00 2,60 0,00 0,02 12,00 0,00 0,00 0,00 0,00 > > Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util > sdb 0,00 0,00 166,20 24,00 0,99 0,17 12,51 0,57 3,02 2,40 45,56 > sda 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 > > Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await svctm %util > sdb 0,00 0,00 188,00 25,40 1,22 0,16 13,23 0,44 2,04 1,78 38,02 > sda 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 0,00 > # vmstat "vmstat 5", not vmstat 5 times.... :/ > >> I tried vanilla Kernel 3.0.30 > >> and 3.3.4 - no difference. Writing to another partition on another > >> xfs array works fine. > >> > >> Details: > >> #~ df -h > >> /dev/sdb1 4,6T 4,4T 207G 96% /mnt > > > > Your filesystem is near full - the allocation algorithms definitely > > slow down as you approach ENOSPC, and IO efficiency goes to hell > > because of a lack of contiguous free space to allocate from. > I've now 94% used but it is still slow. It seems it was just getting > fast with more than 450GB free space. > > /dev/sdb1 4,6T 4,3T 310G 94% /mnt Well, you've probably badly fragmented the free space you have. what does the 'xfs_db -r -c freesp <dev>' command tell you? > >> #~ df -i > >> /dev/sdb1 4875737052 4659318044 216419008 96% /mnt > > You have 4.6 *billion* inodes in your filesystem? > Yes - it backups around 100 servers with a lot of files. So you have what - lots of symlinks? I mean, 4.6 billion inodes alone requires 1.2TB of space, but if I read the fragmentation you only have 82 million files with data extents. The only thing that would other wise use inodes are directories and symlinks.... Still, I can't see how you'd only have 82 million data inodes and 4.5 billion directory inodes - where are all the inodes being consumed? A massive symlink farm? Cheers, Dave. -- Dave Chinner david@fromorbit.com _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 7:17 ` Dave Chinner @ 2012-05-07 7:22 ` Stefan Priebe - Profihost AG 2012-05-07 16:36 ` Stan Hoeppner 2012-05-07 23:42 ` suddenly slow writes on XFS Filesystem Dave Chinner 0 siblings, 2 replies; 33+ messages in thread From: Stefan Priebe - Profihost AG @ 2012-05-07 7:22 UTC (permalink / raw) To: Dave Chinner; +Cc: stan, xfs >> # vmstat > "vmstat 5", not vmstat 5 times.... :/ oh sorry. Sadly the rsync processes do not run right know i've to kill them. Is the output still usable? # vmstat 5 procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 0 1 0 5582136 48 5849956 0 0 176 394 34 54 1 16 82 1 0 1 0 5552180 48 5854280 0 0 2493 2496 3079 2172 1 4 86 9 3 2 0 5601308 48 5857672 0 0 1098 28043 5150 1913 0 10 73 17 0 2 0 5595360 48 5863180 0 0 1098 14336 3945 1897 0 8 69 22 3 2 0 5594088 48 5865280 0 0 432 15897 4209 2366 0 8 71 21 0 2 0 5591068 48 5868940 0 0 854 10989 3519 2107 0 7 70 23 1 1 0 5592004 48 5869872 0 0 180 7886 3605 2436 0 3 76 22 >> /dev/sdb1 4,6T 4,3T 310G 94% /mnt > Well, you've probably badly fragmented the free space you have. what > does the 'xfs_db -r -c freesp <dev>' command tell you? from to extents blocks pct 1 1 942737 942737 0,87 2 3 671860 1590480 1,47 4 7 461268 2416025 2,23 8 15 1350517 18043063 16,67 16 31 111254 2547581 2,35 32 63 192032 9039799 8,35 64 127 33026 3317194 3,07 128 255 14254 2665812 2,46 256 511 12516 4631200 4,28 512 1023 6942 5031081 4,65 1024 2047 4622 6893270 6,37 2048 4095 3268 9412271 8,70 4096 8191 2135 12716435 11,75 8192 16383 338 3974884 3,67 16384 32767 311 7018658 6,49 32768 65535 105 4511372 4,17 65536 131071 29 2577756 2,38 131072 262143 8 1339796 1,24 262144 524287 10 3950416 3,65 524288 1048575 4 2580085 2,38 1048576 2097151 2 3028028 2,80 >>>> #~ df -i >>>> /dev/sdb1 4875737052 4659318044 216419008 96% /mnt >>> You have 4.6 *billion* inodes in your filesystem? >> Yes - it backups around 100 servers with a lot of files. i rechecked this and it seems i sadly copied the wrong output ;-( sorry for that. Here is the correct one: #~ df -i /dev/sdb1 975173568 95212355 879961213 10% /mnt Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 7:22 ` Stefan Priebe - Profihost AG @ 2012-05-07 16:36 ` Stan Hoeppner 2012-05-07 19:08 ` Martin Steigerwald 2012-05-07 20:05 ` Stefan Priebe 2012-05-07 23:42 ` suddenly slow writes on XFS Filesystem Dave Chinner 1 sibling, 2 replies; 33+ messages in thread From: Stan Hoeppner @ 2012-05-07 16:36 UTC (permalink / raw) To: Stefan Priebe - Profihost AG; +Cc: xfs On 5/7/2012 2:22 AM, Stefan Priebe - Profihost AG wrote: > from to extents blocks pct > 1 1 942737 942737 0,87 > 2 3 671860 1590480 1,47 > 4 7 461268 2416025 2,23 > 8 15 1350517 18043063 16,67 > 16 31 111254 2547581 2,35 > 32 63 192032 9039799 8,35 > 64 127 33026 3317194 3,07 > 128 255 14254 2665812 2,46 > 256 511 12516 4631200 4,28 > 512 1023 6942 5031081 4,65 > 1024 2047 4622 6893270 6,37 > 2048 4095 3268 9412271 8,70 > 4096 8191 2135 12716435 11,75 > 8192 16383 338 3974884 3,67 > 16384 32767 311 7018658 6,49 > 32768 65535 105 4511372 4,17 > 65536 131071 29 2577756 2,38 > 131072 262143 8 1339796 1,24 > 262144 524287 10 3950416 3,65 > 524288 1048575 4 2580085 2,38 > 1048576 2097151 2 3028028 2,80 This shows what I originally suspected. Notice how top heavy this histogram is. Over half of your free space sits on little islands of 8MB or less. 17% is in islands of 60KB or less. This is heavily fragmented free space. Contrast this with an XFS from the opposite end of the aging spectrum that is only 1/3rd full and has seen very few deletes as it has aged: from to extents blocks pct 1 1 1028 1028 0.01 2 3 1185 2864 0.02 4 7 1656 9007 0.06 8 15 5232 67674 0.41 16 31 36 775 0.00 32 63 7 292 0.00 64 127 4 350 0.00 128 255 6 1286 0.01 256 511 4 1460 0.01 512 1023 5 4278 0.03 1024 2047 5 6965 0.04 2048 4095 5 12935 0.08 4096 8191 1 8179 0.05 8192 16383 2 29570 0.18 16384 32767 2 60352 0.37 32768 65535 3 148594 0.91 65536 131071 2 195575 1.20 131072 262143 2 420917 2.57 524288 1048575 2 1499689 9.17 4194304 6103694 3 13876795 84.88 Notice how it is very bottom heavy, and that 85% of the free space is in large islands of 16GB to 24GB. Stefan, at this point in your filesystem's aging process, it may not matter how much space you keep freeing up, as your deletion of small files simply adds more heavily fragmented free space to the pool. It's the nature of your workload causing this. If you were rsyncing and deleting 500MB files the story would be much different. What I would suggest is doing an xfsdump to a filesystem on another LUN or machine, expand the size of this LUN by 50% or more (I gather this is an external RAID), format it appropriately, then xfsrestore. This will eliminate your current free space fragmentation, and the 50% size increase will delay the next occurrence of this problem. If you can't expand the LUN, simply do the xfsdump/format/xfsrestore, which will give you contiguous free space. -- Stan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 16:36 ` Stan Hoeppner @ 2012-05-07 19:08 ` Martin Steigerwald 2012-05-07 20:05 ` Stefan Priebe 1 sibling, 0 replies; 33+ messages in thread From: Martin Steigerwald @ 2012-05-07 19:08 UTC (permalink / raw) To: xfs, stan; +Cc: Stefan Priebe - Profihost AG Am Montag, 7. Mai 2012 schrieb Stan Hoeppner: > What I would suggest is doing an xfsdump to a filesystem on another LUN > or machine, expand the size of this LUN by 50% or more (I gather this > is an external RAID), format it appropriately, then xfsrestore. This > will eliminate your current free space fragmentation, and the 50% size > increase will delay the next occurrence of this problem. If you can't > expand the LUN, simply do the xfsdump/format/xfsrestore, which will > give you contiguous free space. Maybe the new dm thin provisioning could help here? But it might just move the fragmentation from the filesystem to the device mapper layer. I did not yet check the implementation details of the new thin provisioning target. -- Martin 'Helios' Steigerwald - http://www.Lichtvoll.de GPG: 03B0 0D6C 0040 0710 4AFA B82F 991B EAAC A599 84C7 _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 16:36 ` Stan Hoeppner 2012-05-07 19:08 ` Martin Steigerwald @ 2012-05-07 20:05 ` Stefan Priebe 2012-05-09 6:57 ` Stan Hoeppner 1 sibling, 1 reply; 33+ messages in thread From: Stefan Priebe @ 2012-05-07 20:05 UTC (permalink / raw) To: stan; +Cc: xfs Am 07.05.2012 18:36, schrieb Stan Hoeppner: > This shows what I originally suspected. Notice how top heavy this > histogram is. Over half of your free space sits on little islands of > 8MB or less. 17% is in islands of 60KB or less. This is heavily > fragmented free space. Contrast this with an XFS from the opposite end > of the aging spectrum that is only 1/3rd full and has seen very few > deletes as it has aged: ... > > Notice how it is very bottom heavy, and that 85% of the free space is in > large islands of 16GB to 24GB. This totally makes sense too me. Thanks for this explanation. > Stefan, at this point in your filesystem's aging process, it may not > matter how much space you keep freeing up, as your deletion of small > files simply adds more heavily fragmented free space to the pool. It's > the nature of your workload causing this. This makes sense - do you have any idea or solution for this? Are Filesystems, Block layers or something else which suits this problem / situation? > What I would suggest is doing an xfsdump to a filesystem on another LUN > or machine, expand the size of this LUN by 50% or more (I gather this is > an external RAID), format it appropriately, then xfsrestore. This will > eliminate your current free space fragmentation, and the 50% size > increase will delay the next occurrence of this problem. If you can't > expand the LUN, simply do the xfsdump/format/xfsrestore, which will give > you contiguous free space. But this will only help for a few month or perhaps a year. Greets, Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 20:05 ` Stefan Priebe @ 2012-05-09 6:57 ` Stan Hoeppner 2012-05-09 7:04 ` Dave Chinner 0 siblings, 1 reply; 33+ messages in thread From: Stan Hoeppner @ 2012-05-09 6:57 UTC (permalink / raw) To: Stefan Priebe; +Cc: xfs On 5/7/2012 3:05 PM, Stefan Priebe wrote: > Am 07.05.2012 18:36, schrieb Stan Hoeppner: >> Stefan, at this point in your filesystem's aging process, it may not >> matter how much space you keep freeing up, as your deletion of small >> files simply adds more heavily fragmented free space to the pool. It's >> the nature of your workload causing this. > This makes sense - do you have any idea or solution for this? Are > Filesystems, Block layers or something else which suits this problem / > situation? The problem isn't the block layer nor the filesystem. The problem is a combination of the workload and filling the FS to near capacity. Any workload that regularly allocates and then deletes large quantities of small files and fills up the filesystem is going to suffer poor performance from free space fragmentation as the water in the FS gets close to the rim of the glass. Two other example workloads are large mail spools on very busy internet mail servers, and maildir storage on IMAP/POP servers. In your case there are two solutions to this problem, the second of which is also the solution for these mail workloads: 1. Use a backup methodology that writes larger files 2. Give your workload a much larger sandbox to play in Regarding #1, if you're using rsnapshot your disk space shouldn't be continuously growing, which is does seem to be. If you're not using rsnapshot look into it. Regarding #2 ... >> What I would suggest is doing an xfsdump to a filesystem on another LUN >> or machine, expand the size of this LUN by 50% or more (I gather this is >> an external RAID), format it appropriately, then xfsrestore. This will >> eliminate your current free space fragmentation, and the 50% size >> increase will delay the next occurrence of this problem. If you can't >> expand the LUN, simply do the xfsdump/format/xfsrestore, which will give >> you contiguous free space. > But this will only help for a few month or perhaps a year. So you are saying your backup solution will fill up an additional 2.3TB in less than a year? In that case I'd say you have dramatically undersized your backup storage and/or are not using file compression to your advantage. And you're obviously not using archiving to your advantage or you'd not have the free space fragmentation issue because you'd be dealing with much larger files. So the best solution to your current problem, and one that will save you disk space and thus $$, is to use a backup solution that makes use of both tar and gzip/bzip2. You can't fix this fundamental small file free space fragmentation problem by tuning/tweaking XFS, or switching to another filesystem, as again, the problem is the workload, not the block layer or FS. -- Stan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-09 6:57 ` Stan Hoeppner @ 2012-05-09 7:04 ` Dave Chinner 2012-05-09 7:36 ` Stefan Priebe - Profihost AG ` (2 more replies) 0 siblings, 3 replies; 33+ messages in thread From: Dave Chinner @ 2012-05-09 7:04 UTC (permalink / raw) To: Stan Hoeppner; +Cc: xfs, Stefan Priebe On Wed, May 09, 2012 at 01:57:01AM -0500, Stan Hoeppner wrote: > On 5/7/2012 3:05 PM, Stefan Priebe wrote: > > Am 07.05.2012 18:36, schrieb Stan Hoeppner: .... > >> What I would suggest is doing an xfsdump to a filesystem on another LUN > >> or machine, expand the size of this LUN by 50% or more (I gather this is > >> an external RAID), format it appropriately, then xfsrestore. This will > >> eliminate your current free space fragmentation, and the 50% size > >> increase will delay the next occurrence of this problem. If you can't > >> expand the LUN, simply do the xfsdump/format/xfsrestore, which will give > >> you contiguous free space. > > But this will only help for a few month or perhaps a year. > > So you are saying your backup solution will fill up an additional 2.3TB > in less than a year? In that case I'd say you have dramatically > undersized your backup storage and/or are not using file compression to > your advantage. And you're obviously not using archiving to your > advantage or you'd not have the free space fragmentation issue because > you'd be dealing with much larger files. > > So the best solution to your current problem, and one that will save you > disk space and thus $$, is to use a backup solution that makes use of > both tar and gzip/bzip2. Why use the slow method? xfsdump will be much faster than tar. :) Cheers, Dave. -- Dave Chinner david@fromorbit.com _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-09 7:04 ` Dave Chinner @ 2012-05-09 7:36 ` Stefan Priebe - Profihost AG 2012-05-09 7:49 ` Stan Hoeppner 2013-02-15 15:06 ` 32bit apps and inode64 Stefan Priebe - Profihost AG 2 siblings, 0 replies; 33+ messages in thread From: Stefan Priebe - Profihost AG @ 2012-05-09 7:36 UTC (permalink / raw) To: Dave Chinner; +Cc: Stan Hoeppner, xfs Am 09.05.2012 09:04, schrieb Dave Chinner: >> So the best solution to your current problem, and one that will save you >> disk space and thus $$, is to use a backup solution that makes use of >> both tar and gzip/bzip2. > > Why use the slow method? xfsdump will be much faster than tar. :) I tried to use it in this special case but it is hanging on dumping directories for 3 hours now or is this normal? Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-09 7:04 ` Dave Chinner 2012-05-09 7:36 ` Stefan Priebe - Profihost AG @ 2012-05-09 7:49 ` Stan Hoeppner 2013-02-15 15:06 ` 32bit apps and inode64 Stefan Priebe - Profihost AG 2 siblings, 0 replies; 33+ messages in thread From: Stan Hoeppner @ 2012-05-09 7:49 UTC (permalink / raw) To: Dave Chinner; +Cc: Stefan Priebe, xfs On 5/9/2012 2:04 AM, Dave Chinner wrote: > On Wed, May 09, 2012 at 01:57:01AM -0500, Stan Hoeppner wrote: >> On 5/7/2012 3:05 PM, Stefan Priebe wrote: >>> Am 07.05.2012 18:36, schrieb Stan Hoeppner: > .... >>>> What I would suggest is doing an xfsdump to a filesystem on another LUN >>>> or machine, expand the size of this LUN by 50% or more (I gather this is >>>> an external RAID), format it appropriately, then xfsrestore. This will >>>> eliminate your current free space fragmentation, and the 50% size >>>> increase will delay the next occurrence of this problem. If you can't >>>> expand the LUN, simply do the xfsdump/format/xfsrestore, which will give >>>> you contiguous free space. >>> But this will only help for a few month or perhaps a year. >> >> So you are saying your backup solution will fill up an additional 2.3TB >> in less than a year? In that case I'd say you have dramatically >> undersized your backup storage and/or are not using file compression to >> your advantage. And you're obviously not using archiving to your >> advantage or you'd not have the free space fragmentation issue because >> you'd be dealing with much larger files. >> >> So the best solution to your current problem, and one that will save you >> disk space and thus $$, is to use a backup solution that makes use of >> both tar and gzip/bzip2. > > Why use the slow method? xfsdump will be much faster than tar. :) Certainly faster. I've assumed Stefan is backing up heterogeneous systems so xfsdump is probably not an option across the board. My thinking directly above related to the possibility that his current software may have integrated functions for tar'ing then compressing at the directory level. This would save space, cause less free space fragmentation when old files are deleted, and allow decompressing files that aren't insanely large in the even a single file needs to be restored from a compressed tar file. Of course the latter can be accomplished with xfsdump/restore, but not on heterogeneous systems. And AFAIK there isn't any nifty FOSS backup software that makes use of xfsdump/restore, requiring some serious custom scripting. If there exists such backup software please do share. :) -- Stan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* 32bit apps and inode64 2012-05-09 7:04 ` Dave Chinner 2012-05-09 7:36 ` Stefan Priebe - Profihost AG 2012-05-09 7:49 ` Stan Hoeppner @ 2013-02-15 15:06 ` Stefan Priebe - Profihost AG 2013-02-15 21:46 ` Ben Myers 2 siblings, 1 reply; 33+ messages in thread From: Stefan Priebe - Profihost AG @ 2013-02-15 15:06 UTC (permalink / raw) To: xfs Hello list, i've discovered some problems on a host with a disk > 1TB. We've some binary 32bit applications which are not able to read some directory anymore after we've formated and installed the system using vanilla 3.7.7 kernel. Right now we're using 3.0.61 kernel on this host - so 64bit apps work well and newly created files get 32bit inode numbers as inode64 is not the default. Is there a way to find / get all 64bit inode files / dies and convert them back to 32bit without a reinstall? Greets, Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: 32bit apps and inode64 2013-02-15 15:06 ` 32bit apps and inode64 Stefan Priebe - Profihost AG @ 2013-02-15 21:46 ` Ben Myers 2013-02-16 10:24 ` Stefan Priebe - Profihost AG 2013-02-17 8:13 ` Jeff Liu 0 siblings, 2 replies; 33+ messages in thread From: Ben Myers @ 2013-02-15 21:46 UTC (permalink / raw) To: Stefan Priebe - Profihost AG; +Cc: xfs Hi Stefan, On Fri, Feb 15, 2013 at 04:06:40PM +0100, Stefan Priebe - Profihost AG wrote: > i've discovered some problems on a host with a disk > 1TB. We've some > binary 32bit applications which are not able to read some directory > anymore after we've formated and installed the system using vanilla > 3.7.7 kernel. > > Right now we're using 3.0.61 kernel on this host - so 64bit apps work > well and newly created files get 32bit inode numbers as inode64 is not > the default. > > Is there a way to find / get all 64bit inode files / dies and convert > them back to 32bit without a reinstall? On IRIX you could use xfs_reno to renumber those inodes. http://techpubs.sgi.com/library/tpl/cgi-bin/getdoc.cgi?coll=0650&db=man&fname=/usr/share/catman/a_man/cat1/xfs_reno.z xfs_reno was ported to linux in '07 and was most recently reposted by Jeff Liu: http://oss.sgi.com/archives/xfs/2012-11/msg00425.html It isn't in xfsprogs today. Regards, Ben _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: 32bit apps and inode64 2013-02-15 21:46 ` Ben Myers @ 2013-02-16 10:24 ` Stefan Priebe - Profihost AG 2013-02-17 21:33 ` Dave Chinner 2013-02-17 8:13 ` Jeff Liu 1 sibling, 1 reply; 33+ messages in thread From: Stefan Priebe - Profihost AG @ 2013-02-16 10:24 UTC (permalink / raw) To: Ben Myers; +Cc: xfs I'm on linux. So no recommended way for production systems? Stefan Am 15.02.2013 um 22:46 schrieb Ben Myers <bpm@sgi.com>: > Hi Stefan, > > On Fri, Feb 15, 2013 at 04:06:40PM +0100, Stefan Priebe - Profihost AG wrote: >> i've discovered some problems on a host with a disk > 1TB. We've some >> binary 32bit applications which are not able to read some directory >> anymore after we've formated and installed the system using vanilla >> 3.7.7 kernel. >> >> Right now we're using 3.0.61 kernel on this host - so 64bit apps work >> well and newly created files get 32bit inode numbers as inode64 is not >> the default. >> >> Is there a way to find / get all 64bit inode files / dies and convert >> them back to 32bit without a reinstall? > > On IRIX you could use xfs_reno to renumber those inodes. > http://techpubs.sgi.com/library/tpl/cgi-bin/getdoc.cgi?coll=0650&db=man&fname=/usr/share/catman/a_man/cat1/xfs_reno.z > > xfs_reno was ported to linux in '07 and was most recently reposted by Jeff Liu: > http://oss.sgi.com/archives/xfs/2012-11/msg00425.html > > It isn't in xfsprogs today. > > Regards, > Ben _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: 32bit apps and inode64 2013-02-16 10:24 ` Stefan Priebe - Profihost AG @ 2013-02-17 21:33 ` Dave Chinner 2013-02-18 8:12 ` Stefan Priebe - Profihost AG 0 siblings, 1 reply; 33+ messages in thread From: Dave Chinner @ 2013-02-17 21:33 UTC (permalink / raw) To: Stefan Priebe - Profihost AG; +Cc: Ben Myers, xfs On Sat, Feb 16, 2013 at 11:24:13AM +0100, Stefan Priebe - Profihost AG wrote: > I'm on linux. So no recommended way for production systems? > > Stefan > > Am 15.02.2013 um 22:46 schrieb Ben Myers <bpm@sgi.com>: > > > Hi Stefan, > > > > On Fri, Feb 15, 2013 at 04:06:40PM +0100, Stefan Priebe - Profihost AG wrote: > >> i've discovered some problems on a host with a disk > 1TB. We've some > >> binary 32bit applications which are not able to read some directory > >> anymore after we've formated and installed the system using vanilla > >> 3.7.7 kernel. > >> > >> Right now we're using 3.0.61 kernel on this host - so 64bit apps work > >> well and newly created files get 32bit inode numbers as inode64 is not > >> the default. > >> > >> Is there a way to find / get all 64bit inode files / dies and convert > >> them back to 32bit without a reinstall? > > > > On IRIX you could use xfs_reno to renumber those inodes. > > http://techpubs.sgi.com/library/tpl/cgi-bin/getdoc.cgi?coll=0650&db=man&fname=/usr/share/catman/a_man/cat1/xfs_reno.z > > > > xfs_reno was ported to linux in '07 and was most recently reposted by Jeff Liu: > > http://oss.sgi.com/archives/xfs/2012-11/msg00425.html > > > > It isn't in xfsprogs today. Simple answer: mount with inode32, run find to print out all the filenames in the filesystem and their inode number, copy the files with inodes numbers greater than 32 bit to a temporary file and then rename them over the top of the original. That's effectively all xfs_reno does, anyway, just with faster algorithms (like bulkstat) and a bunch of crash resiliency semantics wrapped around it.... Cheers, Dave. -- Dave Chinner david@fromorbit.com _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: 32bit apps and inode64 2013-02-17 21:33 ` Dave Chinner @ 2013-02-18 8:12 ` Stefan Priebe - Profihost AG 2013-02-18 22:06 ` Dave Chinner 0 siblings, 1 reply; 33+ messages in thread From: Stefan Priebe - Profihost AG @ 2013-02-18 8:12 UTC (permalink / raw) To: Dave Chinner; +Cc: Ben Myers, xfs Hi, Am 17.02.2013 22:33, schrieb Dave Chinner: >>> xfs_reno was ported to linux in '07 and was most recently reposted by Jeff Liu: >>> http://oss.sgi.com/archives/xfs/2012-11/msg00425.html >>> >>> It isn't in xfsprogs today. > > Simple answer: mount with inode32, run find to print out all the > filenames in the filesystem and their inode number, copy the files > with inodes numbers greater than 32 bit to a temporary file and then > rename them over the top of the original. Thanks, what about directories? Is renaming / moving using mv enough? Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: 32bit apps and inode64 2013-02-18 8:12 ` Stefan Priebe - Profihost AG @ 2013-02-18 22:06 ` Dave Chinner 0 siblings, 0 replies; 33+ messages in thread From: Dave Chinner @ 2013-02-18 22:06 UTC (permalink / raw) To: Stefan Priebe - Profihost AG; +Cc: Ben Myers, xfs On Mon, Feb 18, 2013 at 09:12:03AM +0100, Stefan Priebe - Profihost AG wrote: > Hi, > > Am 17.02.2013 22:33, schrieb Dave Chinner: > > >>> xfs_reno was ported to linux in '07 and was most recently reposted by Jeff Liu: > >>> http://oss.sgi.com/archives/xfs/2012-11/msg00425.html > >>> > >>> It isn't in xfsprogs today. > > > > Simple answer: mount with inode32, run find to print out all the > > filenames in the filesystem and their inode number, copy the files > > with inodes numbers greater than 32 bit to a temporary file and then > > rename them over the top of the original. > > Thanks, what about directories? Is renaming / moving using mv enough? Same thing - create a new temp directory, mv all the entries (which will be a rename operation, anyway) from the old directory into the temp directory, rename the temp dir over the top of the old one. Cheers, Dave. -- Dave Chinner david@fromorbit.com _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: 32bit apps and inode64 2013-02-15 21:46 ` Ben Myers 2013-02-16 10:24 ` Stefan Priebe - Profihost AG @ 2013-02-17 8:13 ` Jeff Liu 2013-02-19 19:11 ` Ben Myers 1 sibling, 1 reply; 33+ messages in thread From: Jeff Liu @ 2013-02-17 8:13 UTC (permalink / raw) To: Ben Myers; +Cc: xfs, Stefan Priebe - Profihost AG Hi Ben, On 02/16/2013 05:46 AM, Ben Myers wrote: > Hi Stefan, > > On Fri, Feb 15, 2013 at 04:06:40PM +0100, Stefan Priebe - Profihost AG wrote: >> i've discovered some problems on a host with a disk > 1TB. We've some >> binary 32bit applications which are not able to read some directory >> anymore after we've formated and installed the system using vanilla >> 3.7.7 kernel. >> >> Right now we're using 3.0.61 kernel on this host - so 64bit apps work >> well and newly created files get 32bit inode numbers as inode64 is not >> the default. >> >> Is there a way to find / get all 64bit inode files / dies and convert >> them back to 32bit without a reinstall? > > On IRIX you could use xfs_reno to renumber those inodes. > http://techpubs.sgi.com/library/tpl/cgi-bin/getdoc.cgi?coll=0650&db=man&fname=/usr/share/catman/a_man/cat1/xfs_reno.z > > xfs_reno was ported to linux in '07 and was most recently reposted by Jeff Liu: > http://oss.sgi.com/archives/xfs/2012-11/msg00425.html The old patch set was belong to the infrastructures of online shrinking support. Recently, I realized that I have made a few mistakes in swap inodes ioctl(2) implementation when I revisit the old patch set at: http://oss.sgi.com/archives/xfs/2012-11/msg00414.html Since we have user request and this function is independent to the shrinking feature, I'd like to work on it at first if you like it. Thanks, -Jeff _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: 32bit apps and inode64 2013-02-17 8:13 ` Jeff Liu @ 2013-02-19 19:11 ` Ben Myers 0 siblings, 0 replies; 33+ messages in thread From: Ben Myers @ 2013-02-19 19:11 UTC (permalink / raw) To: Jeff Liu; +Cc: xfs, Stefan Priebe - Profihost AG Hey Jeff, On Sun, Feb 17, 2013 at 04:13:28PM +0800, Jeff Liu wrote: > Hi Ben, > > On 02/16/2013 05:46 AM, Ben Myers wrote: > > Hi Stefan, > > > > On Fri, Feb 15, 2013 at 04:06:40PM +0100, Stefan Priebe - Profihost AG wrote: > >> i've discovered some problems on a host with a disk > 1TB. We've some > >> binary 32bit applications which are not able to read some directory > >> anymore after we've formated and installed the system using vanilla > >> 3.7.7 kernel. > >> > >> Right now we're using 3.0.61 kernel on this host - so 64bit apps work > >> well and newly created files get 32bit inode numbers as inode64 is not > >> the default. > >> > >> Is there a way to find / get all 64bit inode files / dies and convert > >> them back to 32bit without a reinstall? > > > > On IRIX you could use xfs_reno to renumber those inodes. > > http://techpubs.sgi.com/library/tpl/cgi-bin/getdoc.cgi?coll=0650&db=man&fname=/usr/share/catman/a_man/cat1/xfs_reno.z > > > > xfs_reno was ported to linux in '07 and was most recently reposted by Jeff Liu: > > http://oss.sgi.com/archives/xfs/2012-11/msg00425.html > The old patch set was belong to the infrastructures of online shrinking > support. Recently, I realized that I have made a few mistakes in swap > inodes ioctl(2) implementation when I revisit the old patch set at: > http://oss.sgi.com/archives/xfs/2012-11/msg00414.html > > Since we have user request and this function is independent to the > shrinking feature, I'd like to work on it at first if you like it. That sounds good to me. xfs_reno is a worthwhile feature. Regards, Ben _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 7:22 ` Stefan Priebe - Profihost AG 2012-05-07 16:36 ` Stan Hoeppner @ 2012-05-07 23:42 ` Dave Chinner 1 sibling, 0 replies; 33+ messages in thread From: Dave Chinner @ 2012-05-07 23:42 UTC (permalink / raw) To: Stefan Priebe - Profihost AG; +Cc: stan, xfs On Mon, May 07, 2012 at 09:22:42AM +0200, Stefan Priebe - Profihost AG wrote: > > >> # vmstat > > "vmstat 5", not vmstat 5 times.... :/ > oh sorry. Sadly the rsync processes do not run right know i've to kill > them. Is the output still usable? > # vmstat 5 > procs -----------memory---------- ---swap-- -----io---- -system-- > ----cpu---- > r b swpd free buff cache si so bi bo in cs us sy > id wa > 0 1 0 5582136 48 5849956 0 0 176 394 34 54 1 > 16 82 1 > 0 1 0 5552180 48 5854280 0 0 2493 2496 3079 2172 1 > 4 86 9 > 3 2 0 5601308 48 5857672 0 0 1098 28043 5150 1913 0 > 10 73 17 > 0 2 0 5595360 48 5863180 0 0 1098 14336 3945 1897 0 > 8 69 22 > 3 2 0 5594088 48 5865280 0 0 432 15897 4209 2366 0 > 8 71 21 > 0 2 0 5591068 48 5868940 0 0 854 10989 3519 2107 0 > 7 70 23 > 1 1 0 5592004 48 5869872 0 0 180 7886 3605 2436 0 > 3 76 22 It tells me that there is still quite an IO load on the system even when the rsyncs are not running... > >> /dev/sdb1 4,6T 4,3T 310G 94% /mnt > > Well, you've probably badly fragmented the free space you have. what > > does the 'xfs_db -r -c freesp <dev>' command tell you? > > from to extents blocks pct > 1 1 942737 942737 0,87 > 2 3 671860 1590480 1,47 > 4 7 461268 2416025 2,23 > 8 15 1350517 18043063 16,67 > 16 31 111254 2547581 2,35 > 32 63 192032 9039799 8,35 So that's roughly 3.7 million free space extents of 256kB or less totalling about 32% of the freespace (~100GB). That's pretty badly fragmented, and given the workload, probably unrecoverable. Dump, mkfs and restore is probably the only way to unfragment the free space now, but that would only be a temporary solution if you continue to run at >90% full. Even if you do keep it at below 90% full, such a workload will age the filesystem and slowly fragment free space, but it should take a lot longer to get to this state... > >>>> /dev/sdb1 4875737052 4659318044 216419008 96% /mnt > >>> You have 4.6 *billion* inodes in your filesystem? > >> Yes - it backups around 100 servers with a lot of files. > i rechecked this and it seems i sadly copied the wrong output ;-( sorry > for that. > > Here is the correct one: > #~ df -i > /dev/sdb1 975173568 95212355 879961213 10% /mnt That makes more sense. :) Cheers, Dave. -- Dave Chinner david@fromorbit.com _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 6:39 ` Stefan Priebe - Profihost AG 2012-05-07 7:17 ` Dave Chinner @ 2012-05-07 8:21 ` Martin Steigerwald 2012-05-07 16:44 ` Stan Hoeppner 2012-05-07 8:31 ` Martin Steigerwald 2 siblings, 1 reply; 33+ messages in thread From: Martin Steigerwald @ 2012-05-07 8:21 UTC (permalink / raw) To: xfs; +Cc: stan, Stefan Priebe - Profihost AG Am Montag, 7. Mai 2012 schrieb Stefan Priebe - Profihost AG: > Hi, > > after deleting 400GB it was faster. Now there are still 300GB free but > it is slow as hell again ;-( Well, before you had 200 GB free, now you have 300 GB free. Thats not that much more. You have 5 AGs, so if distributed evenly it would be 40 to 60 GB free space per AG. The 10% Dave recommended of 4,6 TB would be 460 GB. Are other volumes of similar size are similarily full but faster? Would be interesting to see whether you are able to sustain faster speeds by keeping the volume above 500 GB free or so. > >> It is a raid 10 of 20 SATA Disks and i can only write to them with > >> about 700kb/s while doing random i/o. > > > > What sort of random IO? size, read, write, direct or buffered, data > > or metadata, etc? > > There are 4 rsync processes running and doing backups of other severs. Which rsync version is involved? I strongly suggest to use rsync version 3 or above. > > Your filesystem is near full - the allocation algorithms definitely > > slow down as you approach ENOSPC, and IO efficiency goes to hell > > because of a lack of contiguous free space to allocate from. > > I've now 94% used but it is still slow. It seems it was just getting > fast with more than 450GB free space. > > /dev/sdb1 4,6T 4,3T 310G 94% /mnt Then: Will it stay fast if you keep it above 450 GB free space? And is it as fast then as it used to be before you faced performance issues? Ciao, -- Martin 'Helios' Steigerwald - http://www.Lichtvoll.de GPG: 03B0 0D6C 0040 0710 4AFA B82F 991B EAAC A599 84C7 _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 8:21 ` Martin Steigerwald @ 2012-05-07 16:44 ` Stan Hoeppner 0 siblings, 0 replies; 33+ messages in thread From: Stan Hoeppner @ 2012-05-07 16:44 UTC (permalink / raw) To: Martin Steigerwald; +Cc: Stefan Priebe - Profihost AG, xfs On 5/7/2012 3:21 AM, Martin Steigerwald wrote: > Am Montag, 7. Mai 2012 schrieb Stefan Priebe - Profihost AG: >> after deleting 400GB it was faster. Now there are still 300GB free but >> it is slow as hell again ;-( > > Well, before you had 200 GB free, now you have 300 GB free. Thats not that > much more. You have 5 AGs, so if distributed evenly it would be 40 to 60 > GB free space per AG. > > The 10% Dave recommended of 4,6 TB would be 460 GB. As I mentioned, increasing the size of the free space pool will only help for a short duration. Most of the files Stefan is deleting are small files. Thus he's simply punching out small holes of free space, which is the problem he already has. Now if he could delete a handful of large files (10GB+) that are mostly contiguous now, that would help quite a bit, as the resulting free space wouldn't be horribly fragmented. So the takeaway is that simply deleting files to get more space isn't the right way to go. Deleting larger files that are currently contiguous is the way to go. At least for some temporary breathing room. The "permanent" solution I described in another reply. -- Stan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 6:39 ` Stefan Priebe - Profihost AG 2012-05-07 7:17 ` Dave Chinner 2012-05-07 8:21 ` Martin Steigerwald @ 2012-05-07 8:31 ` Martin Steigerwald 2012-05-07 13:57 ` Stefan Priebe - Profihost AG 2 siblings, 1 reply; 33+ messages in thread From: Martin Steigerwald @ 2012-05-07 8:31 UTC (permalink / raw) To: xfs; +Cc: stan, Stefan Priebe - Profihost AG Am Montag, 7. Mai 2012 schrieb Stefan Priebe - Profihost AG: > > iostat -x -d -m 5 and vmstat 5 traces would be > > useful to see if it is your array that is slow..... > > ~ # iostat -x -d -m 5 > Linux 2.6.40.28intel (server844-han) 05/07/2012 _x86_64_ > (8 CPU) > > Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s > avgrq-sz avgqu-sz await svctm %util > sdb 0,00 0,00 254,80 25,40 1,72 0,16 > 13,71 0,86 3,08 2,39 67,06 > sda 0,00 0,20 0,00 1,20 0,00 0,00 > 6,50 0,00 0,00 0,00 0,00 > > Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s > avgrq-sz avgqu-sz await svctm %util > sdb 0,00 0,00 187,40 24,20 1,26 0,19 > 14,05 0,75 3,56 3,33 70,50 > sda 0,00 0,00 0,00 0,40 0,00 0,00 > 4,50 0,00 0,00 0,00 0,00 > > Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s > avgrq-sz avgqu-sz await svctm %util > sdb 0,00 11,20 242,40 92,00 1,56 0,89 > 15,00 4,70 14,06 1,58 52,68 > sda 0,00 0,20 0,00 2,60 0,00 0,02 > 12,00 0,00 0,00 0,00 0,00 > > Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s > avgrq-sz avgqu-sz await svctm %util > sdb 0,00 0,00 166,20 24,00 0,99 0,17 > 12,51 0,57 3,02 2,40 45,56 > sda 0,00 0,00 0,00 0,00 0,00 0,00 > 0,00 0,00 0,00 0,00 0,00 > > qDevice: rrqm/s wrqm/s r/s w/s rMB/s wMB/s > avgrq-sz avgqu-sz await svctm %util > sdb 0,00 0,00 188,00 25,40 1,22 0,16 > 13,23 0,44 2,04 1,78 38,02 > sda 0,00 0,00 0,00 0,00 0,00 0,00 > 0,00 0,00 0,00 0,00 0,00 Disk utilization seems to be quite high but it seems there not near to 90 to 100%. So there might be other overheads involved - like network or (unlikely) CPU. Did you verify that at the time you perceive slowness the servers you backup can deliver data fast enough? I would like to now, whether there are really processes waiting for I/O during rsync workload. Can you try vmstat 5 and while true; do ps aux | grep " D" | grep -v grep ; sleep 1; done while the backup workload is running and slow? Like this: merkaba:~> while true; do ps aux | grep " D" | grep -v grep ; sleep 1; done root 1508 0.0 0.0 0 0 ? D Mai06 0:00 [flush-253:2] root 1508 0.0 0.0 0 0 ? D Mai06 0:00 [flush-253:2] martin 28374 100 0.0 9800 652 pts/7 D+ 10:27 0:02 dd if=/dev/zero of=fhgs root 1508 0.0 0.0 0 0 ? D Mai06 0:00 [flush-253:2] (this is with an Ext4, so its using the flush daemon, with XFS you probably see xfssyncd or xfsbufd instead if I am not mistaken and if rsync processes are waiting for I/O they should appear there too) And yes, its important to have vmstat 5 output during workload is happening to see the amount of CPU time that the kernel cannot use for processing cause all processes that are runnable wait for I/O. -- Martin 'Helios' Steigerwald - http://www.Lichtvoll.de GPG: 03B0 0D6C 0040 0710 4AFA B82F 991B EAAC A599 84C7 _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 8:31 ` Martin Steigerwald @ 2012-05-07 13:57 ` Stefan Priebe - Profihost AG 2012-05-07 14:32 ` Martin Steigerwald 0 siblings, 1 reply; 33+ messages in thread From: Stefan Priebe - Profihost AG @ 2012-05-07 13:57 UTC (permalink / raw) To: Martin Steigerwald; +Cc: stan, xfs Am 07.05.2012 10:31, schrieb Martin Steigerwald: > Am Montag, 7. Mai 2012 schrieb Stefan Priebe - Profihost AG: > Did you verify that at the time you perceive slowness the servers you > backup can deliver data fast enough? Yes. It works fine to another partition. > I would like to now, whether there are really processes waiting for I/O > during rsync workload. > > Can you try vmstat 5 and > while true; do ps aux | grep " D" | grep -v grep ; sleep 1; done Here it is: # vmstat 5 procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 0 1 0 396688 48 7728136 0 0 229 374 22 32 1 12 85 2 2 0 0 421744 48 7717348 0 0 3722 1169 9015 8389 1 2 95 2 0 1 0 405780 48 7751904 0 0 8937 207 10780 9290 2 2 88 7 0 0 0 486928 48 7733300 0 0 2526 1277 8275 7791 1 2 96 1 0 0 0 416692 48 7778164 0 0 5046 43750 8548 8141 1 2 95 2 0 0 0 444968 48 7777792 0 0 8021 1709 9315 8573 2 2 94 2 2 0 0 357924 48 7946532 0 0 48181 1031 17646 12684 10 4 87 0 1 0 0 348696 48 8137200 0 0 74366 1056 24362 16775 15 5 81 0 1 0 0 391552 48 8279000 0 0 54693 1242 19224 13957 11 4 85 0 # while true; do ps aux | grep " D" | grep -v grep ; sleep 1; done root 12493 2.0 0.2 101780 48392 ? D 15:35 0:24 rsync --daemon root 13581 3.5 0.1 76832 20828 ? D 15:50 0:10 rsync --daemon root 12493 2.0 0.2 101268 48508 ? D 15:35 0:24 rsync --daemon root 12494 3.9 0.2 128220 44328 ? D 15:35 0:47 rsync --daemon root 12493 2.0 0.2 101268 48508 ? D 15:35 0:24 rsync --daemon root 13581 3.5 0.1 76832 20828 ? D 15:50 0:10 rsync --daemon root 12493 2.0 0.2 101268 48508 ? D 15:35 0:24 rsync --daemon root 13581 3.5 0.1 76832 20828 ? D 15:50 0:10 rsync --daemon root 12493 2.0 0.2 86420 48120 ? D 15:35 0:25 rsync --daemon root 12493 2.0 0.2 86420 48120 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 21720 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86420 48120 ? D 15:35 0:25 rsync --daemon root 12493 2.0 0.2 86420 48120 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 21720 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86420 48120 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 21980 ? D 15:50 0:11 rsync --daemon root 13582 1.7 0.0 74308 13280 ? D 15:50 0:05 rsync --daemon root 12493 2.0 0.2 87956 48212 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 21980 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86420 48156 ? D 15:35 0:25 rsync --daemon root 12493 2.0 0.2 86420 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22244 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22244 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22244 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22244 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22244 ? D 15:50 0:11 rsync --daemon root 13581 3.5 0.1 76832 22508 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22508 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22508 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22772 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22772 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22772 ? D 15:50 0:11 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22772 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.5 0.1 76832 22772 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 3378 0.1 0.0 0 0 ? D 13:59 0:09 [flush-8:16] root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon root 12493 2.0 0.2 86676 48180 ? D 15:35 0:25 rsync --daemon root 13581 3.4 0.1 76832 23036 ? D 15:50 0:12 rsync --daemon Stefan _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: suddenly slow writes on XFS Filesystem 2012-05-07 13:57 ` Stefan Priebe - Profihost AG @ 2012-05-07 14:32 ` Martin Steigerwald 0 siblings, 0 replies; 33+ messages in thread From: Martin Steigerwald @ 2012-05-07 14:32 UTC (permalink / raw) To: Stefan Priebe - Profihost AG; +Cc: stan, xfs Am Montag, 7. Mai 2012 schrieb Stefan Priebe - Profihost AG: > Am 07.05.2012 10:31, schrieb Martin Steigerwald: > > Am Montag, 7. Mai 2012 schrieb Stefan Priebe - Profihost AG: > > Did you verify that at the time you perceive slowness the servers you > > backup can deliver data fast enough? > > Yes. It works fine to another partition. With more free space I suppose? > > > I would like to now, whether there are really processes waiting for > > I/O during rsync workload. > > > > Can you try vmstat 5 and > > while true; do ps aux | grep " D" | grep -v grep ; sleep 1; done > > Here it is: > # vmstat 5 > procs -----------memory---------- ---swap-- -----io---- -system-- > ----cpu---- > r b swpd free buff cache si so bi bo in cs us sy > id wa > 0 1 0 396688 48 7728136 0 0 229 374 22 32 1 > 12 85 2 > 2 0 0 421744 48 7717348 0 0 3722 1169 9015 8389 1 > 2 95 2 > 0 1 0 405780 48 7751904 0 0 8937 207 10780 9290 2 > 2 88 7 > 0 0 0 486928 48 7733300 0 0 2526 1277 8275 7791 1 > 2 96 1 > 0 0 0 416692 48 7778164 0 0 5046 43750 8548 8141 1 > 2 95 2 > 0 0 0 444968 48 7777792 0 0 8021 1709 9315 8573 2 > 2 94 2 > 2 0 0 357924 48 7946532 0 0 48181 1031 17646 12684 10 > 4 87 0 > 1 0 0 348696 48 8137200 0 0 74366 1056 24362 16775 15 > 5 81 0 > 1 0 0 391552 48 8279000 0 0 54693 1242 19224 13957 11 > 4 85 0 Hmm, there are some I/O wait times, but they are lower than I expected. In the last lines the throughput is higher than in the first lines. Does that correlate to higher speed in rsync? It may also be excessive writes due to free space fragmentation, but Dave or someone else might be able to say whether such a huge difference can be blamed to free space fragmentation alone. > # while true; do ps aux | grep " D" | grep -v grep ; sleep 1; done > root 12493 2.0 0.2 101780 48392 ? D 15:35 0:24 rsync > --daemon > root 13581 3.5 0.1 76832 20828 ? D 15:50 0:10 rsync > --daemon > root 12493 2.0 0.2 101268 48508 ? D 15:35 0:24 rsync > --daemon > root 12494 3.9 0.2 128220 44328 ? D 15:35 0:47 rsync > --daemon > root 12493 2.0 0.2 101268 48508 ? D 15:35 0:24 rsync > --daemon Still these rsyncs appear to be in uninteruptible sleep quite consistently. They shouldn´t be in that state when waiting on network socket readyness. So this workloads seems to be I/O bound to me. I suggest trying to keep the volume above 500 GB free and see whether that works consistently. Thanks, -- Martin 'Helios' Steigerwald - http://www.Lichtvoll.de GPG: 03B0 0D6C 0040 0710 4AFA B82F 991B EAAC A599 84C7 _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 33+ messages in thread
end of thread, other threads:[~2013-02-19 19:11 UTC | newest] Thread overview: 33+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2012-05-06 9:01 suddenly slow writes on XFS Filesystem Stefan Priebe 2012-05-06 10:31 ` Martin Steigerwald 2012-05-06 10:33 ` Martin Steigerwald 2012-05-06 15:45 ` Stan Hoeppner 2012-05-06 19:25 ` Stefan Priebe 2012-05-07 1:39 ` Dave Chinner 2012-05-06 21:43 ` Martin Steigerwald 2012-05-07 6:40 ` Stefan Priebe - Profihost AG 2012-05-07 1:34 ` Dave Chinner 2012-05-07 6:39 ` Stefan Priebe - Profihost AG 2012-05-07 7:17 ` Dave Chinner 2012-05-07 7:22 ` Stefan Priebe - Profihost AG 2012-05-07 16:36 ` Stan Hoeppner 2012-05-07 19:08 ` Martin Steigerwald 2012-05-07 20:05 ` Stefan Priebe 2012-05-09 6:57 ` Stan Hoeppner 2012-05-09 7:04 ` Dave Chinner 2012-05-09 7:36 ` Stefan Priebe - Profihost AG 2012-05-09 7:49 ` Stan Hoeppner 2013-02-15 15:06 ` 32bit apps and inode64 Stefan Priebe - Profihost AG 2013-02-15 21:46 ` Ben Myers 2013-02-16 10:24 ` Stefan Priebe - Profihost AG 2013-02-17 21:33 ` Dave Chinner 2013-02-18 8:12 ` Stefan Priebe - Profihost AG 2013-02-18 22:06 ` Dave Chinner 2013-02-17 8:13 ` Jeff Liu 2013-02-19 19:11 ` Ben Myers 2012-05-07 23:42 ` suddenly slow writes on XFS Filesystem Dave Chinner 2012-05-07 8:21 ` Martin Steigerwald 2012-05-07 16:44 ` Stan Hoeppner 2012-05-07 8:31 ` Martin Steigerwald 2012-05-07 13:57 ` Stefan Priebe - Profihost AG 2012-05-07 14:32 ` Martin Steigerwald
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.