From mboxrd@z Thu Jan 1 00:00:00 1970 From: Joe Landman Subject: Re: hung grow Date: Wed, 4 Oct 2017 14:29:10 -0400 Message-ID: <0001704a-fe2f-e164-7694-f294a427ed83@gmail.com> References: Mime-Version: 1.0 Content-Type: text/plain; charset=utf-8; format=flowed Content-Transfer-Encoding: 8bit Return-path: In-Reply-To: Content-Language: en-US Sender: linux-raid-owner@vger.kernel.org To: Curt , Anthony Youngman Cc: linux-raid@vger.kernel.org List-Id: linux-raid.ids On 10/04/2017 02:16 PM, Curt wrote: > Hi, > > I was reading this one https://raid.wiki.kernel.org/index.php/RAID_Recovery > > I don't have any spare bays on that server...I'd have to make a trip > to my datacenter and bring the drives back to my house. The bad thing > is the 2 drives I replaced, failed a while ago, so they were behind. > I was hoping I could still use the 4 drives I had before I did a grow > on them. Do they need to be up-to-date or do I just need the config > from them to recover the 3 drives that were still good? > > Oh, I originally started with 7, 2 failed a few moths back and the 3rd > one just recently. FML Er ... honestly, I hope you have a backup. If the drives are really dead, and can't be seen with lsscsi or cat /proc/scsi/scsi , then your raid is probably gone. If they can be seen, the ddrescue is your best option right now. Do not grow the system.  Stop that.  Do nothing that changes metadata. You may (remotely possibly) recover if you can copy the "dead" drives to two new live ones. > > Cheers, > Curt > > On Wed, Oct 4, 2017 at 1:51 PM, Anthony Youngman > wrote: >> On 04/10/17 18:18, Curt wrote: >>> Is my raid completely fucked or can I still recover some data with >>> doing the create assume clean? >> >> PLEASE PLEASE PLEASE DON'T !!!!!! >> >> I take it you haven't read the raid wiki? >> >> https://raid.wiki.kernel.org/index.php/Linux_Raid#When_Things_Go_Wrogn >> >> The bad news is your array is well borked. The good news is I don't think >> you have - YET - managed to bork it irretrievably. A create will almost >> certainly trash it beyond recovery!!! >> >> I think we can stop/revert the grow, and get the array back to a usable >> state, where we can force an assemble. If a bit of data gets lost, sorry. >> >> Do you have spare SATA ports? So you have the bad drives you replaced (can >> you ddrescue them on to new drives?). What was the original configuration of >> the raid - you say you lost three drives, but how many did you have to start >> with? >> >> I'll let the experts talk you through the actual recovery, but the steps >> need to be to revert the grow, ddrescue the best of your failed drives, >> force an assembly, and then replace the other two failed drives. No >> guarantees as to how much data will be left at the end, although hopefully >> we'll save most of it. >> >> Cheers, >> Wol > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Joe Landman e: joe.landman@gmail.com t: @hpcjoe w: https://scalability.org g: https://github.com/joelandman