linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: David T-G <davidtg-robot@justpickone.org>
To: Linux RAID list <linux-raid@vger.kernel.org>
Subject: timing (was "Re: all of my drives are spares")
Date: Sun, 10 Sep 2023 03:44:28 +0000	[thread overview]
Message-ID: <20230910034427.GF1085@jpo> (raw)
In-Reply-To: <20230910025554.GD1085@jpo>

One more time this evening ...

...and then David T-G home said...
% 
% ...and then Wol said...
...
% % that's just struck me, this is often caused by a drive failing some while
% % back, and then a glitch on a second drive brings the whole thing down. When
% % did you last check your array was fully functional?
% 
% Let me get back to you on that.  It's actually been a couple of weeks in
% this state just waiting to get to it; life has been interesting here,
[snip]

Apparently less than a couple of weeks after all.  That's what I get for
not knowing where I'll sleep each night and losing track of the days as a
result...

Anyway, here are a couple of clips from 08/29:

  ######################################################################
   02:55:01  up  22:42,  0 users,  load average: 6.41, 6.73, 6.49
  Personalities : [raid1] [raid6] [raid5] [raid4] [linear]
  md50 : active linear md52[1] md54[3] md56[5] md51[0] md53[2] md55[4]
	29289848832 blocks super 1.2 0k rounding

  md4 : active raid1 sde4[0] sda4[2]
	142972224 blocks super 1.2 [2/2] [UU]
	bitmap: 1/2 pages [4KB], 65536KB chunk

  md3 : active raid1 sde3[0] sda3[2]
	35617792 blocks super 1.2 [2/2] [UU]

  md56 : active raid5 sdd56[3] sdc56[1] sdb56[0] sdf56[5] sdl56[4] sdk56[6]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/6] [UUUUUU]
	  resync=DELAYED
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md55 : active raid5 sdd55[3] sdc55[1] sdb55[0] sdf55[5] sdl55[4] sdk55[6]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/6] [UUUUUU]
	  resync=DELAYED
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md53 : active raid5 sdd53[3] sdc53[1] sdb53[0] sdf53[5] sdl53[4] sdk53[6]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/6] [UUUUUU]
	[>....................]  check =  4.2% (69399936/1627261952) finish=7901.9min speed=3285K/sec
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md51 : active raid5 sdb51[0] sdd51[3] sdc51[1] sdf51[5] sdl51[4] sdk51[6]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/6] [UUUUUU]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md54 : active raid5 sdd54[3] sdc54[1] sdb54[0] sdf54[5] sdl54[4] sdk54[6]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/6] [UUUUUU]
	  resync=DELAYED
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md52 : active raid5 sdd52[3] sdc52[1] sdb52[0] sdf52[5] sdl52[4] sdk52[6]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/6] [UUUUUU]
	  resync=DELAYED
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md2 : active raid1 sde2[0] sda2[2]
	35617792 blocks super 1.2 [2/2] [UU]

  md1 : active raid1 sde1[0] sda1[2]
	35609600 blocks super 1.2 [2/2] [UU]

  unused devices: <none>
  ...
  ######################################################################
   03:00:01  up  22:47,  0 users,  load average: 3.75, 5.86, 6.28
  Personalities : [raid1] [raid6] [raid5] [raid4] [linear]
  md50 : active linear md52[1] md54[3] md56[5] md51[0] md53[2] md55[4]
	29289848832 blocks super 1.2 0k rounding

  md4 : active raid1 sde4[0] sda4[2]
	142972224 blocks super 1.2 [2/2] [UU]
	bitmap: 1/2 pages [4KB], 65536KB chunk

  md3 : active raid1 sde3[0] sda3[2]
	35617792 blocks super 1.2 [2/2] [UU]

  md56 : active raid5 sdd56[3] sdc56[1] sdb56[0] sdf56[5] sdl56[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md55 : active raid5 sdd55[3] sdc55[1] sdb55[0] sdf55[5] sdl55[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md53 : active raid5 sdd53[3] sdc53[1] sdb53[0] sdf53[5] sdl53[4] sdk53[6](F)
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	[>....................]  check =  4.2% (69789932/1627261952) finish=17276.4min speed=1502K/sec
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md51 : active raid5 sdb51[0] sdd51[3] sdc51[1] sdf51[5] sdl51[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md54 : active raid5 sdd54[3] sdc54[1] sdb54[0] sdf54[5] sdl54[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md52 : active raid5 sdd52[3] sdc52[1] sdb52[0] sdf52[5] sdl52[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md2 : active raid1 sde2[0] sda2[2]
	35617792 blocks super 1.2 [2/2] [UU]

  md1 : active raid1 sde1[0] sda1[2]
	35609600 blocks super 1.2 [2/2] [UU]

  unused devices: <none>
  /dev/sdn: Avolusion PRO-5X:  drive supported, but it doesn't have a temperature sensor.
  /dev/sdo: Seagate BUP BL:  drive supported, but it doesn't have a temperature sensor.
  /dev/sda: SATA SSD: 33302260C
  /dev/sdb: TOSHIBA HDWR11A: 43302260C
  /dev/sdc: TOSHIBA HDWR11A: 41302260C
  /dev/sdd: TOSHIBA HDWR11A: 42302260C
  /dev/sde: SATA SSD: 33302260C
  /dev/sdf: : S.M.A.R.T. not available
  /dev/sdg: : S.M.A.R.T. not available
  /dev/sdh: : S.M.A.R.T. not available
  /dev/sdp: WDC WD2500BEKT-75A25T0: S.M.A.R.T. not available
  /dev/sdq: WDC WD3200BEVT-60ZCT0: S.M.A.R.T. not available
  /dev/sdr: WD easystore 25FB: S.M.A.R.T. not available
  /dev/sds: WD easystore 264D: S.M.A.R.T. not available
  /dev/sdt: ST9120822A: S.M.A.R.T. not available
  /dev/sdu: WD Elements 25A3: S.M.A.R.T. not available

That's where sdk, a brand new EXOS 20T drive, apparently keeled over.
Hmmmmm.  Notice the temps check display; half of the SATA expansion card
(sdf - sdm) is missing.  Ouch.

Things ran fine like that for a day, until early on 08/30 we seem to
have keeled over.

  ######################################################################
   01:10:01  up 1 day 20:57,  0 users,  load average: 2.31, 2.11, 1.12
  Personalities : [raid1] [raid6] [raid5] [raid4] [linear]
  md50 : active linear md52[1] md54[3] md56[5] md51[0] md53[2] md55[4]
	29289848832 blocks super 1.2 0k rounding

  md4 : active raid1 sde4[0] sda4[2]
	142972224 blocks super 1.2 [2/2] [UU]
	bitmap: 0/2 pages [0KB], 65536KB chunk

  md3 : active raid1 sde3[0] sda3[2]
	35617792 blocks super 1.2 [2/2] [UU]

  md56 : active raid5 sdd56[3] sdc56[1] sdb56[0] sdf56[5] sdl56[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md55 : active raid5 sdd55[3] sdc55[1] sdb55[0] sdf55[5] sdl55[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md53 : active raid5 sdd53[3] sdc53[1] sdb53[0] sdf53[5] sdl53[4] sdk53[6](F)
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md51 : active raid5 sdb51[0] sdd51[3] sdc51[1] sdf51[5] sdl51[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md54 : active raid5 sdd54[3] sdc54[1] sdb54[0] sdf54[5] sdl54[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 1/13 pages [4KB], 65536KB chunk

  md52 : active raid5 sdd52[3] sdc52[1] sdb52[0] sdf52[5] sdl52[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [UUUU_U]
	bitmap: 0/13 pages [0KB], 65536KB chunk

  md2 : active raid1 sde2[0] sda2[2]
	35617792 blocks super 1.2 [2/2] [UU]

  md1 : active raid1 sde1[0] sda1[2]
	35609600 blocks super 1.2 [2/2] [UU]

  unused devices: <none>
  ...
  ######################################################################
   01:15:02  up 1 day 21:02,  0 users,  load average: 0.16, 0.84, 0.84
  Personalities : [raid1] [raid6] [raid5] [raid4] [linear]
  md50 : active linear md52[1] md54[3] md56[5] md51[0] md53[2] md55[4]
	29289848832 blocks super 1.2 0k rounding

  md4 : active raid1 sde4[0] sda4[2]
	142972224 blocks super 1.2 [2/2] [UU]
	bitmap: 0/2 pages [0KB], 65536KB chunk

  md3 : active raid1 sde3[0] sda3[2]
	35617792 blocks super 1.2 [2/2] [UU]

  md56 : active raid5 sdd56[3] sdc56[1] sdb56[0] sdf56[5] sdl56[4]
	8136309760 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/5] [

We came up again at 0240 the next day (08/31), and everything was a
spare.

  ######################################################################
   02:40:01  up   0:18,  27 users,  load average: 0.00, 0.03, 0.10
  Personalities : [raid1]
  md4 : active raid1 sde4[0] sda4[2]
	142972224 blocks super 1.2 [2/2] [UU]
	bitmap: 0/2 pages [0KB], 65536KB chunk

  md3 : active raid1 sde3[0] sda3[2]
	35617792 blocks super 1.2 [2/2] [UU]

  md55 : inactive sdd55[3](S) sdc55[1](S) sdb55[0](S) sdl55[4](S) sdk55[6](S) sdf55[5](S)
	9763571712 blocks super 1.2

  md56 : inactive sdd56[3](S) sdb56[0](S) sdc56[1](S) sdl56[4](S) sdk56[6](S) sdf56[5](S)
	9763571712 blocks super 1.2

  md53 : inactive sdd53[3](S) sdc53[1](S) sdb53[0](S) sdl53[4](S) sdk53[6](S) sdf53[5](S)
	9763571712 blocks super 1.2

  md54 : inactive sdd54[3](S) sdc54[1](S) sdb54[0](S) sdl54[4](S) sdk54[6](S) sdf54[5](S)
	9763571712 blocks super 1.2

  md52 : inactive sdd52[3](S) sdb52[0](S) sdc52[1](S) sdl52[4](S) sdk52[6](S) sdf52[5](S)
	9763571712 blocks super 1.2

  md51 : inactive sdd51[3](S) sdb51[0](S) sdc51[1](S) sdl51[4](S) sdk51[6](S) sdf51[5](S)
	9763571712 blocks super 1.2

  md1 : active raid1 sde1[0] sda1[2]
	35609600 blocks super 1.2 [2/2] [UU]

  md2 : active raid1 sde2[0] sda2[2]
	35617792 blocks super 1.2 [2/2] [UU]

  unused devices: <none>
  /dev/sdb: TOSHIBA HDWR11A: drive is sleeping
  /dev/sdc: TOSHIBA HDWR11A: drive is sleeping
  /dev/sdd: TOSHIBA HDWR11A: drive is sleeping
  /dev/sdf: ST20000NM007D-3DJ103: drive is sleeping
  /dev/sdk: ST20000NM007D-3DJ103: drive is sleeping
  /dev/sdl: TOSHIBA HDWR11A: drive is sleeping
  /dev/sdn: Avolusion PRO-5X:  drive supported, but it doesn't have a temperature sensor.
  /dev/sdo: Seagate BUP BL:  drive supported, but it doesn't have a temperature sensor.
  /dev/sda: SATA SSD: 33302260C
  /dev/sde: SATA SSD: 33302260C
  /dev/sdg: WDC WD7500BPKX-75HPJT0: 31302260C
  /dev/sdh: TOSHIBA MQ01ABD064: 33302260C
  /dev/sdi: ST3500413AS: 37302260C
  /dev/sdj: TOSHIBA MQ01ABD100: 33302260C
  /dev/sdm: Hitachi HDE721010SLA330: 43302260C
  /dev/sdp: WDC WD2500BEKT-75A25T0: S.M.A.R.T. not available
  /dev/sdq: WDC WD3200BEVT-60ZCT0: S.M.A.R.T. not available
  /dev/sdr: ST9120822A: S.M.A.R.T. not available
  /dev/sds: WD Elements 25A3: S.M.A.R.T. not available
  /dev/sdt: WD easystore 264D: S.M.A.R.T. not available
  /dev/sdu: WD easystore 25FB: S.M.A.R.T. not available

The whole SATA card is present, too; yay.  So rebooting helps.  But ...
Now I'm not sure how to get back to reassembly.


Thanks again and good night to all

:-D
-- 
David T-G
See http://justpickone.org/davidtg/email/
See http://justpickone.org/davidtg/tofu.txt


      parent reply	other threads:[~2023-09-10  3:44 UTC|newest]

Thread overview: 7+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-09-08  2:50 all of my drives are spares David T-G
2023-09-09 11:26 ` David T-G
2023-09-09 18:28   ` Wol
2023-09-10  2:55     ` David T-G
2023-09-10  3:11       ` assemble didn't quite (was "Re: all of my drives are spares") David T-G
2023-09-14 15:59         ` assemble didn't quite David T-G
2023-09-10  3:44       ` David T-G [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20230910034427.GF1085@jpo \
    --to=davidtg-robot@justpickone.org \
    --cc=linux-raid@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).