NAS540: Volume down, no option to repair
All Replies
-
From this information that volume seems healthy.The raidarray md4 is build from sdc3, and is up.Yet there is something strange with md0 and md1 (firmware and swap). Normally a 4 disk ZyXEL nas will create a 4 member raid1 array for these volumes, and the 4 disks have roles 0,1,2 and 3 in that array.But here we have role 0,1 and 4:
md0 : active raid1 sda1[0] sdb1[1] sdc1[4]
*Maybe* the firmware expects role 3 here, and is whining because it can't find that.It is possible to repair that manually.su mdadm /dev/md0 -fail /dev/sdc1 --remove /dev/sdc1 mdadm /dev/md1 --fail /dev/sdc2 --remove /dev/sdc2 mdadm /dev/md0 --fail detached --remove detached mdadm /dev/md1 --fail detached --remove detached
After this both md0 and md1 should be a 2 member raid1 array, something likemd0 : active raid1 sda1[0] sdb1[1] 1997760 blocks super 1.2 [2/2] [UU]
Now you can add sdc again:mdadm /dev/md0 --add /dev/sdc1 mdadm /dev/md1 --add /dev/sdc2
I expect now a 3 disk array:md0 : active raid1 sda1[0] sdb1[1] sdc1[2] 1997760 blocks super 1.2 [3/3] [UUU]
And finally you can add a (missing) 4th membermdadm /dev/md0 --add missing mdadm /dev/md1 --add missing
And I expect:md0 : active raid1 sda1[0] sdb1[1] sdc1[2] 1997760 blocks super 1.2 [4/3] [UUU_]
Don't know if that helps, though.1 -
And just because I'm a complete noob when it comes to stuff like this; will this touch the data on the disks in slot 1 and 2? I'm guessing not as I presume this will rebuild how the NAS sees drives and doesn't touch the actual drives, but I wanted to confirm before I do anything0
-
Indeed these commands are not supposed to do anything with the data in slot1&2.That data is in the raid array md2 and md3, which use the partitions sda3 and sdb3. These are not touched by the commands.1
-
Many thanks for your continued help.I started to run the commands, but are getting stuck on readding sdc. When I runI get this message:
mdadm /dev/md0 --add /dev/sdc1
mdadm: Cannot open /dev/sdc1: Device or resource busy
I'm finding quite a few results with this error on Google, but am not really versed enough in Linux or raid setup to make head or tail of the results, and don't really want to enter commands willy nilly.Worth noting is also that I got a reply from Zyxel support and was told that they would see if they could repro and/or investigate it on Monday.0 -
Did 'get /proc/mdstat' show the expected output? Strange. What keeps sdc1 busy? Unless it is automagically re-added to the array by the firmware after you removed it, I wouldn't know.1
-
Hmm, most have been some temporary thing. I was away for a day and retried your suggested commands today, and now everything worked. The only thing I got an error on was
mdadm /dev/md0 --add missing
where I got a message sayingmdadm: 'missing' only meaningful with --re-add
which I tried instead, and gave no error.Here's my whole input log just for documentations sake:~ # mdadm /dev/md0 --fail /dev/sdc1 --remove /dev/sdc1<br>mdadm: set /dev/sdc1 faulty in /dev/md0<br>mdadm: hot removed /dev/sdc1 from /dev/md0<br>~ # mdadm /dev/md1 --fail /dev/sdc2 --remove /dev/sdc2<br>mdadm: set /dev/sdc2 faulty in /dev/md1<br>mdadm: hot removed /dev/sdc2 from /dev/md1<br>~ # mdadm /dev/md0 --fail detached --remove detached<br>~ # mdadm /dev/md1 --fail detached --remove detached<br>~ # mdadm /dev/md0 --add /dev/sdc1<br>mdadm: added /dev/sdc1<br>~ # mdadm /dev/md1 --add /dev/sdc2<br>mdadm: added /dev/sdc2<br>~ # mdadm /dev/md0 --add missing<br>mdadm: 'missing' only meaningful with --re-add<br>~ # mdadm /dev/md0 --re-add missing<br>~ # mdadm /dev/md1 --re-add missing
And here's the output from mdstat and partitions again:~ # cat /proc/mdstat<br>Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]<br>md4 : active raid1 sdc3[0]<br> 240067392 blocks super 1.2 [1/1] [U]<br><br>md3 : active raid1 sdb3[0]<br> 3902886720 blocks super 1.2 [1/1] [U]<br><br>md2 : active raid1 sda3[0]<br> 3902886720 blocks super 1.2 [1/1] [U]<br><br>md1 : active raid1 sdc2[4] sda2[0] sdb2[1]<br> 1998784 blocks super 1.2 [4/3] [UU_U]<br><br>md0 : active raid1 sdc1[4] sda1[0] sdb1[1]<br> 1997760 blocks super 1.2 [4/3] [UU_U]<br><br>unused devices: <none><br>~ # cat /proc/partitions<br>major minor #blocks name<br><br> 7 0 147456 loop0<br> 31 0 256 mtdblock0<br> 31 1 512 mtdblock1<br> 31 2 256 mtdblock2<br> 31 3 10240 mtdblock3<br> 31 4 10240 mtdblock4<br> 31 5 112640 mtdblock5<br> 31 6 10240 mtdblock6<br> 31 7 112640 mtdblock7<br> 31 8 6144 mtdblock8<br> 8 0 3907018584 sda<br> 8 1 1998848 sda1<br> 8 2 1999872 sda2<br> 8 3 3903017984 sda3<br> 8 16 3907018584 sdb<br> 8 17 1998848 sdb1<br> 8 18 1999872 sdb2<br> 8 19 3903017984 sdb3<br> 8 32 244198584 sdc<br> 8 33 1998848 sdc1<br> 8 34 1999872 sdc2<br> 8 35 240198656 sdc3<br> 8 48 244198584 sdd<br> 8 49 1998848 sdd1<br> 8 50 1999872 sdd2<br> 31 9 102424 mtdblock9<br> 9 0 1997760 md0<br> 9 1 1998784 md1<br> 31 10 4464 mtdblock10<br> 9 2 3902886720 md2<br> 9 3 3902886720 md3<br> 9 4 240067392 md4<br>~ #<br>
which looks like it has the same output as in my previous post a couple of days ago, and I still get the volume down message on disk 3 when I log into the web UI.Not sure if it makes any sense, but again, thanks for your help0 -
I did some more testing after a promt from Zyxel support. It seems like I get the volume down error on a new volume whenever I already have a volume created.1 disk in any slot with a basic volume -> No error2 disks in any slots with a raid1 volume -> No error2 or more disks in any slots with a basic volume each -> Error on the volume that was created lastI also tried to do another factoyr reset, as well as reinstalling the firmware, and I still get errors on the above pattern.0
-
SOP?Anyway, this forum is falling apart. Some of the listings which were there are lost.Can you enable the ssh server, login over ssh as root (admin password), and post the output ofcat /proc/mdstatmdadm --examine /dev/sd[abcd]30
-
# cat /proc/mdstatPersonalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]md2 : inactive sda3[0](S) sdd3[4](S) sdc3[2](S)11708660736 blocks super 1.2md1 : active raid1 sdb2[5](F) sda2[0] sdd2[4] sdc2[2]1998784 blocks super 1.2 [4/3] [U_UU]md0 : active raid1 sdb1[5](F) sda1[0] sdd1[4] sdc1[2]1997760 blocks super 1.2 [4/3] [U_UU]unused devices: <none># mdadm --examine /dev/sd[abcd]3/dev/sda3:Magic : a92b4efcVersion : 1.2Feature Map : 0x0Array UUID : 14e886d5:f3e1b36a:b3cd558d:574f21a2Name : NAS542:2 (local to host NAS542)Creation Time : Fri Sep 20 13:17:43 2019Raid Level : raid5Raid Devices : 4Avail Dev Size : 7805773824 (3722.08 GiB 3996.56 GB)Array Size : 11708660736 (11166.25 GiB 11989.67 GB)Data Offset : 262144 sectorsSuper Offset : 8 sectorsState : cleanDevice UUID : 3fe9b787:1a0fa640:04a63619:8a65fb5bUpdate Time : Thu Nov 21 22:12:54 2019Checksum : 1b9dbffb - correctEvents : 100Layout : left-symmetricChunk Size : 64KDevice Role : Active device 0Array State : A.A. ('A' == active, '.' == missing)mdadm: No md superblock detected on /dev/sdb3./dev/sdc3:Magic : a92b4efcVersion : 1.2Feature Map : 0x0Array UUID : 14e886d5:f3e1b36a:b3cd558d:574f21a2Name : NAS542:2 (local to host NAS542)Creation Time : Fri Sep 20 13:17:43 2019Raid Level : raid5Raid Devices : 4Avail Dev Size : 7805773824 (3722.08 GiB 3996.56 GB)Array Size : 11708660736 (11166.25 GiB 11989.67 GB)Data Offset : 262144 sectorsSuper Offset : 8 sectorsState : cleanDevice UUID : 930a65f9:196ec7bc:dc154c61:76ab86eaUpdate Time : Thu Nov 21 22:12:54 2019Checksum : e00cf615 - correctEvents : 100Layout : left-symmetricChunk Size : 64KDevice Role : Active device 2Array State : A.A. ('A' == active, '.' == missing)/dev/sdd3:Magic : a92b4efcVersion : 1.2Feature Map : 0x0Array UUID : 14e886d5:f3e1b36a:b3cd558d:574f21a2Name : NAS542:2 (local to host NAS542)Creation Time : Fri Sep 20 13:17:43 2019Raid Level : raid5Raid Devices : 4Avail Dev Size : 7805773824 (3722.08 GiB 3996.56 GB)Array Size : 11708660736 (11166.25 GiB 11989.67 GB)Data Offset : 262144 sectorsSuper Offset : 8 sectorsState : cleanDevice UUID : ccf2036a:a7ac4e63:9aa38f9d:165784aaUpdate Time : Thu Nov 21 22:12:54 2019Checksum : f374563b - correctEvents : 100Layout : left-symmetricChunk Size : 64KDevice Role : spareArray State : A.A. ('A' == active, '.' == missing)0
Categories
- All Categories
- 415 Beta Program
- 2.4K Nebula
- 144 Nebula Ideas
- 94 Nebula Status and Incidents
- 5.6K Security
- 238 USG FLEX H Series
- 267 Security Ideas
- 1.4K Switch
- 71 Switch Ideas
- 1.1K Wireless
- 40 Wireless Ideas
- 6.3K Consumer Product
- 247 Service & License
- 384 News and Release
- 83 Security Advisories
- 29 Education Center
- 10 [Campaign] Zyxel Network Detective
- 3.2K FAQ
- 34 Documents
- 34 Nebula Monthly Express
- 83 About Community
- 71 Security Highlight