NSA310 - eSata disk status is DOWN
csipetcsapat
Posts: 3
Hello,
I have a NSA310 with 2TB internal HDD and 4TB external HDD connected to esata.
After a power failure, the external disk status was changed to Down and it is not shared anymore (the internal HDD is working).
The S.M.A.R.T status shows no errors.I tried to fix it via the admin web UI, but "Scan Volume" returns a "Failed to scan volume." error.
I have tested it inPC with Linux (using commends from https://kb.zyxel.com/KB/searchArticle!gwsViewDetail.action?articleOid=013308&lang=EN ), I can mount it and the file system & files are not corrupted.
I connect to NAS with putty and the disk looks good:
root@nsa310:~# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1]
md1 : active linear sdb2[0]
3906516992 blocks super 1.2 0k rounding
md0 : active linear sda2[0]
1952997888 blocks 64k rounding
unused devices: <none>
Personalities : [linear] [raid0] [raid1]
md1 : active linear sdb2[0]
3906516992 blocks super 1.2 0k rounding
md0 : active linear sda2[0]
1952997888 blocks 64k rounding
unused devices: <none>
e2fsck founds some incorrect nodes, but it fixes the errors, after that the disk is good:
root@nsa310:~# umount /mnt/extdisk/
root@nsa310:~# e2fsck.new /dev/md1
e2fsck 1.41.14 (22-Dec-2010)
/dev/md1: clean, 124570/244162560 files, 654272573/976629248 blocks
root@nsa310:~# e2fsck.new /dev/md1
e2fsck 1.41.14 (22-Dec-2010)
/dev/md1: clean, 124570/244162560 files, 654272573/976629248 blocks
If I mount it manually, the disk is mounted correctly:
root@nsa310:~# mkdir /mnt/extdisk
root@nsa310:~# mount /dev/md1 /mnt/extdisk/
root@nsa310:~# ls -l /mnt/extdisk/
total 52
drwxrwsrwx 17 admin root 4096 May 31 2021 DATA
drwxrwxrwx 5 root root 4096 Jul 29 2013 admin
-rw------- 1 root root 7168 Dec 11 21:30 aquota.user
drwxrwsrwx 2 root root 16384 Feb 26 2020 lost+found
drwxrwxrwx 2 root root 4096 Feb 26 2020 music
drwxrwxrwx 3 root root 4096 Sep 6 2020 nfs
drwxrwxrwx 2 root root 4096 Feb 26 2020 photo
drwxrwxrwx 2 root root 4096 Feb 26 2020 public
drwxrwxrwx 2 root root 4096 Feb 26 2020 video
root@nsa310:~# mount /dev/md1 /mnt/extdisk/
root@nsa310:~# ls -l /mnt/extdisk/
total 52
drwxrwsrwx 17 admin root 4096 May 31 2021 DATA
drwxrwxrwx 5 root root 4096 Jul 29 2013 admin
-rw------- 1 root root 7168 Dec 11 21:30 aquota.user
drwxrwsrwx 2 root root 16384 Feb 26 2020 lost+found
drwxrwxrwx 2 root root 4096 Feb 26 2020 music
drwxrwxrwx 3 root root 4096 Sep 6 2020 nfs
drwxrwxrwx 2 root root 4096 Feb 26 2020 photo
drwxrwxrwx 2 root root 4096 Feb 26 2020 public
drwxrwxrwx 2 root root 4096 Feb 26 2020 video
How can I switch back the 2nd HDD to Healthy state?
0
All Replies
-
You can try the Telnet Backdoor and run dmesg on the NAS to find out why the disk isn't mounted.
0 -
The dmesg shows no error:root@nsa310:~# dmesg
Marvell USB EHCI Host controller #0: c8040740
PEX0 interface detected Link X1
pci 0000:00:01.0: reg 10 io port: [0xf0000000-0xf00000ff]
pci 0000:00:01.0: reg 18 64bit mmio: [0x40000000-0x40000fff]
pci 0000:00:01.0: reg 20 64bit mmio: [0x40004000-0x40007fff]
pci 0000:00:01.0: reg 30 32bit mmio: [0x000000-0x00ffff]
pci 0000:00:01.0: supports D1 D2
pci 0000:00:01.0: PME# supported from D0 D1 D2 D3hot D3cold
pci 0000:00:01.0: PME# disabled
PCI: bus0: Fast back to back transfers disabled
mvPexLocalBusNumSet: ERR. Invalid PEX interface 1
bio: create slab <bio-0> at 0
SCSI subsystem initialized
usbcore: registered new interface driver usbfs
usbcore: registered new interface driver hub
usbcore: registered new device driver usb
cfg80211: Calling CRDA to update world regulatory domain
NET: Registered protocol family 2
IP route cache hash table entries: 2048 (order: 1, 8192 bytes)
TCP established hash table entries: 8192 (order: 4, 65536 bytes)
TCP bind hash table entries: 8192 (order: 3, 32768 bytes)
TCP: Hash tables configured (established 8192 bind 8192)
TCP reno registered
NET: Registered protocol family 1
rtc mv_rtc: rtc core: registered kw-rtc as rtc0
RTC registered
cpufreq: Init kirkwood cpufreq driver
cpufreq: High frequency: 1200000KHz - Low frequency: 400000KHz
cpufreq: Setting CPU Frequency to 1200000 KHz
cpufreq: Setting PowerSaveState to off
XOR registered 4 channels
XOR 2nd invalidate WA enabled
mvCesaInit: sessions=640, queue=64, pSram=f0000000
Warning: TS unit is powered off.
gpio_dev = fd00000
LED_SET_CTL_IOC_NUM= 6702
High gpio polarity = 50
High gpio polarity = 70
Enable Interrupt
mv88f6281_wdt: Watchdog Timer timeout = 8 sec.
mv88f6281_wdt: HW Watchdog timeout = 15 sec.
VFS: Disk quotas dquot_6.5.2
Dquot-cache hash table entries: 1024 (order 0, 4096 bytes)
Installing knfsd (copyright (C) 1996 okir@monad.swb.de).
JFFS2 version 2.2. (NAND) \xffffffc2\xffffffa9 2001-2006 Red Hat, Inc.
fuse init (API version 7.12)
SGI XFS with security attributes, large block/inode numbers, no debug enabled
SGI XFS Quota Management subsystem
yaffs built Jan 27 2016 11:05:07 Installing.
0 -
I think you left out a lot of lines. Here you can find the (serial) bootlog of a 325. The serial bootlog of a 310 shouldn't be much different. The lines from 'Linux version 2.6.31.8' until 'Freeing init memory: 3164K' should all be in dmesg, unless that much extra logging is added that the oldest lines are gone. (It's a 64 or 128kB memory buffer).Between those lines I recognize these sata/disk related lines:SCSI subsystem initialized
Integrated Sata device found
IRQ 21/mvSata: IRQF_DISABLED is not guaranteed on shared IRQs
Wake up disk0......
Wake up disk1......
scsi0 : Marvell SCSI to SATA adapter
scsi1 : Marvell SCSI to SATA adapter
scsi 0:0:0:0: Direct-Access WD800JD-00JNC0 05.0 PQ: 0 ANSI: 5
sd 0:0:0:0: [sda] Sector size 0 reported, assuming 512.
sd 0:0:0:0: [sda] 156301488 512-byte logical blocks: (80.0 GB/74.5 GiB)
sd 0:0:0:0: [sda] 0-byte physical blocks
sd 0:0:0:0: Attached scsi generic sg0 type 0
sd 0:0:0:0: [sda] Write Protect is off
sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA
sd 0:0:0:0: [sda] Sector size 0 reported, assuming 512.
sda: sda1 sda2
sd 0:0:0:0: [sda] Sector size 0 reported, assuming 512.
sd 0:0:0:0: [sda] Attached SCSI diskI expect on your box either some low level sata error, or a mounting problem, or no trace of a 2nd disk at all. But at least the 1st disk should show up, somehow.
0 -
Thank You for your support.The previous output from the dmesg is full, no other lines are displayed ( I do not see any logs under the /var/log).The 1st HDD is mounted correctly, and the 2nd HDD on esata can be monted manually (if I connect to telnet and mount it with command mount /dev/md1 /mnt/extdisk/ - it means the JBOD partitions are correct, just the NAS cannot mount it automatically from any reason).Anyway, I will copying data to another external disk and format the disk again, it seems to be an internal issue with autodetecting of the disk during boot.
0
Categories
- All Categories
- 415 Beta Program
- 2.4K Nebula
- 151 Nebula Ideas
- 98 Nebula Status and Incidents
- 5.7K Security
- 277 USG FLEX H Series
- 277 Security Ideas
- 1.4K Switch
- 74 Switch Ideas
- 1.1K Wireless
- 42 Wireless Ideas
- 6.4K Consumer Product
- 250 Service & License
- 395 News and Release
- 85 Security Advisories
- 29 Education Center
- 10 [Campaign] Zyxel Network Detective
- 3.6K FAQ
- 34 Documents
- 34 Nebula Monthly Express
- 85 About Community
- 75 Security Highlight