EX4100 Raid 5 Degraded and i cant find acurate steps for correction that relate specifically to my situation

Today my Ex4100 had the 4th drive go red on me, the system just locked up, all you got was the following errors after logging into the admin.

Warning: Unknown: write failed: No space left on device (28) in Unknown on line 0

Warning: Unknown: Failed to write session data (files). Please verify that the current setting of session.save_path is correct () in Unknown on line 0

After a standard reboot (had to press the button) the drive was blue during boot and then went red soon after the box was up.

The drive shows a status of healthy, as to the other 3 drives.

Drive4 4 TB 37 °C Good

I can pull the SN etc when i click on the link for Drive 4.
all 4 drives are the same WD 4gb Red all bought less than a year ago.

I ran a quick diagnostic, and every thing came back as good.

SMART Data is as follows

ID Attribute Current Worst Threshold
1 Raw_Read_Error_Rate 200 200 51
3 Spin_Up_Time 180 177 21
4 Start_Stop_Count 83 83 0
5 Reallocated_Sector_Ct 200 200 140
7 Seek_Error_Rate 200 200 0
9 Power_On_Hours 84 84 0
10 Spin_Retry_Count 100 100 0
11 Calibration_Retry_Count 100 253 0
12 Power_Cycle_Count 100 100 0
192 Power-Off_Retract_Count 200 200 0
193 Load_Cycle_Count 194 194 0
194 Temperature_Celsius 115 114 0
196 Reallocated_Event_Count 200 200 0
197 Current_Pending_Sector 200 200 0
198 Offline_Uncorrectable 100 253 0
199 UDMA_CRC_Error_Count 200 200 0
200 Multi_Zone_Error_Rate 100 253 0

everything the system says means that the drive is fine, and yet its screaming for a rebuild as if the drive isnt even there or that its a new drive, which its not.

i just dont get what it really needs, and the support site is lacking in answers, and whats there is not relevant to my situation in any way.

@jerzey4life,

What is the output from “mdstat” ? Does it look similar to what I get on mine below or is it showing RAID members as missing ?

root@WDMyCloudEX4100 / # cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
md1 : active raid5 sdc2[0] sda2[3] sdb2[2] sdd2[1]
8778211776 blocks super 1.0 level 5, 64k chunk, algorithm 2 [4/4] [UUUU]
bitmap: 2/2 pages [64KB], 131072KB chunk

md0 : active raid1 sda1[3] sdb1[2] sdd1[1] sdc1[0]
2097088 blocks [4/4] [UUUU]
bitmap: 0/16 pages [0KB], 8KB chunk

unused devices: none

Cheers,

JediNite

Also show the output of the “df” command.

Given that the errors are saying “no space left on device” that might indicate one of the partitions is full, and if that happens on a NAS, LOTS of weird things can happen.

Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
md1 : active raid5 sdc2[0] sdb2[2] sdd2[1]
11708466624 blocks super 1.0 level 5, 64k chunk, algorithm 2 [4/3] [UUU_]
bitmap: 1/1 pages [32KB], 262144KB chunk

md0 : active raid1 sda1[3] sdb1[2] sdd1[1] sdc1[0]
2097088 blocks [4/4] [UUUU]
bitmap: 0/16 pages [0KB], 8KB chunk

unused devices:

i think that was really related to the fact that the system didnt know what to do when it went wonky, which is not a shock given the ways that this thing likes to throw errors with no meaning or hangs up because it doesnt like the shade of blue the sky is that day.

I have run the full disk test and everything came back green

Self Test Results
Disk1 Passed - Full disk test completed successfully.
Disk2 Passed - Full disk test completed successfully.
Disk3 Passed - Full disk test completed successfully.
Disk4 Passed - Full disk test completed successfully.

Interesting. md0 and md1 are four-disk RAID groups; SDA is fine on md0, but missing from md1. So it’s not entirely a drive issue, but a partition issue.

unfortunately the EX series is woefully short on system logs; you could try dumping the detailed logs from the GUI, or just use mdadm to try to add sda2 back into the raid group.

mdadm --manage /dev/md1 --re-add /dev/sda2

…and see what happens.

1 Like

i will give that a try but not before i try and back up everything i have on the other 3 disks first.

i tried an older external drive i had and that failed badly 3 times because im sure its old an arthritic, but i just grabbed a new drive off the shelf and will back up today\tonight, and then give that command a whirl.

Its amazing how much unix i have forgotten after 15 years, ugh.