Line 7: |
Line 7: |
| | | |
| [[User:Stephdl|Stephdl]] ([[User talk:Stephdl|talk]]) 12:24, 6 February 2013 (MST) ok i work on the howTo...work in progress, don't disturb :p | | [[User:Stephdl|Stephdl]] ([[User talk:Stephdl|talk]]) 12:24, 6 February 2013 (MST) ok i work on the howTo...work in progress, don't disturb :p |
| + | |
| + | |
| + | |
| + | |
| + | == HowTo: Remove a disk from the RAID1 Array from the command Line == |
| + | |
| + | === Look at the mdstat === |
| + | |
| + | First we must determine which drive is in default. |
| + | |
| + | |
| + | [root@sme8-64-dev ~]# cat /proc/mdstat |
| + | Personalities : [raid1] |
| + | md1 : active raid1 sdb1[1] sda1[0] |
| + | 104320 blocks [2/2] [UU] |
| + | |
| + | md2 : active raid1 sdb2[2](F) sda2[0] |
| + | 52323584 blocks [2/1] [U_] |
| + | |
| + | unused devices: <none> |
| + | |
| + | (S)= Spare |
| + | (F)= Fail |
| + | [0]= number of the disk |
| + | |
| + | {{notebox|as we can see the partition sdb2 is in default, we can see the flag: sdb2 [2] (F) |
| + | we need to resynchronize the disk sdb to the existing raid md2.}} |
| + | |
| + | === Fail and remove the disk, '''sdb''' in this case === |
| + | |
| + | [root@ ~]# '''mdadm --manage /dev/md2 --fail /dev/sdb2''' |
| + | mdadm: set /dev/sdb2 faulty in /dev/md2 |
| + | [root@ ~]# '''mdadm --manage /dev/md2 --remove /dev/sdb2''' |
| + | mdadm: hot removed /dev/sdb2 |
| + | [root@ ~]# '''mdadm --manage /dev/md1 --fail /dev/sdb1''' |
| + | mdadm: set /dev/sdb1 faulty in /dev/md1 |
| + | [root@ ~]# '''mdadm --manage /dev/md1 --remove /dev/sdb1''' |
| + | mdadm: hot removed /dev/sdb1 |
| + | |
| + | === Do your Disk Maintenance here === |
| + | |
| + | At this point the disk is idle. |
| + | |
| + | [root@sme8-64-dev ~]# cat /proc/mdstat |
| + | Personalities : [raid1] |
| + | md1 : active raid1 sda1[0] |
| + | 104320 blocks [2/1] [U_] |
| + | |
| + | md2 : active raid1 sda2[0] |
| + | 52323584 blocks [2/1] [U_] |
| + | |
| + | unused devices: <none> |
| + | |
| + | {{notebox|You'll have to determine if your disk can be reinstated at the array. In fact sometimes a raid can get out of sync after a power failure but also some outages times for physical disk itself. It is necessary to test the hard drive if this occurs repeatedly. For this we will use smartctl.}} |
| + | |
| + | |
| + | |
| + | |
| + | === Add the partitions back === |
| + | |
| + | [root@ ~]# '''mdadm --manage /dev/md1 --add /dev/sdb1''' |
| + | mdadm: hot added /dev/sdb1 |
| + | [root@ ~]# '''mdadm --manage /dev/md2 --add /dev/sdb2''' |
| + | mdadm: hot added /dev/sdb2 |
| + | |
| + | === Another Look at the mdstat === |
| + | |
| + | [root@ ~]# '''cat /proc/mdstat''' |
| + | Personalities : [raid1] |
| + | md2 : active raid1 sdb2[2] sda2[0] |
| + | 488279488 blocks [2/1] [U_] |
| + | [=>...................] recovery = 6.3% (31179264/488279488) finish=91.3min speed=83358K/sec |
| + | md1 : active raid1 sdb1[1] sda1[0] |
| + | 104320 blocks [2/2] [UU] |
| + | |
| + | unused devices: <none> |