Changes

From SME Server
Jump to navigationJump to search
no edit summary
Line 7: Line 7:     
[[User:Stephdl|Stephdl]] ([[User talk:Stephdl|talk]]) 12:24, 6 February 2013 (MST) ok i work on the howTo...work in progress, don't disturb :p
 
[[User:Stephdl|Stephdl]] ([[User talk:Stephdl|talk]]) 12:24, 6 February 2013 (MST) ok i work on the howTo...work in progress, don't disturb :p
 +
 +
 +
 +
 +
== HowTo: Remove a disk from the RAID1 Array from the command Line ==
 +
 +
=== Look at the mdstat ===
 +
 +
First we must determine which drive is in default.
 +
 +
 +
[root@sme8-64-dev ~]# cat /proc/mdstat
 +
Personalities : [raid1]
 +
md1 : active raid1 sdb1[1] sda1[0]
 +
      104320 blocks [2/2] [UU]
 +
     
 +
md2 : active raid1 sdb2[2](F) sda2[0]
 +
      52323584 blocks [2/1] [U_]
 +
     
 +
unused devices: <none>
 +
 +
(S)= Spare
 +
(F)= Fail
 +
[0]= number of the disk
 +
 +
{{notebox|as we can see the partition sdb2 is in default, we can see the flag: sdb2 [2] (F)
 +
we need to resynchronize the disk sdb to the existing raid md2.}}
 +
 +
=== Fail and remove the disk, '''sdb''' in this case ===
 +
 +
[root@ ~]# '''mdadm --manage /dev/md2 --fail /dev/sdb2'''
 +
mdadm: set /dev/sdb2 faulty in /dev/md2
 +
[root@ ~]# '''mdadm --manage /dev/md2 --remove /dev/sdb2'''
 +
mdadm: hot removed /dev/sdb2
 +
[root@ ~]# '''mdadm --manage /dev/md1 --fail /dev/sdb1'''
 +
mdadm: set /dev/sdb1 faulty in /dev/md1
 +
[root@ ~]# '''mdadm --manage /dev/md1 --remove /dev/sdb1'''
 +
mdadm: hot removed /dev/sdb1
 +
 +
=== Do your Disk Maintenance here ===
 +
 +
At this point the disk is idle.
 +
 +
[root@sme8-64-dev ~]# cat /proc/mdstat
 +
Personalities : [raid1]
 +
md1 : active raid1 sda1[0]
 +
      104320 blocks [2/1] [U_]
 +
     
 +
md2 : active raid1 sda2[0]
 +
      52323584 blocks [2/1] [U_]
 +
     
 +
unused devices: <none>
 +
 +
{{notebox|You'll have to determine if your disk can be reinstated at the array. In fact sometimes a raid can get out of sync after a power failure but also some outages times for physical disk itself. It is necessary to test the hard drive if this occurs repeatedly. For this we will use smartctl.}}
 +
 +
 +
 +
 +
=== Add the partitions back ===
 +
 +
[root@ ~]# '''mdadm --manage /dev/md1 --add /dev/sdb1'''
 +
mdadm: hot added /dev/sdb1
 +
[root@ ~]# '''mdadm --manage /dev/md2 --add /dev/sdb2'''
 +
mdadm: hot added /dev/sdb2
 +
 +
=== Another Look at the mdstat ===
 +
 +
[root@ ~]# '''cat /proc/mdstat'''
 +
Personalities : [raid1]
 +
md2 : active raid1 sdb2[2] sda2[0]
 +
      488279488 blocks [2/1] [U_]
 +
      [=>...................]  recovery =  6.3% (31179264/488279488) finish=91.3min speed=83358K/sec
 +
md1 : active raid1 sdb1[1] sda1[0]
 +
      104320 blocks [2/2] [UU]
 +
 +
unused devices: <none>

Navigation menu