Changes

Jump to navigation Jump to search
m
no edit summary
Line 24: Line 24:     
Install the required tools using the Yum commands below.
 
Install the required tools using the Yum commands below.
  yum Install parted<br>
+
  yum install parted<br>
 
  yum install xfsprogs
 
  yum install xfsprogs
 
   
 
   
Line 39: Line 39:  
  print
 
  print
   −
Remember to set the max size (4.00TB above) to whatever size hard disk drives you are using also you will need tochange the device sdX value for each drive in the chain.
+
Remember to set the max size (4.00TB above) to whatever size hard disk drives you are using also you will need to change the device sdX value for each drive in the chain.
    
===Creating the array===
 
===Creating the array===
Line 46: Line 46:  
Change the md3 number to the one you have selected, you will also need to change the number of raid devices and list each of the devices to use in the array.
 
Change the md3 number to the one you have selected, you will also need to change the number of raid devices and list each of the devices to use in the array.
   −
{{Note box|Be Advised this command will process quickly but in the background it will be conducted a re-sync of the RAID array you can monitor this by running cat /proc/mdstatIt highly advisable to wait for the re-sync to be completed before proceeding.}}
+
{{Note box|Be Advised this command will process quickly but will continue to run a re-sync of the RAID array in the background. You can monitor this re-sync by running cat /proc/mdstat, It highly advisable to wait for the re-sync to be completed before proceeding.}}
    
===Save raid array structure===
 
===Save raid array structure===
Line 53: Line 53:  
  mdadm --detail --scan | grep md3 >> /etc/mdadm.conf
 
  mdadm --detail --scan | grep md3 >> /etc/mdadm.conf
   −
Once that command is run it would be a good idea to look inside it. You should only see one new line in that file.
+
Once that command is run it would be a good advisable to view the file, You should only see one new line in the file.
It is the then advisable to create a copy of the mdadm.conf run the following
+
It is the then a good idea to create a copy of the mdadm.conf by running the following command
 
  cp /etc/mdadm.conf /etc/mdadm.conf.bak
 
  cp /etc/mdadm.conf /etc/mdadm.conf.bak
 
Keep this file permanently as it is possible in the future something may occur on your SME Server to cause the mdadm.conf file to be trashed or reset, in that case you can use this backup file to rewrite your required raid information.
 
Keep this file permanently as it is possible in the future something may occur on your SME Server to cause the mdadm.conf file to be trashed or reset, in that case you can use this backup file to rewrite your required raid information.
Line 66: Line 66:  
{{Note box|I have noticed these commands do take a while, be patience..}}
 
{{Note box|I have noticed these commands do take a while, be patience..}}
    +
===Format your new Partition and testing===
 +
Run the following
 +
mkfs.xfs /dev/vg_DATA/lv_DATA
    +
If you want to be sure everything went ok you can run a file check on the new LVM once the format is complete.
 +
xfs_check /dev/vg_DATA/lv_DATA
   −
===Adding partitions===
+
{{Note box|I found that I could not use EXT3 or EXT4 based file systems due to problems with the block sizes and my 20TB setup there may be work-around for this, but I didn’t find anything solid so instead I decided use XFS file system as it does what I need it too.}}
{{Note box|msg=The process can take many hours or even days. There is a critical section at start, which cannot be backed up. To allow recovery after unexpected power failure, an additional option <code>--backup-file=</code> can be specified. Make sure this file is on a different disk or it defeats the purpose.
     −
mdadm --grow --raid-devices=5 --backup-file=/root/grow_md1.bak /dev/md1
+
===Mount your new partition to a directory===
mdadm --grow --raid-devices=4 --backup-file=/root/grow_md2.bak /dev/md2}}
+
Finally open /etc/fstab and edit the bottom line to mount the new area be sure to leave a new line feed at the bottom, and use proper spacing.
   −
Now we need to add the first partition /dev/sde1 to /dev/md1
+
For Example in my file I entered
 +
/dev/vg_DATA/lv_DATA /TESTFOLDER XFS defaults 0 0
   −
[root@smeraid5 ~]# mdadm --add /dev/md1 /dev/sde1
+
You trigger a remount by using
mdadm: added /dev/sde1
+
  mount –a
[root@smeraid5 ~]# mdadm --grow --raid-devices='''5''' /dev/md1
  −
  −
Here we use the option --raid-devices='''5''' because raid1 uses all drives. You can see how the array looks by:
  −
{{Warning box|During the raid growing step you DO not shutdown your computer, or experienced an electrical failure, these issues can let your computer in a badly status and you can loose your data}}
  −
[root@smeraid5 ~]# mdadm --detail /dev/md1
  −
/dev/md1:
  −
        Version : 0.90
  −
  Creation Time : Tue Oct 29 21:04:15 2013
  −
    Raid Level : raid1
  −
    Array Size : 104320 (101.89 MiB 106.82 MB)
  −
  Used Dev Size : 104320 (101.89 MiB 106.82 MB)
  −
  Raid Devices : 5
  −
  Total Devices : 5
  −
Preferred Minor : 1
  −
    Persistence : Superblock is persistent
  −
  −
    Update Time : Tue Oct 29 21:39:00 2013
  −
          State : clean
  −
Active Devices : 5
  −
Working Devices : 5
  −
Failed Devices : 0
  −
  Spare Devices : 0
  −
  −
          UUID : 15eb70b1:3d0293bb:f3c49d70:6fc5aa4d
  −
        Events : 0.4
  −
  −
    Number  Major  Minor  RaidDevice State
  −
      0      8        1        0      active sync  /dev/sda1
  −
      1      8      17        1      active sync  /dev/sdb1
  −
      2      8      33        2      active sync  /dev/sdc1
  −
      3      8      49        3      active sync  /dev/sdd1
  −
      4      8      65        4      active sync  /dev/sde1
  −
  −
After that we have to do the same thing with the md2 which is a raid5 array.
  −
 
  −
[root@smeraid5 ~]# mdadm --add /dev/md2 /dev/sde2
  −
mdadm: added /dev/sde2
  −
 
  −
[root@smeraid5 ~]# mdadm --grow --raid-devices='''4''' /dev/md2
  −
mdadm: Need to backup 14336K of critical section..
  −
mdadm: ... critical section passed.
  −
 
  −
{{tip box|msg=You need to keep --raid-devices='''4''' if you want to have an array of 4 drives+1spare, However if you do not want a spare drive, you should  set --raid-devices='''5'''. This command can be used to grow an array of raid on the spare drive, just say to mdadm that you want to use all disks connected to the computer.}}
  −
 
  −
{{Warning box|During the raid growing step you DO not shutdown your computer, or experienced an electrical failure, these issues can let your computer in a badly status and you can loose your data}}
  −
 
  −
we can take a look to the md2 array
  −
  −
[root@smeraid5 ~]# mdadm --detail /dev/md2
  −
/dev/md2:
  −
        Version : 0.90
  −
  Creation Time : Tue Oct 29 21:04:28 2013
  −
    Raid Level : raid5
  −
    Array Size : 32644096 (30.28 GiB 31.39 GB)
  −
  Used Dev Size : 7377728 (7.90 GiB 9.63 GB)
  −
  Raid Devices : 4
  −
  Total Devices : 5
  −
Preferred Minor : 2
  −
    Persistence : Superblock is persistent
  −
  −
    Update Time : Tue Oct 29 21:39:29 2013
  −
          State : clean
  −
Active Devices : 4
  −
Working Devices : 5
  −
Failed Devices : 0
  −
  Spare Devices : 1
  −
  −
        Layout : left-symmetric
  −
    Chunk Size : 256K
  −
  −
          UUID : d2c26bed:b5251648:509041c5:fab64ab4
  −
        Events : 0.462
  −
  −
    Number  Major  Minor  RaidDevice State
  −
      0      8        2        0      active sync  /dev/sda2
  −
      1      8      18        1      active sync  /dev/sdb2
  −
      3      8      34        2      active sync  /dev/sdc2
  −
      4      8      50        3      active sync  /dev/sde2
  −
  −
      2      8      114        -      spare  /dev/sdd2
  −
 
  −
===LVM: Growing the PV===
  −
 
  −
{{Note box|Once the construction is complete, we have to set the LVM to use the whole space}}
  −
 
  −
* In a root terminal, issue the following command lines
  −
 
  −
[root@smeraid5 ~]# pvresize /dev/md2
  −
  Physical volume "/dev/md2" changed
  −
  1 physical volume(s) resized / 0 physical volume(s) not resized
  −
 
  −
* after that we can resize the LVM
  −
 
  −
[root@smeraid5 ~]# lvresize -l +100%FREE  /dev/main/root
  −
  Extending logical volume root to 30,25 GB
  −
  Logical volume root successfully resized
  −
 
  −
{{tip box|/dev/main/root is the default name, but if you have changed this you can find it by typing the command : lvdisplay}}
  −
 
  −
  [root@smeraid5 ~]# resize2fs  /dev/main/root
  −
resize2fs 1.39 (29-May-2006)
  −
Filesystem at /dev/main/root is mounted on /; on-line resizing required
  −
Performing an on-line resize of /dev/main/root to 19726336 (4k) blocks.
  −
 
  −
* You should verify that your LVM use the whole drive space with the command
     −
[root@smeraid5 ~]# pvdisplay
+
You can also check whether it has been successful mounted easily by running it should list your mount location and size in use.
  --- Physical volume ---
+
df -h
  PV Name              /dev/md2
  −
  VG Name              main
  −
  PV Size              30.25 GB / not usable 8,81 MB
  −
  Allocatable          yes (but full)
  −
  PE Size (KByte)      32768
  −
  Total PE              1533
  −
  '''Free PE              0'''
  −
  Allocated PE          1533
  −
  PV UUID              a31UBW-2SN6-CXFk-qLOZ-qrsQ-BIYo-nZexXo
     −
if you can see that you have no more '''FREE PE''' you are the king of raid. But you can see also with the command
+
*This setup in /etc/fstab should be maintained when updates or upgrades are conducted however if you want a more definite solution I would advise reading up on templates in SME Server.
   −
[root@smeraid5 ~]# lvdisplay
+
<noinclude>
<noinclude>[[Category:Howto]][[Category:Administration:Storage]]</noinclude>
+
[[Category:Howto]]
 +
[[Category:Administration:Storage]]
 +
</noinclude>
3,054

edits

Navigation menu