CentOS 磁盤陣列(raid10)


1、通過mdadm命令進行磁盤陣列部署

mdadm是multiple devices admin的簡稱,它是Linux下的一款標准的軟件 RAID 管理工具
如果沒有mdadm命令,通過yum安裝一下
yum install -y mdadm
mdadm的主要參數

-a 檢測設備名稱 -n 指定設備數量 -l 指定 RAID 級別 -C 創建 -v 顯示過程 -f 模擬設備損壞 -r 移除設備 -Q 查看摘要信息 -D 查看詳細信息 -S 停止 RAID 磁盤陣列

更多的參數說明使用man mdadm來查看。

2、創建raid10

1)查看硬盤情況

#先看看磁盤情況
[root@bigdata-senior01 ~]# lsblk
NAME            MAJ:MIN RM  SIZE RO TYPE MOUNTPOINT
sda               8:0    0   20G  0 disk 
├─sda1            8:1    0    1G  0 part /boot
└─sda2            8:2    0   19G  0 part 
  ├─centos-root 253:0    0   32G  0 lvm  /
  └─centos-swap 253:1    0    2G  0 lvm  [SWAP]
sdb               8:16   0   20G  0 disk 
├─sdb1            8:17   0    2G  0 part /backup
├─sdb2            8:18   0    3G  0 part [SWAP]
└─sdb3            8:19   0   15G  0 part 
  └─centos-root 253:0    0   32G  0 lvm  /
sdc               8:32   0    1G  0 disk 
sdd               8:48   0    1G  0 disk 
sde               8:64   0    1G  0 disk 
sdf               8:80   0    1G  0 disk 
sr0              11:0    1  4.2G  0 rom 

2)開始創建:

-C 參數代表創建一個 RAID 陣列卡;-v 參數顯示創建的過程,后面是一個設備名稱/dev/md0,是創建后的RAID磁盤陣列的名稱;
-a yes 參數代表自動創建設備文件;-n 4 參數代表使用 4 塊硬盤來部署這個RAID 磁盤陣列;
而-l 10 參數則代表 RAID 10 方案;最后 4 塊硬盤設備的名稱

[root@bigdata-senior01 ~]# mdadm -Cv /dev/md0 -a yes -n 4 -l 10 /dev/sdc /dev/sdd /dev/sde /dev/sdf mdadm: layout defaults to n2 mdadm: layout defaults to n2 mdadm: chunk size defaults to 512K mdadm: size set to 1046528K mdadm: Fail create md0 when using /sys/module/md_mod/parameters/new_array mdadm: Defaulting to version 1.2 metadata mdadm: array /dev/md0 started.
[root
@bigdata-senior01 ~]# file /dev/md0 /dev/md0: block special

3)格式化陣列

[root@bigdata-senior01 ~]# mkfs
mkfs         mkfs.btrfs   mkfs.cramfs  mkfs.ext2    mkfs.ext3    mkfs.ext4    mkfs.minix   mkfs.xfs
[root@bigdata-senior01 ~]# mkfs.ext4 /dev/md0
mke2fs 1.42.9 (28-Dec-2013)
文件系統標簽=
OS type: Linux
塊大小=4096 (log=2)
分塊大小=4096 (log=2)
Stride=128 blocks, Stripe width=256 blocks
130816 inodes, 523264 blocks
26163 blocks (5.00%) reserved for the super user
第一個數據塊=0
Maximum filesystem blocks=536870912
16 block groups
32768 blocks per group, 32768 fragments per group
8176 inodes per group
Superblock backups stored on blocks: 
    32768, 98304, 163840, 229376, 294912

Allocating group tables: 完成                            
正在寫入inode表: 完成                            
Creating journal (8192 blocks): 完成
Writing superblocks and filesystem accounting information: 完成 

4)掛載陣列

[root@bigdata-senior01 ~]# mount /dev/md0 /raid/raid10

[root@bigdata-senior01 ~]# df -h
文件系統                 容量  已用  可用 已用% 掛載點
/dev/mapper/centos-root   32G  4.2G   28G   13% /
devtmpfs                 901M     0  901M    0% /dev
tmpfs                    912M     0  912M    0% /dev/shm
tmpfs                    912M  8.6M  904M    1% /run
tmpfs                    912M     0  912M    0% /sys/fs/cgroup
/dev/sda1               1014M  143M  872M   15% /boot
/dev/sdb1                2.0G   67M  2.0G    4% /backup
tmpfs                    183M     0  183M    0% /run/user/1004
/dev/md0                 2.0G  6.0M  1.9G    1% /raid/raid10

陣列容量2G,符合預期

5)查看陣列信息

[root@bigdata-senior01 ~]# mdadm -D /dev/md0
/dev/md0:
           Version : 1.2
     Creation Time : Wed Jan 16 16:23:44 2019
        Raid Level : raid10
        Array Size : 2093056 (2044.00 MiB 2143.29 MB)
     Used Dev Size : 1046528 (1022.00 MiB 1071.64 MB)
      Raid Devices : 4
     Total Devices : 4
       Persistence : Superblock is persistent

       Update Time : Wed Jan 16 16:31:05 2019
             State : clean 
    Active Devices : 4
   Working Devices : 4
    Failed Devices : 0
     Spare Devices : 0

            Layout : near=2
        Chunk Size : 512K

Consistency Policy : resync

              Name : bigdata-senior01.home.com:0  (local to host bigdata-senior01.home.com)
              UUID : 1ad7fbe9:3ed88dbf:48408f25:9d49eae2
            Events : 17

    Number   Major   Minor   RaidDevice State
       0       8       32        0      active sync set-A   /dev/sdc
       1       8       48        1      active sync set-B   /dev/sdd
       2       8       64        2      active sync set-A   /dev/sde
       3       8       80        3      active sync set-B   /dev/sdf

6)編輯/etc/fstab永久掛載陣列

vi /etc/fstab
# /etc/fstab
# Created by anaconda on Sun Apr 29 17:25:33 2018
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
/dev/mapper/centos-root /                       xfs     defaults        0 0
UUID=fa75616e-a122-4c73-9fd4-b1d50a4af91a /boot                   xfs     defaults        0 0
/dev/mapper/centos-swap swap                    swap    defaults        0 0
/dev/sdb1               /backup                 xfs     defaults        0 0
/dev/sdb2               swap                    swap    defaults        0 0
/dev/md0                /raid/raid10            ext4    defaults        0 0

7)模擬磁盤損壞及修復

#模擬/dev/sdc損壞
[root@bigdata-senior01 ~]# mdadm /dev/md0 -f /dev/sdc
mdadm: set /dev/sdc faulty in /dev/md0

[root@bigdata-senior01 ~]# mdadm -D /dev/md0
/dev/md0:
           Version : 1.2
     Creation Time : Wed Jan 16 16:23:44 2019
        Raid Level : raid10
        Array Size : 2093056 (2044.00 MiB 2143.29 MB)
     Used Dev Size : 1046528 (1022.00 MiB 1071.64 MB)
      Raid Devices : 4
     Total Devices : 4
       Persistence : Superblock is persistent

       Update Time : Wed Jan 16 16:42:27 2019
             State : clean, degraded 
    Active Devices : 3
   Working Devices : 3
    Failed Devices : 1
     Spare Devices : 0

            Layout : near=2
        Chunk Size : 512K

Consistency Policy : resync

              Name : bigdata-senior01.home.com:0  (local to host bigdata-senior01.home.com)
              UUID : 1ad7fbe9:3ed88dbf:48408f25:9d49eae2
            Events : 19

    Number   Major   Minor   RaidDevice State
       -       0        0        0      removed
       1       8       48        1      active sync set-B   /dev/sdd
       2       8       64        2      active sync set-A   /dev/sde
       3       8       80        3      active sync set-B   /dev/sdf

       0       8       32        -      faulty   /dev/sdc

#查看raid狀態,缺了一個U,說明4個盤有一個down了
[root@bigdata-senior01 ~]# cat /proc/mdstat 
Personalities : [raid10] 
md0 : active raid10 sdf[3] sde[2] sdd[1] sdc[0](F)
      2093056 blocks super 1.2 512K chunks 2 near-copies [4/3] [_UUU]

#移除/dev/sdc
root@bigdata-senior01 ~]# mdadm /dev/md0 -r /dev/sdc
mdadm: hot removed /dev/sdc from /dev/md0

#通過lsblk以及mdadm -D /dev/md0命令可以查看一下情況
[root@bigdata-senior01 ~]# lsblk
NAME            MAJ:MIN RM  SIZE RO TYPE   MOUNTPOINT
sda               8:0    0   20G  0 disk   
├─sda1            8:1    0    1G  0 part   /boot
└─sda2            8:2    0   19G  0 part   
  ├─centos-root 253:0    0   32G  0 lvm    /
  └─centos-swap 253:1    0    2G  0 lvm    [SWAP]
sdb               8:16   0   20G  0 disk   
├─sdb1            8:17   0    2G  0 part   /backup
├─sdb2            8:18   0    3G  0 part   [SWAP]
└─sdb3            8:19   0   15G  0 part   
  └─centos-root 253:0    0   32G  0 lvm    /
sdc               8:32   0    1G  0 disk   #已經離開陣列
sdd               8:48   0    1G  0 disk   
└─md0             9:0    0    2G  0 raid10 /raid/raid10
sde               8:64   0    1G  0 disk   
└─md0             9:0    0    2G  0 raid10 /raid/raid10
sdf               8:80   0    1G  0 disk   
└─md0             9:0    0    2G  0 raid10 /raid/raid10
sr0              11:0    1  4.2G  0 rom 

#現在假設/dev/sdc已經修復,重新加回到陣列里
[root@bigdata-senior01 ~]# mdadm /dev/md0 -a /dev/sdc
 mdadm: added /dev/sdc

#四個U,表示4塊盤都正常了
[root@bigdata-senior01 ~]# cat /proc/mdstat 
Personalities : [raid10] 
md0 : active raid10 sdc[4] sdf[3] sde[2] sdd[1]
      2093056 blocks super 1.2 512K chunks 2 near-copies [4/4] [UUUU]
      
unused devices: <none>

用mdadm -D /dev/md0也可以驗證一下

最后:
#重要 mdadm運行時會自動檢查/etc/mdadm.conf  文件並嘗試自動裝配,因此第一次配置raid后可以將信息導入到/etc/mdadm.conf中
[root@bigdata-senior01 ~]# mdadm -Ds /dev/md0 > /etc/mdadm.conf
[root@bigdata-senior01 ~]# cat /etc/mdadm.conf
ARRAY /dev/md0 metadata=1.2 spares=1 name=bigdata-senior01.home.com:0 UUID=8eafb085:276f669d:c1b5cdd7:5ccd72d7

如果不生成mdadm.conf文件,mdadm會嘗試自己加載raid,最后可能會出現把raid生成到/dev/md127里(不一定是127)

3、磁盤陣列+備份盤,實現4+1方式,准備一塊多余的硬盤,這塊硬盤平時處於閑置狀態,一旦 RAID磁盤陣列中有硬盤出現故障后則會馬上自動頂替

1)創建

[root@bigdata-senior01 ~]# cat /proc/partitions #查看硬盤和分區情況
major minor  #blocks  name

   8        0   20971520 sda
   8        1    1048576 sda1
   8        2   19921920 sda2
   8       16   20971520 sdb
   8       17    2097152 sdb1
   8       18    3145728 sdb2
   8       19   15727616 sdb3
   8       32    1048576 sdc
   8       48    1048576 sdd
   8       64    1048576 sde
   8       80    1048576 sdf
   8       96    1048576 sdg
  11        0    4365312 sr0
 253        0   33546240 dm-0
 253        1    2097152 dm-1
[root@bigdata-senior01 ~]# cat /proc/mdstat #查看有沒有陣列
Personalities : 
unused devices: <none>

#普通陣列多一個參數-x,表示空余的盤 [root@bigdata-senior01 ~]# mdadm -Cv /dev/md0 -a yes -n 4 -l 10 -x 1 /dev/sdc /dev/sdd /dev/sde /dev/sdf /dev/sdg
mdadm: layout defaults to n2
mdadm: layout defaults to n2 mdadm: chunk size defaults to 512K mdadm: size set to 1046528K mdadm: Fail create md0 when using /sys/module/md_mod/parameters/new_array mdadm: Defaulting to version 1.2 metadata mdadm: array /dev/md0 started.

#查看陣列創建進度,最后等進度結束后再格式化 [root
@bigdata-senior01 ~]# cat /proc/mdstat Personalities : [raid10] md0 : active raid10 sdg[4](S) sdf[3] sde[2] sdd[1] sdc[0] 2093056 blocks super 1.2 512K chunks 2 near-copies [4/4] [UUUU] [=============>.......] resync = 67.8% (1419776/2093056) finish=0.0min speed=236629K/sec unused devices: <none> [root@bigdata-senior01 ~]# cat /proc/mdstat Personalities : [raid10] md0 : active raid10 sdg[4](S) sdf[3] sde[2] sdd[1] sdc[0] 2093056 blocks super 1.2 512K chunks 2 near-copies [4/4] [UUUU] unused devices: <none> [root@bigdata-senior01 ~]# mdadm -D /dev/md0 /dev/md0: Version : 1.2 Creation Time : Wed Jan 16 17:55:46 2019 Raid Level : raid10 Array Size : 2093056 (2044.00 MiB 2143.29 MB) Used Dev Size : 1046528 (1022.00 MiB 1071.64 MB) Raid Devices : 4 Total Devices : 5 Persistence : Superblock is persistent Update Time : Wed Jan 16 17:55:56 2019 State : clean Active Devices : 4 Working Devices : 5 Failed Devices : 0 Spare Devices : 1 Layout : near=2 Chunk Size : 512K Consistency Policy : resync Name : bigdata-senior01.home.com:0 (local to host bigdata-senior01.home.com) UUID : 8eafb085:276f669d:c1b5cdd7:5ccd72d7 Events : 17 Number Major Minor RaidDevice State 0 8 32 0 active sync set-A /dev/sdc 1 8 48 1 active sync set-B /dev/sdd 2 8 64 2 active sync set-A /dev/sde 3 8 80 3 active sync set-B /dev/sdf 4 8 96 - spare /dev/sdg [root@bigdata-senior01 ~]# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 20G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 19G 0 part ├─centos-root 253:0 0 32G 0 lvm / └─centos-swap 253:1 0 2G 0 lvm [SWAP] sdb 8:16 0 20G 0 disk ├─sdb1 8:17 0 2G 0 part /backup ├─sdb2 8:18 0 3G 0 part [SWAP] └─sdb3 8:19 0 15G 0 part └─centos-root 253:0 0 32G 0 lvm / sdc 8:32 0 1G 0 disk └─md0 9:0 0 2G 0 raid10 sdd 8:48 0 1G 0 disk └─md0 9:0 0 2G 0 raid10 sde 8:64 0 1G 0 disk └─md0 9:0 0 2G 0 raid10 sdf 8:80 0 1G 0 disk └─md0 9:0 0 2G 0 raid10 sdg 8:96 0 1G 0 disk └─md0 9:0 0 2G 0 raid10 sr0 11:0 1 4.2G 0 rom
#重要 mdadm運行時會自動檢查/etc/mdadm.conf  文件並嘗試自動裝配,因此第一次配置raid后可以將信息導入到/etc/mdadm.conf中 [root
@bigdata-senior01 ~]# mdadm -Ds /dev/md0 > /etc/mdadm.conf

2) 格式化

[root@bigdata-senior01 ~]# mkfs.ext4 /dev/md0
mke2fs 1.42.9 (28-Dec-2013)
文件系統標簽=
OS type: Linux
塊大小=4096 (log=2)
分塊大小=4096 (log=2)
Stride=128 blocks, Stripe width=256 blocks
130816 inodes, 523264 blocks
26163 blocks (5.00%) reserved for the super user
第一個數據塊=0
Maximum filesystem blocks=536870912
16 block groups
32768 blocks per group, 32768 fragments per group
8176 inodes per group
Superblock backups stored on blocks: 
    32768, 98304, 163840, 229376, 294912

Allocating group tables: 完成                            
正在寫入inode表: 完成                            
Creating journal (8192 blocks): 完成
Writing superblocks and filesystem accounting information: 完成

3)掛載:

[root@bigdata-senior01 ~]# mount /dev/md0 /raid/raid10
[root@bigdata-senior01 ~]# df -h
文件系統                 容量  已用  可用 已用% 掛載點
/dev/mapper/centos-root   32G  4.2G   28G   13% /
devtmpfs                 901M     0  901M    0% /dev
tmpfs                    912M     0  912M    0% /dev/shm
tmpfs                    912M  8.7M  903M    1% /run
tmpfs                    912M     0  912M    0% /sys/fs/cgroup
/dev/sda1               1014M  143M  872M   15% /boot
/dev/sdb1                2.0G   67M  2.0G    4% /backup
tmpfs                    183M     0  183M    0% /run/user/1004
/dev/md0                 2.0G  6.0M  1.9G    1% /raid/raid10

4)故障測試

#設置/dev/sdc故障
[root@bigdata-senior01 ~]# mdadm /dev/md0 -f /dev/sdc mdadm: set /dev/sdc faulty in /dev/md0 [root@bigdata-senior01 ~]# mdadm -D /dev/md0 /dev/md0: Version : 1.2 Creation Time : Wed Jan 16 17:55:46 2019 Raid Level : raid10 Array Size : 2093056 (2044.00 MiB 2143.29 MB) Used Dev Size : 1046528 (1022.00 MiB 1071.64 MB) Raid Devices : 4 Total Devices : 5 Persistence : Superblock is persistent Update Time : Wed Jan 16 19:09:03 2019 State : clean, degraded, recovering Active Devices : 3 Working Devices : 4 Failed Devices : 1 Spare Devices : 1 Layout : near=2 Chunk Size : 512K Consistency Policy : resync Rebuild Status : 87% complete Name : bigdata-senior01.home.com:0 (local to host bigdata-senior01.home.com) UUID : 8eafb085:276f669d:c1b5cdd7:5ccd72d7 Events : 32 Number Major Minor RaidDevice State 4 8 96 0 spare rebuilding /dev/sdg 1 8 48 1 active sync set-B /dev/sdd 2 8 64 2 active sync set-A /dev/sde 3 8 80 3 active sync set-B /dev/sdf 0 8 32 - faulty /dev/sdc

#/dev/sdg自動頂替了/dev/sdc

5)移除故障盤/dev/sdc,然后重新加入

[root@bigdata-senior01 ~]# mdadm /dev/md0 -r /dev/sdc
mdadm: hot removed /dev/sdc from /dev/md0
[root@bigdata-senior01 ~]# mdadm -D /dev/md0
/dev/md0:
           Version : 1.2
     Creation Time : Wed Jan 16 17:55:46 2019
        Raid Level : raid10
        Array Size : 2093056 (2044.00 MiB 2143.29 MB)
     Used Dev Size : 1046528 (1022.00 MiB 1071.64 MB)
      Raid Devices : 4
     Total Devices : 4
       Persistence : Superblock is persistent

       Update Time : Wed Jan 16 19:13:08 2019
             State : clean 
    Active Devices : 4
   Working Devices : 4
    Failed Devices : 0
     Spare Devices : 0

            Layout : near=2
        Chunk Size : 512K

Consistency Policy : resync

              Name : bigdata-senior01.home.com:0  (local to host bigdata-senior01.home.com)
              UUID : 8eafb085:276f669d:c1b5cdd7:5ccd72d7
            Events : 37

    Number   Major   Minor   RaidDevice State
       4       8       96        0      active sync set-A   /dev/sdg
       1       8       48        1      active sync set-B   /dev/sdd
       2       8       64        2      active sync set-A   /dev/sde
       3       8       80        3      active sync set-B   /dev/sdf

[root@bigdata-senior01 ~]# mdadm /dev/md0 -a /dev/sdc
mdadm: added /dev/sdc
[root@bigdata-senior01 ~]# mdadm -D /dev/md0
/dev/md0:
           Version : 1.2
     Creation Time : Wed Jan 16 17:55:46 2019
        Raid Level : raid10
        Array Size : 2093056 (2044.00 MiB 2143.29 MB)
     Used Dev Size : 1046528 (1022.00 MiB 1071.64 MB)
      Raid Devices : 4
     Total Devices : 5
       Persistence : Superblock is persistent

       Update Time : Wed Jan 16 19:14:38 2019
             State : clean 
    Active Devices : 4
   Working Devices : 5
    Failed Devices : 0
     Spare Devices : 1

            Layout : near=2
        Chunk Size : 512K

Consistency Policy : resync

              Name : bigdata-senior01.home.com:0  (local to host bigdata-senior01.home.com)
              UUID : 8eafb085:276f669d:c1b5cdd7:5ccd72d7
            Events : 38

    Number   Major   Minor   RaidDevice State
       4       8       96        0      active sync set-A   /dev/sdg
       1       8       48        1      active sync set-B   /dev/sdd
       2       8       64        2      active sync set-A   /dev/sde
       3       8       80        3      active sync set-B   /dev/sdf

       5       8       32        -      spare   /dev/sdc

6)備盤可以加多塊,可以在陣列已經創建后hot-add

#按上例,創建是用mdadm -Cv /dev/md0 -a yes -n 4 -l 10 /dev/sdc /dev/sdd /dev/sde /dev/sdf完成,備盤可以后加
[root@bigdata-senior01 ~]# mdadm /dev/md0 --add-spare /dev/sdg
mdadm: added /dev/sdg
[root@bigdata-senior01 ~]# mdadm -D /dev/md0
/dev/md0:
           Version : 1.2
     Creation Time : Wed Jan 16 22:16:28 2019
        Raid Level : raid10
        Array Size : 2093056 (2044.00 MiB 2143.29 MB)
     Used Dev Size : 1046528 (1022.00 MiB 1071.64 MB)
      Raid Devices : 4
     Total Devices : 5
       Persistence : Superblock is persistent

       Update Time : Wed Jan 16 22:19:25 2019
             State : clean 
    Active Devices : 4
   Working Devices : 5
    Failed Devices : 0
     Spare Devices : 1

            Layout : near=2
        Chunk Size : 512K

Consistency Policy : resync

              Name : bigdata-senior01.home.com:0  (local to host bigdata-senior01.home.com)
              UUID : c9d67482:d9fb7776:3384ac75:2d46763a
            Events : 18

    Number   Major   Minor   RaidDevice State
       0       8       32        0      active sync set-A   /dev/sdc
       1       8       48        1      active sync set-B   /dev/sdd
       2       8       64        2      active sync set-A   /dev/sde
       3       8       80        3      active sync set-B   /dev/sdf

       4       8       96        -      spare   /dev/sdg

 

4、移除陣列,不正確移除陣列會引起各種奇葩的問題

1)卸載raid設備

[root@bigdata-senior01 ~]# umount /dev/md0

2)停止陣列

[root@bigdata-senior01 ~]# mdadm -S /dev/md0
mdadm: stopped /dev/md0

PS:重啟陣列
[root@bigdata-senior01 ~]# mdadm -A -s /dev/md0
mdadm: Fail create md0 when using /sys/module/md_mod/parameters/new_array
mdadm: /dev/md0 has been started with 4 drives and 1 spare.

[root@bigdata-senior01 ~]# mdadm -D /dev/md0
mdadm: cannot open /dev/md0: No such file or directory
[root@bigdata-senior01 ~]# cat /proc/mdstat
Personalities : [raid10]
unused devices: <none>

3)刪除磁盤

[root@bigdata-senior01 ~]# mdadm --misc --zero-superblock /dev/sdc
[root@bigdata-senior01 ~]# mdadm --misc --zero-superblock /dev/sdd
[root@bigdata-senior01 ~]# mdadm --misc --zero-superblock /dev/sde
[root@bigdata-senior01 ~]# mdadm --misc --zero-superblock /dev/sdf
[root@bigdata-senior01 ~]# mdadm --misc --zero-superblock /dev/sdg

4)刪除或注銷配置文件mdadm.conf和fstab等

[root@bigdata-senior01 ~]# cat /etc/mdadm.conf 
# ARRAY /dev/md0 metadata=1.2 spares=1 name=bigdata-senior01.home.com:0 UUID=8eafb085:276f669d:c1b5cdd7:5ccd72d7

[root@bigdata-senior01 ~]# cat /etc/fstab # # /etc/fstab # Created by anaconda on Sun Apr 29 17:25:33 2018 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # /dev/mapper/centos-root / xfs defaults 0 0 UUID=fa75616e-a122-4c73-9fd4-b1d50a4af91a /boot xfs defaults 0 0 /dev/mapper/centos-swap swap swap defaults 0 0 /dev/sdb1 /backup xfs defaults 0 0 /dev/sdb2 swap swap defaults 0 0 # /dev/md0 /raid/raid10 ext4 defaults 0 0

reboot看看效果

 


免責聲明!

本站轉載的文章為個人學習借鑒使用,本站對版權不負任何法律責任。如果侵犯了您的隱私權益,請聯系本站郵箱yoyou2525@163.com刪除。



 
粵ICP備18138465號   © 2018-2025 CODEPRJ.COM