# clair hwinfo --disk --short # un peu plus complet fdisk -l # illisible chez moi lsblk
En général c'est
/dev/sda /dev/sdb /dev/sdc ...
Les partition dans chaque disques sont :
/dev/sda1 /dev/sda2 /dev/sda3 ...
cat /proc/mdstat
Les partitions raid sont notées
/dev/md0 /dev/md1 ...
On voit les X disques
et le U dans les crochet dit que c'est up, et le _ dit que c'est mort
la ya bien [UU_] donc 2 disques ok et un mort
cat /sys/block/md*/md/dev-*/state
https://www.kernel.org/doc/html/v4.15/admin-guide/md.html
faulty : device has been kicked from active use due to a detected fault, or it has unacknowledged bad blocks
in_sync : device is a fully in-sync member of the array
writemostly : device will only be subject to read requests if there are no other options. This applies only to raid1 arrays.
blocked: device has failed, and the failure hasn’t been acknowledged yet by the metadata handler. Writes that would write to this device if it were not faulty are blocked.
spare: device is working, but not a full member. This includes spares that are in the process of being recovered to.
write_error : device has ever seen a write error.
want_replacement: device is (mostly) working but probably should be replaced, either due to errors or due to user request.
replacement : device is a replacement for another active device with same raid_disk.
mdadm --detail /dev/md0 | grep -e '^\s*State : ' | awk '{ print $NF; }'
will output "clean" or "active" for a good array. You can also loop over /dev/md/* to get all arrays.
i have a raid1 on 3 disks.
i see "clean" when it is fully nice.
and "active, degraded, recovering" when one disk is nice, the second is removed/dead, and the third is recovering
The possible values (can be comma separated) are described here:
mdadm -D /dev/md0
les dernières lignes listent les disques
Exemple avec un disque mort et un deuxième disque en cours de synchro :
cat /proc/mdstat Personalities : [raid1] [raid0] [raid6] [raid5] [raid4] [linear] [multipath] [raid10] md2 : active raid1 sdb2[1] sda2[3] 1952461760 blocks [3/1] [_U_] [========>............] recovery = 42.7% (834347520/1952461760) finish=1703.6min speed=10937K/sec bitmap: 15/15 pages [60KB], 65536KB chunk md1 : active raid1 sdb1[1] sda1[0] 523200 blocks [3/2] [UU_] unused devices:
root@raphaelpiccolo:~# mdadm -D /dev/md2 /dev/md2: Version : 0.90 Creation Time : Wed Jun 10 15:12:05 2020 Raid Level : raid1 Array Size : 1952461760 (1862.01 GiB 1999.32 GB) Used Dev Size : 1952461760 (1862.01 GiB 1999.32 GB) Raid Devices : 3 Total Devices : 2 Preferred Minor : 2 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Sun Feb 12 21:57:17 2023 State : active, degraded, recovering Active Devices : 1 Working Devices : 2 Failed Devices : 0 Spare Devices : 1 Consistency Policy : bitmap Rebuild Status : 42% complete UUID : e879a2ab:989cc6fb:a4d2adc2:26fd5302 Events : 0.2129993 Number Major Minor RaidDevice State 3 8 2 0 spare rebuilding /dev/sda2 1 8 18 1 active sync /dev/sdb2 - 0 0 2 removed
smartctl -i /dev/sdc # si besoin lancer une analyse smartctl -a /dev/sdc smartctl -t short /dev/sdc
Vérifier le disque :
fsck -f -y /dev/sdc1 fsck -f -y /dev/sdc2 fsck -f -y /dev/sdc3
choisir le bon raid (md1) et la bonne partition (sdc1) en se basant sur le retour de la commande "mdadm -D /dev/md1"
mdadm /dev/md1 --add /dev/sdc1 mdadm /dev/md2 --add /dev/sdc2