RAID1 array created with mdadm; array was rebuilding when there was a power failure. On coming back up, the array seems to be dead.
# cat /proc/mdstat
Personalities : [raid1]
md0 : inactive sdb1[2](S) sda1[3](S)
15627751424 blocks super 1.2
# mdadm --stop /dev/md0
mdadm: stopped /dev/md0
# mdadm --assemble /dev/md1 /dev/sda1 /dev/sdb1
mdadm: /dev/md1 assembled from 0 drives and 1 rebuilding - not enough to start the array.
I don't know why it says 0 drives - I expected the 1 rebuilding but the other drive should have been fine. I really don't want to accidentally do anything destructive - what should I do to get the array back together to a state I can run fsck on it?
/dev/sda1:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x3
Array UUID : c4ea2289:c63bc8ce:e6fe5806:5bebe020
Name : ******:0 (local to host ******)
Creation Time : Thu Aug 20 20:48:53 2020
Raid Level : raid1
Raid Devices : 2
Avail Dev Size : 15627718656 (7451.88 GiB 8001.39 GB)
Array Size : 7813859328 (7451.88 GiB 8001.39 GB)
Data Offset : 264192 sectors
Super Offset : 8 sectors
Recovery Offset : 7331083264 sectors
Unused Space : before=264112 sectors, after=0 sectors
State : clean
Device UUID : ae984a41:f3e421f4:f10e1fac:d7955178
Internal Bitmap : 8 sectors from superblock
Update Time : Sat Oct 31 17:13:53 2020
Bad Block Log : 512 entries available at offset 40 sectors
Checksum : 4be47968 - correct
Events : 1079809
Device Role : Active device 0
Array State : AA ('A' == active, '.' == missing, 'R' == replacing)
/dev/sdb1:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x1
Array UUID : c4ea2289:c63bc8ce:e6fe5806:5bebe020
Name : ******:0 (local to host ******)
Creation Time : Thu Aug 20 20:48:53 2020
Raid Level : raid1
Raid Devices : 2
Avail Dev Size : 15627784192 (7451.91 GiB 8001.43 GB)
Array Size : 7813859328 (7451.88 GiB 8001.39 GB)
Used Dev Size : 15627718656 (7451.88 GiB 8001.39 GB)
Data Offset : 264192 sectors
Super Offset : 8 sectors
Unused Space : before=264112 sectors, after=65536 sectors
State : active
Device UUID : 0abf70e2:7e0e43ca:a22548a2:ef87e9c0
Internal Bitmap : 8 sectors from superblock
Update Time : Sat Oct 31 17:13:31 2020
Bad Block Log : 512 entries available at offset 40 sectors
Checksum : 4cf65df4 - correct
Events : 1079799
Device Role : Active device 1
Array State : AA ('A' == active, '.' == missing, 'R' == replacing)
Looking at the drives with xxd, both do still seem to contain data. Is this something I could reasonable manage myself or is it time for a data recovery company?
Edit: Wazoox's answer worked.
# mdadm --assemble --scan --force
mdadm: forcing event count in /dev/sda1(1) from 1079799 upto 1079809
mdadm: /dev/md0 has been started with 1 drive (out of 2) and 1 rebuilding.
# cat /proc/mdstat
Personalities : [raid1]
md0 : active (auto-read-only) raid1 sdb1[3] sda1[2]
7813859328 blocks super 1.2 [2/1] [_U]
bitmap: 59/59 pages [236KB], 65536KB chunk