Koozali.org: home of the SME Server

RAID Management Question

stefan_gk

RAID Management Question
« on: April 20, 2007, 09:12:21 AM »
Yesterday I decide to check RAID status of my server via admin console and I got the following:

Code: [Select]
Disk redundancy status as of Friday April 20, 2007 09:38:59
Current RAID status:
Personalities: [raidl]
md1: active raidl hda1 [0] hdb1 [1]
102208 blocks [2/2] [UU] md2: active raidl hdb2 [1]
79931328 blocks [2/1] [_U] unused devices: <none>
Only some of the RAID devices are unclean.
Manual intervention may be required.


Only some of the RAID devices are unclean.
Manual intervention may be required.


This one make me very concerned!!!

I don't have any expeience with mdadm and chkdsk with new versions of SME and kernel.

Any help what should I do and some explanation will be very appreciated.

Offline TrevorB

  • *
  • 259
  • +0/-0
    • http://www.batley.id.au
Re: RAID Management Question
« Reply #1 on: April 20, 2007, 10:52:00 AM »
Quote from: "stefan_gk"
Code: [Select]
md1: active raidl hda1 [0] hdb1 [1]
102208 blocks [2/2] [UU] md2: active raidl hdb2 [1]
79931328 blocks [2/1] [_U] unused devices: <none>
Only some of the RAID devices are unclean.
Manual intervention may be required.

Any help what should I do and some explanation will be very appreciated.
Fist thing is to get a little more detail on md2
Code: [Select]
mdadm --detail /dev/md2Which should show you something that shows that /dev/hda2 has been failed (or somesuch).

As /dev/hda1 seems OK, you could try to fail, remove and then re-add /dev/hda2 (this will then get it resynced).
Code: [Select]
mdadm /dev/md2 -f /dev/hda2 -r /dev/hda2
mdadm /dev/md2 -a /dev/hda2
mdadm --detail /dev/md2
Hopefully you will now have something like
Code: [Select]
       Version : 00.90.01
  Creation Time : Sun May 28 15:42:05 2006
     Raid Level : raid1
     Array Size : 156183808 (148.95 GiB 159.93 GB)
    Device Size : 156183808 (148.95 GiB 159.93 GB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 2
    Persistence : Superblock is persistent

    Update Time : Fri Apr 20 18:41:16 2007
          State : clean
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0


    Number   Major   Minor   RaidDevice State
       0       3        2        0      resyncing   /dev/hda2
       1      22       66        1      active sync   /dev/hdb2
And it will resync /dev/hda2 within the existing array.

If not we'll have to look at some more drastic measures  :wink:

stefan_gk

Re: RAID Management Question
« Reply #2 on: April 20, 2007, 11:15:51 AM »
I have done what you propose and after that I have

Code: [Select]
[root@server ~]# mdadm --detail --verbose /dev/md2
/dev/md2:
        Version : 00.90.01
  Creation Time : Sat Oct 28 22:10:16 2006
     Raid Level : raid1
     Array Size : 79931328 (76.23 GiB 81.85 GB)
    Device Size : 79931328 (76.23 GiB 81.85 GB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 2
    Persistence : Superblock is persistent

    Update Time : Fri Apr 20 12:12:17 2007
          State : clean, degraded, recovering
 Active Devices : 1
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 1


 Rebuild Status : 17% complete

    Number   Major   Minor   RaidDevice State
       0       0        0       -1      removed
       1       3       66        1      active sync   /dev/hdb2
       2       3        2        0      spare   /dev/hda2
           UUID : 9a5bd1ac:20d83580:40e21508:30f62149
         Events : 0.2976151


and in admin console in states that now resyncing /dev/hda2 from /dev/hdb2 ...

stefan_gk

RAID Management Question
« Reply #3 on: April 20, 2007, 11:55:01 AM »
Last message after finish of sync is:

Now all RAID devices are in clean state.

Thanks a lot for help!

Offline TrevorB

  • *
  • 259
  • +0/-0
    • http://www.batley.id.au
RAID Management Question
« Reply #4 on: April 20, 2007, 12:12:37 PM »
Quote from: "stefan_gk"
Last message after finish of sync is:

Now all RAID devices are in clean state.
Great!

Now keep an eye on the raid stats to see if anything happens again. The drive itself didn't fail (otherwise there should have been a problem with /dev/hd1 as well), but it may be a symptom.

Good Luck

Offline ltwally

  • ****
  • 108
  • +0/-0
mdadm
« Reply #5 on: April 23, 2007, 08:38:16 PM »
I'm new to SME, and am wondering:  what is the command that you're using to check the software raid status.  (And, yes, I did check the SME manual.  I even googled around a bit, before posting.)

Offline skydivers

  • *
  • 178
  • +0/-0
Re: mdadm
« Reply #6 on: April 23, 2007, 08:54:20 PM »
Quote from: "ltwally"
I'm new to SME, and am wondering:  what is the command that you're using to check the software raid status.  (And, yes, I did check the SME manual.  I even googled around a bit, before posting.)


Login as admin in your terminal. Check manage disk redundancy!

Greets

Offline bpivk

  • *
  • 908
  • +0/-0
    • http://www.bezigrad.com
RAID Management Question
« Reply #7 on: April 23, 2007, 10:12:46 PM »
I have the same raid problem (this topic reminded me to check my raid array).  :lol:

But i have the problem with removing md2.
mdadm: hot remove failed for /dev/hda2: Device or resource busy

I don't want to manually unplug and mess with my server and i tryed rebooting so is there any other way?

 cat /proc/mdstat gives:
Code: [Select]
Personalities : [raid1]
md1 : active raid1 hda1[0] hdd1[1]
      104320 blocks [2/2] [UU]

md2 : active raid1 hda2[0]
      38973568 blocks [2/1] [U_]


mdadm --detail /dev/md2 gives:

Code: [Select]
mdadm --detail /dev/md2
/dev/md2:
        Version : 00.90.01
  Creation Time : Mon Oct 16 18:40:05 2006
     Raid Level : raid1
     Array Size : 38973568 (37.17 GiB 39.91 GB)
    Device Size : 38973568 (37.17 GiB 39.91 GB)
   Raid Devices : 2
  Total Devices : 1
Preferred Minor : 2
    Persistence : Superblock is persistent

    Update Time : Mon Apr 23 23:09:58 2007
          State : clean, degraded
 Active Devices : 1
Working Devices : 1
 Failed Devices : 0
  Spare Devices : 0


    Number   Major   Minor   RaidDevice State
       0       3        2        0      active sync   /dev/hda2
       1       0        0       -1      removed
           UUID : 6bad674e:eceb395b:ad0d495c:0498790d
         Events : 0.8110286
"It should just work" if it doesn't report it. Thanks!

Offline Gaston94

  • *****
  • 184
  • +0/-0
RAID Management Question
« Reply #8 on: April 23, 2007, 11:18:16 PM »
Hi,
you don't need to remove anything : the systeme did it already.

You have to try to join back you second disk on the second raid array
Code: [Select]
#mdadm /dev/md2 -a /dev/hdd2

Note : the message you provided shows thta your main array device, the one were all the system, the i-bay, ... resides.
This is an alert you have to care about, and monitor carefully how this  goes in the next days, weeks, ....
You should have received some mail alert, look at the mdadm monitor option it was no enabled for you.

G.

Offline bpivk

  • *
  • 908
  • +0/-0
    • http://www.bezigrad.com
RAID Management Question
« Reply #9 on: April 23, 2007, 11:22:15 PM »
Well i'm glad that the remove command worked but
Code: [Select]
#mdadm /dev/md2 -a /dev/hdd2 doesn't work.
I still get the same screen and no rebuild process.

Edit: Forget that. It just seems my server is slow. It is rebuilding now. (It took me 10 minutes to remove the disk. And another 10 to add it. :) Snail speed!!! ) Now i just don't know... Is my server this slow to follow commands or my internet line.  :lol:

Gaston thanks. Have enterd the same commands but they seem to work better if i paste them from you.  :shock:
"It should just work" if it doesn't report it. Thanks!

Offline ltwally

  • ****
  • 108
  • +0/-0
Re: mdadm
« Reply #10 on: April 23, 2007, 11:44:01 PM »
Quote from: "skydivers"
Quote from: "ltwally"
I'm new to SME, and am wondering:  what is the command that you're using to check the software raid status.  (And, yes, I did check the SME manual.  I even googled around a bit, before posting.)


Login as admin in your terminal. Check manage disk redundancy!

Greets


Yeah.. that was absolutely no help.  You might as well have just said what you meant, "RTFM NOOB!!!"

I'm looking for the mdadm console command that generated the output of the parent post.  If someone knows it, I'd appreciate you posting it.  Thanks.

Offline Gaston94

  • *****
  • 184
  • +0/-0
RAID Management Question
« Reply #11 on: April 24, 2007, 12:24:09 AM »
Hi,
the answer was exactly the correct one.
Let me be more detailled : when you connect to the server console, you have a function menu, and one of the choices is "manage disk redundancy" (choice 5)
From there you have the exact output displayed
From there you should be able to add a disk to the array, replace defective raid part, ...
Sometime I do like to use the terminal command instead ;) , and the same display can be retrieved (more or less) with a "cat /proc/mdstat" command.
All output on this thread, unless TrevorB's one, came from the command line

G.

markdeblois

Dell PowerEdge SC1430 Hardware Raid
« Reply #12 on: May 08, 2007, 12:09:26 AM »
Hi there,

Sorry for my newbie questions. I have bought a Dell PowerEdge SC1430 server with 2 raid 1 disks (250Gb). I was planning to install SME Server with a software raid but the machine came with a hardware raid (SAS 5iR U320 SAS Controller). Installation went fine and it works well.

What I am not sure of is whether the raid was configured correctly automatically. In the server-manager menu Disk Redundancy it tells me that it might be using hardware mirroring, which I believe is the case. But whether this is actually working I am not sure. I would greatly appreciate your help!

If I try: cat /proc/mdstat , I get

Personalities : [raid1]
md1 : active raid1 sda1[0]
      104320 blocks [2/1] [U_]
     
md2 : active raid1 sda2[0]
      243055296 blocks [2/1] [U_]
     
unused devices: <none>

The result of fdisk -l is:

Disk /dev/sda: 248.9 GB, 248999051264 bytes
255 heads, 63 sectors/track, 30272 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sda1   *           1          13      104391   fd  Linux raid autodetect
/dev/sda2              14       30272   243055417+  fd  Linux raid autodetect

Disk /dev/md2: 248.8 GB, 248888623104 bytes
2 heads, 4 sectors/track, 60763824 cylinders
Units = cylinders of 8 * 512 = 4096 bytes


Disk /dev/md1: 106 MB, 106823680 bytes
2 heads, 4 sectors/track, 26080 cylinders
Units = cylinders of 8 * 512 = 4096 bytes


And if I try mdadm --detail /dev/md2 I get:
/dev/md2:
        Version : 00.90.01
  Creation Time : Fri Apr  6 02:48:56 2007
     Raid Level : raid1
     Array Size : 243055296 (231.80 GiB 248.89 GB)
    Device Size : 243055296 (231.80 GiB 248.89 GB)
   Raid Devices : 2
  Total Devices : 1
Preferred Minor : 2
    Persistence : Superblock is persistent

    Update Time : Mon May  7 23:22:30 2007
          State : clean, degraded
 Active Devices : 1
Working Devices : 1
 Failed Devices : 0
  Spare Devices : 0


    Number   Major   Minor   RaidDevice State
       0       8        2        0      active sync   /dev/sda2
       1       0        0       -1      removed
           UUID : 212b400a:2e501081:36defda4:6d35900e
         Events : 0.46590

I am not 100% sure how to interpret this information. Thanks so much for pointing me in the right direction!

cheers,
Mark

Offline jumba

  • *****
  • 291
  • +0/-0
  • Donations: July 2007 - $ 20.00
    • Smeserver på svenska!
RAID Management Question
« Reply #13 on: August 02, 2007, 09:44:13 AM »
I have a similar issue, with two SCSI disks:

Code: [Select]
[root@server ~]# mdadm --detail --verbose /dev/md1
/dev/md1:
        Version : 00.90.01
  Creation Time : Mon Nov 21 18:27:53 2005
     Raid Level : raid1
     Array Size : 104320 (101.89 MiB 106.82 MB)
    Device Size : 104320 (101.89 MiB 106.82 MB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 1
    Persistence : Superblock is persistent

    Update Time : Wed Aug  1 21:41:11 2007
          State : clean
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0

           UUID : 38c0db3c:5b2ad44d:86e2fb95:39a1b377
         Events : 0.7744

    Number   Major   Minor   RaidDevice State
       0       8        1        0      active sync   /dev/sda1
       1       8       17        1      active sync   /dev/sdb1
[root@server ~]# mdadm --detail --verbose /dev/md  
mdadm: cannot open /dev/md: No such file or directory
[root@server ~]# mdadm --detail --verbose /dev/md1
/dev/md1:
        Version : 00.90.01
  Creation Time : Mon Nov 21 18:27:53 2005
     Raid Level : raid1
     Array Size : 104320 (101.89 MiB 106.82 MB)
    Device Size : 104320 (101.89 MiB 106.82 MB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 1
    Persistence : Superblock is persistent

    Update Time : Wed Aug  1 21:41:11 2007
          State : clean
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0

           UUID : 38c0db3c:5b2ad44d:86e2fb95:39a1b377
         Events : 0.7744

    Number   Major   Minor   RaidDevice State
       0       8        1        0      active sync   /dev/sda1
       1       8       17        1      active sync   /dev/sdb1


Code: [Select]
[root@server ~]# mdadm --detail --verbose /dev/md1
/dev/md1:
        Version : 00.90.01
  Creation Time : Mon Nov 21 18:27:53 2005
     Raid Level : raid1
     Array Size : 104320 (101.89 MiB 106.82 MB)
    Device Size : 104320 (101.89 MiB 106.82 MB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 1
    Persistence : Superblock is persistent

    Update Time : Wed Aug  1 21:41:11 2007
          State : clean
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0

           UUID : 38c0db3c:5b2ad44d:86e2fb95:39a1b377
         Events : 0.7744

    Number   Major   Minor   RaidDevice State
       0       8        1        0      active sync   /dev/sda1
       1       8       17        1      active sync   /dev/sdb1
[root@server ~]# mdadm --detail --verbose /dev/md2
/dev/md2:
        Version : 00.90.01
  Creation Time : Mon Nov 21 18:26:51 2005
     Raid Level : raid1
     Array Size : 143267584 (136.63 GiB 146.71 GB)
    Device Size : 143267584 (136.63 GiB 146.71 GB)
   Raid Devices : 2
  Total Devices : 1
Preferred Minor : 2
    Persistence : Superblock is persistent

    Update Time : Thu Aug  2 09:41:34 2007
          State : clean, degraded
 Active Devices : 1
Working Devices : 1
 Failed Devices : 0
  Spare Devices : 0

           UUID : 7ed9fc4a:085df8f6:4737e5b4:f7188a71
         Events : 0.22491650

    Number   Major   Minor   RaidDevice State
       0       8        2        0      active sync   /dev/sda2
       1       0        0        -      removed


...Since there seems to be people around here with much better experience on RAID i thought someone could give me a hint on how to solve this one.

(The server, BTW, is 7.2 with latest updates applied)

Offline NickR

  • *
  • 283
  • +0/-0
    • http://www.witzendcs.co.uk/
RAID Management Question
« Reply #14 on: August 02, 2007, 10:39:41 AM »
You need to add the removed partition sdb2 back into md2


#mdadm /dev/md2 -a /dev/sdb2

then doing

#mdadm --detail --verbose /dev/md2

should show 2 active disks and a remirror in operation
--
Nick......