From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from lists.gentoo.org (pigeon.gentoo.org [208.92.234.80]) by finch.gentoo.org (Postfix) with ESMTP id 036241381F3 for ; Thu, 27 Jun 2013 18:53:50 +0000 (UTC) Received: from pigeon.gentoo.org (localhost [127.0.0.1]) by pigeon.gentoo.org (Postfix) with SMTP id 3957BE0946; Thu, 27 Jun 2013 18:53:38 +0000 (UTC) Received: from plane.gmane.org (plane.gmane.org [80.91.229.3]) (using TLSv1 with cipher AES256-SHA (256/256 bits)) (No client certificate requested) by pigeon.gentoo.org (Postfix) with ESMTPS id 0996FE087D for ; Thu, 27 Jun 2013 18:53:36 +0000 (UTC) Received: from list by plane.gmane.org with local (Exim 4.69) (envelope-from ) id 1UsHJs-00059u-7V for gentoo-amd64@lists.gentoo.org; Thu, 27 Jun 2013 20:53:24 +0200 Received: from ip68-231-22-224.ph.ph.cox.net ([68.231.22.224]) by main.gmane.org with esmtp (Gmexim 0.1 (Debian)) id 1AlnuQ-0007hv-00 for ; Thu, 27 Jun 2013 20:53:24 +0200 Received: from 1i5t5.duncan by ip68-231-22-224.ph.ph.cox.net with local (Gmexim 0.1 (Debian)) id 1AlnuQ-0007hv-00 for ; Thu, 27 Jun 2013 20:53:24 +0200 X-Injected-Via-Gmane: http://gmane.org/ To: gentoo-amd64@lists.gentoo.org From: Duncan <1i5t5.duncan@cox.net> Subject: [gentoo-amd64] Re: Can initrd and/or RAID be disabled at boot? Date: Thu, 27 Jun 2013 18:53:08 +0000 (UTC) Message-ID: References: Precedence: bulk List-Post: List-Help: List-Unsubscribe: List-Subscribe: List-Id: Gentoo Linux mail X-BeenThere: gentoo-amd64@lists.gentoo.org Reply-to: gentoo-amd64@lists.gentoo.org Mime-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-Complaints-To: usenet@ger.gmane.org X-Gmane-NNTP-Posting-Host: ip68-231-22-224.ph.ph.cox.net User-Agent: Pan/0.140 (Chocolate Salty Balls; GIT 368aae4 /usr/src/portage/src/egit-src/pan2) X-Archives-Salt: 4c47e166-3d9a-47a7-b739-adc0f300a582 X-Archives-Hash: e8382a49f1a89e90e160d2fa0bc4bdf9 Mark Knecht posted on Tue, 25 Jun 2013 15:51:14 -0700 as excerpted: > This is related to my thread from a few days ago about the > disappointing speed of my RAID6 root partition. The goal here is to get > the machine booting from an SSD so that I can free up my five hard > drives to play with. FWIW, this post covers a lot of ground, too much I think to really cover in one post. Which is why I've delayed replying until now. I expect I'll punt on some topics this first time thru, but we'll see how it goes... > SHORT SUMMATION: I've tried noninitrd and noraid in the kernel line of > grub.conf but I keep booting from old RAID instead of the new SSD. > What am I doing wrong? > > What I've done so far: > > 1) I've removed everything relatively non-essential from the HDD-based > RAID6. It's still a lot of data (40GB) but my Windows VMs are moved to > an external USB drive as is all the video content which is on a second > USB drive so the remaining size is pretty manageable. OK... > 2) In looking around for ways to get / copied to the SDD I ran across > this Arch Linux page called "Full System Backup with rsync": > > https://wiki.archlinux.org/index.php/Full_System_Backup_with_rsync > Basically it boiled down to just a straight-forward rsync command, but > what I liked about the description what that it can be done on a live > system. The command in the page is > > rsync -aAXv /* /path/to/backup/folder > --exclude={/dev/*,/proc/*,/sys/*,/tmp/*,/run/*,/mnt/*,/media/*,/lost +found} > > which I have modified to > > rsync -avx /* /path/to/backup/folder > --exclude={/dev/*,/proc/*,/sys/*,/tmp/*,/run/*,/mnt/*,/media/*,/lost +found} > > because I don't use (AFAICT) any of the ACL stuff and the command simply > wouldn't do anything. For ACL you're probably correct. But you might be using xattributes. Do you have any of the following USE flags turned on: caps xattr filecaps ? (Without going into an explanation of the specific distinction between the USE flags above, particularly caps and filecaps.) What filesystem do you use on / (and /usr if separate), and if appropriate, are the extended attributes and security label kernel options enabled for it? For example, here I have ext4, reiserfs and btrfs enabled and use or have used them on my various root filesystems, as well as tmpfs with the appropriate options since I have PORTAGE_TMPDIR pointed at tmpfs (and also devtmpfs needs some of the options): zgrep 'REISER\|EXT4\|TMPFS\|BTRFS' /proc/config.gz CONFIG_DEVTMPFS=y CONFIG_DEVTMPFS_MOUNT=y CONFIG_EXT4_FS=y CONFIG_EXT4_USE_FOR_EXT23=y # CONFIG_EXT4_FS_POSIX_ACL is not set CONFIG_EXT4_FS_SECURITY=y # CONFIG_EXT4_DEBUG is not set CONFIG_REISERFS_FS=y # CONFIG_REISERFS_CHECK is not set # CONFIG_REISERFS_PROC_INFO is not set CONFIG_REISERFS_FS_XATTR=y # CONFIG_REISERFS_FS_POSIX_ACL is not set CONFIG_REISERFS_FS_SECURITY=y CONFIG_BTRFS_FS=y # CONFIG_BTRFS_FS_POSIX_ACL is not set # CONFIG_BTRFS_FS_CHECK_INTEGRITY is not set # CONFIG_BTRFS_FS_RUN_SANITY_TESTS is not set # CONFIG_BTRFS_DEBUG is not set CONFIG_TMPFS=y CONFIG_TMPFS_POSIX_ACL=y CONFIG_TMPFS_XATTR=y tmpfs only has ACL on for devtmpfs (and I'm not sure I need that, but to avoid both security issues and broken device functionality...). The others don't have that on, but where appropriate, they have XATTR on, as well as FS_SECURITY. (Again, this is really only surface coverage, here. TBH I don't fully understand the depths myself, certainly not well enough to be comfortable discussing it in depth, tho I'm reasonably sure I have the options I want enabled, here.) The deal is that file capacities in one form or another can be used to avoid having to SETUID root various executables that would otherwise need it, which is a good thing since that reduces the security vulnerability window that SETUID root otherwise opens, often necessarily. And these file capacities are implemented using xattrs. So if your system is setup to use them, a good thing from a security perspective but somewhat complicated by the kernel config requirements in addition to the USE flags, you'll probably want to use the -X option, tho you should still be safe without -A (tho it shouldn't hurt). However, the penalty for NOT using -X, provided you're not using xattrs for anything else, should simply be that you'll need to become root to run some commands that would otherwise be runnable without root (with the corresponding open security window, should it be possible for a cracker to get those commands running as root to do something unintended). So the potential cost of getting it wrong is actually quite limited, unless you happen to be the target of a cracker with both good timing and a reasonable skill level, as well. And of course if you have only part of the pieces above enabled, say the appropriate filesystem options in the kernel but not the USE flags, or the reverse, then you're not covered and the rsync options won't matter in any case. But the -AX options shouldn't do any harm in any case, so here I'd have just left them on, making it -avxAX. Meanwhile, while I always see people worried about copying a live filesystem around, I've never had a problem here simply doing a cp --archive, or the equivalent in mc (midnight commander, ncurses-based commander-style dual-pane file manager). What I do for root is use a root-bind script: #!/bin/bash me=${0##*/} case $me in rootbind) mount --bind / /mnt/rootbind;; rootbindu) umount /mnt/rootbind;; *) echo rootbind: bad call; exit 1;; esac (That allows the script to be called rootbind, with a symlink to it called rootbindu, that does the corresponding umount.) What a bind-mount does is mount an already mounted filesystem at a different mountpoint. In particular, it does NOT do recursive mounts (tho there's another mount option that copies the full mount tree, it's just not what I want here), so what I'm using it for here is to get a "clean" copy of the rootfs, WITHOUT other filesystems such as /dev and /home mounted on top. Then I can do a nice clean cp --archive of my rootfs to a (normally freshly formated, so cp and rsync would accomplish the same thing) backup root, and that's what I've used for backup, for years. And I test those backups too, and occasionally reboot to the backup and do a clean mkfs and copy back from the backup to the normally working copy too, just to take care of fragmentation and any possibility of unresolved filesystem damage or bitrot that might have set in, as well as ensuring that I can switch to the backups for operational use by actually doing so. So I know the technique works for me. Now if I was running some active database that was continuing to update as I did my copy, yes, that would be a problem, and I'd want to do a snapshot or otherwise "freeze" the live filesystem in ordered to get a good self-consistent copy. But, for root anyway, unless I'm trying to do an emerge --update in the background or something at the same time (and why would I, both the copy and the update could be trying to access the filesystem at once, slowing both down, and it needlessly complicates things, so there's no purpose to doing so), a simple cp --archive of the live filesystem, from the bind-mount so I get JUST the root filesystem, no more no less, is sufficient. For /home, there's a /bit/ more concern, say with the firefox sqlite databases, if I'm browsing at the same time I'm trying to do the backup. However, that's simple enough to avoid. Just don't do anything that's going to be actively changing the filesystem at the same time I'm trying to make an accurate backup of it. Of course with your VMs that's a bit different story, rather like the active database case. A snapshotting filesystem (like btrfs) or sub- filesystem block-device layer (like lvm2) can be used here, taking the snapshot and copying it while the activity continues on the live filesystem, or, likely a simpler solution for those where it's possible, just do the copy when the database/vms aren't active and in use. But unless your vms/databases are files on your rootfs, that shouldn't be a problem with the rootfs backup, in any case. And if they are and you can't shut down the vms/databases for long enough to do a backup, I'd personally have questions of the strategy that put them on rootfs to begin with, but whatever, THAT is when you'd need to worry about taking an accurate copy of the live rootfs, but ideally, that's not a case you need to worry about, and indeed, from what I've read it's not a problem in your case at all. =:^) > I ran this command the first time to get 98% of everything copied while > in KDE, but before I moved forward I exited KDE, stopped X and ran it as > root from the console. After the second run it didn't pick up any new > file changes so I suspect it's pretty close to what I'd get dealing with > a Live CD boot. (COMMENTS?) As the above commentary should indicate, if anything I think you're being overly cautious. In the vast majority of cases, a simple cp --archive, or your equivalent rsync, should be fine. The caveat would be if you were trying to backup the vms while they were in operation, but you've taken care of that separately, so (with the possible caveat about file capacities and xattrs) I believe you're good to go. > 3) I added a new boot options in grub.conf: (Long) Note in passing: You should probably look into upgrading to grub2 at some point. Now may not be a good time, as you've got a lot on your plate right now as it is, but at some point. Because while there's a bit of a learning curve to getting up and running on grub2, it's a lot more flexible than grub1, with a lot more troubleshooting possible if you're not getting the boot you expect, and direct support of all sorts of fancy stuff like mdadm, lvm2, btrfs, zfs, etc, as well as an advanced command- line shell much like sh/bash itself, so it's very possible to browse your whole filesystem directly from inside grub, as I said, making troubleshooting **MUCH** easier. Plus its scripting (including if/then conditionals and variable handling much like bash) and menu system make all sorts of advanced boot configs possible. And while I'm at it, I'd strongly recommend switching to gpt partitioning from the old mbr style partitions, either before switching to grub2 or at the same time. GPT is more reliable (checksummed partitioned table with two copies, one at the beginning and one at the end of the device, unlike mbr with just one, with no checking, that if it goes bad...) and less complicated (no primary/secondary/logical partition distinction, up to 128 partitions handled by default, with a possibility for even more if you setup a larger gpt than mbr. Plus, it allows partition labels much like the filesystem labels people already use, only on the partitions themselves, so they don't change with the filesystem. That in itself makes things much easier, since with labels it's much easier to keep track of what each partition is actually for. The reason I recommend switching to gpt before switching to grub2, is that gpt has a special BIOS-reserved partition type, that grub2 can make use of to store its core (like grub1's stage-1.5 and 2), making the grub2 administration and updates less problematic than they might be otherwise. (This of course assumes bios, not efi, but grub2 works with efi as well, I'm just not familiar with its efi operation, and besides, efi folks are likely to already be running grub2 or something else, instead of legacy grub1, so it's likely a reasonably safe assumption.) Actually, when I switched to gpt here, while still on grub1, I was forward thinking enough to setup both a bios-reserved partition, and an efi-reserved partition, even tho neither one was used at the time. They were small (a couple MB for the BIOS partition, just under 128 MB for the efi partition, so the both fit in 128 MB, an eighth of a gig). Then I upgraded to grub2 and it found and used the gpt bios partition without issue, instead of having to worry about fitting it in slack space before the first partition or whatever. The efi-reserved partition is still unused, but it's there in case I upgrade to efi on this machine (I doubt I will as I have no reason to), or decide to fit the existing disk into a new machine at some point, without full repartitioning. (FWIW, I use gptfdisk, aka gdisk, as my gpt-partitioner analogous to fdisk. However, gparted has supported gpt for awhile, and even standard fdisk, from the util-linux package, has (still experimental) gpt support now. Tho the cfdisk variant (also from util-linux) doesn't have gpt support yet, but cgdisk, from the gptfdisk package, does, and that's the executable from the gptfdisk package I tend to use here. (I use gdisk -l to spit out the partition list on the commandline, similar to cat-ing a file. That's about it. I use cgdisk for actual gpt partition table editing.)) It's just that reading your post, I'm translating to grub2 in my head, and thinking how much simpler grub2 makes troubleshooting, when you can effectively browse all hard drives in read-only mode directly from grub, not only browsing around to know for sure that a particular partition is the one you need, but paging thru various files in the kernel Documentation dir, for instance, to get options to plug in on the kernel commandline in grub, etc. It really does make troubleshooting early boot problems MUCH easier, because grub2 simply gives you far more to work with in terms of troubleshooting tools available to use at the grub prompt. The one caveat for gpt is for people multi-booting to other than Linux. >From what I've read, MS does support GPT, but with substantially less flexibility (especially for XP, 7 is better) than Linux. I think it can install to either, but switching from one to the other without reinstalling is problematic, or something like that, whereas with Linux it's simply ensuring the appropriate support is configured into (or available as modules if you're running an initr*) the kernel. (I have little idea how Apple or the BSDs work with GPT.) But while you do MS, AFAIK it's all in VMs, so that shouldn't be a problem for you, so gpt should be fine. And of course grub2 should be fine as well, gpt or not, but based on my experience, gpt makes the grub2 upgrade far easier, at least as long as there's a bios-reserved partition setup in gpt already, as there was here when I did my grub2 upgrade, since I'd already done the gpt upgrade previously. But as I said, now may not be the best time to think about that as you have enough on your plate ATM. Maybe something for later, tho... Or maybe consider doing gpt now, since you're repartitioning now, and grub2 later... (grub1 menu entries:) > title fastVM 3.8.13-gentoo using LABEL (SSD, initramfs in kernel) > root (hd5,0) > kernel (hd0,0)/boot/bzImage-3.8.13-gentoo root=LABEL=fastVM video=vesafb > vga=0x307title > > fastVM 3.8.13-gentoo using LABEL (SSD, initramfs in kernel) > root (hd5,0) > kernel (hd0,0)/boot/bzImage-3.8.13-gentoo root=/dev/sda1 video=vesafb > vga=0x307 I'll assume that "vga=0x307title" is a typo, and that "title" starts the second menu entry... ... Making the difference between the two entries the root=LABEL=fastVM, vs root=/dev/sda1 > I am relatively confident that (hd5,0) is the SSD. I have 6 drives in > the system - the 5 HDDs and the SSD. The 5 hard drives all have multiple > partitions which is what grub tells me using tab completion for the line > > root(hdX, > > Additionally the SDD has a single partition to tab completion on > root(hd5 finishes with root(hd5,0). I used /dev/sda as that's how it's > identified when I boot using RAID. This is actually what triggered the long grub2 note above. "Relatively confident", vs. knowing, because with grub2's mdadm support, you can (read-only) browse all the filesystems in the raid, etc (lvm2, etc, if you're using that...), as well. So you know what's what, because you can actually browse it, direct from the grub2 boot prompt. However, while my grub1 knowledge is getting a bit rusty now, I think you're mixing up grub's root(hdX,Y) notation, which can be thought of as sort of like a cd in bash, simply changing the location you're starting from if you don't type in the full path, with the kernel's root= commandline option. Once the kernel loads (from hd0,0 in both entries), its root= line may have an entirely DIFFERENT device ordering, depending on the order in which it loaded its (sata chipset, etc) drivers and the order the devices came back in the device probes it did as it loaded them. That's actually why kernel devs and udev folks plus many distros tend to recommend the LABEL= (or alternatively UUID=) option for the kernel's root= commandline option, these days, instead of the old /dev/sdX style, because in theory at least, the numbering of /dev/sdX devices can change arbitrarily. In fact, on most home systems with a consistent set of devices appearing at boot, the order seldom changes, and it's *OFTEN* the same as the order as seen by grub, but that doesn't HAVE to be the case. Of course the monkey wrench in all this is that as far as I'm aware anyway, the LABEL= and UUID= variants of the root= kernel commandline option *REQUIRE* an initr* with working udev or similar (I'm not sure if busybox's mdev supports LABEL=/UUID= or not), which might well be a given on binary-based distros that handle devices using kernel modules instead of custom built-in kernel device support, and thus require an initr* to handle the modules load anyway, but it's definitely *NOT* a given on a distro like gentoo, which strongly encourages building from source and where many, perhaps most users, use a custom-built kernel with the drivers necessary to boot builtin, and thus may well not require an initr* at all. For initr*-less boots, AFAIK root=/dev/* is the only usable alternative, because the /dev/disk/by-*/ subdirs that LABEL= and UUID= depends on are udev userspace created, and those won't be available for rootfs mount in an initr*-less boot. > Now, the kernel has the initrd built into it so if it cannot be > turned off I guess I'll try building a new kernel without it. However I > found a few web pages that also said RAID could be disabled using a > 'noraid' option which I thought should stop the system from finding the > exiting RAID6 but no luck. FWIW, the best reference for kernel commandline options is the kernel documentation itself. Sometimes you need more, but that's always the place to look first. Specifically, $KERNELDIR/Documentation/kernel-parameters.txt , for the big list in one place, with additional documentation often provided in the various individual files documenting specific features. kernel-parameters.txt lists noinitrd: noinitrd [RAM] Tells the kernel not to load any configured initial RAM disk. So that should work. It doesn't say anything about it not working with a built-in initramfs, either, so if it doesn't, there's a bug in that it either should say something about it, or it should work. FWIW, depending on what initramfs creation script you're using and its content, you should be able to tell whether the initramfs activated or not. Here, I /just/ /recently/ started using dracut, since it seems multi- device btrfs as root doesn't work reliably otherwise, and that's what I'm using as my rootfs now (btrfs raid1 mode on dual SSDs, I could only get it to mount the dual-device btrfs raid1 in degraded mode, seeing only one of the two devices, without the btrfs device scan in the initramfs, tho a google says some people have it working, ). But even booting to my more traditional reiserfs rootfs backups on the "spinning rust", where booting from the initramfs isn't mandatory, I can tell whether the initramfs was loaded or not by the boot-time console output. Among other things, if the initramfs is loaded and run, then /proc and /run are already loaded when the openrc service that would normally mount them gets run, because the initramfs mounted them. But apparently the initramfs mounts at least /run with different permissions, so openrc mentions that it's changing permissions on /run when it runs after the initramfs, but simply mounts it with the permissions it wants, when the initramfs hasn't run. But unfortunately, I've not actually tried the noinitrd kernel commandline option, so I can't VERIFY that it works here, with my now builtin initramfs. I'll have to reboot to try that, and will try to get back to you on that. (Note to self. Test the root=LABEL with initramfs-less boot too, while I'm at it.) If you're using a dracut-created initr*, then there's several other helpful kernel commandline options that it hooks. See the dracut.comandline manpage for the full list, but rd.break and its rd.break= variants allow dropping to the initr*'s builtin shell (AFAIK dash by default for dracut, but bash is an option... which I've enabled) at various points, say right before or right after the initr*'s udev runs, right before mounting the real rootfs, or right before the final switchroot and start of the init on the real rootfs. If you're using some other initr* creator, obviously you'd check its documentation for similar options. I know rd.break works here, as I tested it while I was figuring out how to work this new-to-me initramfs thing. And it's obvious that I'm in the initr*'s bash, because its shell prompt isn't anything like my customized shell prompt. Meanwhile, I DO NOT see "noraid" listed in kernel-parameters.txt, altho that doesn't mean it doesn't (or didn't at some point) exist. I DO see a couple raid-options, md= and raid=, however, with references to Documentation/md.txt. Based on the md.txt file, it appears raid=noautodetect is the option you're looking for. This also matches my now slightly rusty recollection from when I ran mdraid before. noraid didn't look quite right, but raid=noautodetect looks much closer to what I remember. (If you're using dracut-based initr*, there's a similar option for it, rd.auto, rd.auto=(0|1), that defaults off with current versions of dracut, according to the dracut.cmdline manpage. That governs autoassembly of raid, lvm, etc. But since it already defaults off, unless you're running an old version where that defaulted on, or have it as part of your builtin commandline as configured in your kernel or something, that shouldn't be your problem.) > Does anyone here have any ideas? fdisk info follows at the end. > Ask for anything else you want to see. > > If I can get to booting off the SSD then for the next few days I > could build different RAIDs and do some performance testing. Hmm... This didn't turn out to be so hard to reply to after all. Maybe because I kept my initr* remarks to dracut-based, which is all I know anyway... Some other remarks... FWIW, if you're running an md-raid rootfs, at least with gpt and a dedicated bios partition, installing grub2 is easier than installing or updating grub1, as well. I remember the pain it was to install grub1 to each of the four drives composing my raid, back when I had that setup, in particular, the pain of trying to be sure I was installing to the physical drive I /thought/ I was installing to, while at the same time ensuring it was actually pointed at the /boot on the same drive, not at the /boot on a different drive, so that if that drive was the only one I had left, I could still boot from it. The problem was that because I was on mdraid, grub1 was detecting that and I had to specify the physical device one way to tell it where to install stage1, and a different way to tell it where to put stage2 in /boot. With grub2, things were so much easier that I had trouble believing I'd actually installed it already. But I rebooted and it worked just fine, so I had. Same thing when I switched to the pair of ssds with btrfs in raid1 mode as my rootfs. I installed to the first one... and thought surely there was another step I had missed, but there it was. After reboot to test, I installed to the second one, and rebooted to it (using the boot selector in the BIOS) to test. All fine. =:^) Of course part of that, again, is due to using gpt with the reserved bios partition for grub to put its stage2 core in, quite apart from what it puts in /boot. I suppose I'd have had a similar problem as I did with grub1, if I was still using mbr or didn't have a reserved bios partition in my gpt layout, and grub had to stick the stage2 core either in slack space before the first partition (if there was room to do so), or in /boot itself, and hope the filesystem didn't move things around afterward (which reiserfs did do a couple of times to me back with grub1, tho it wasn't usually a problem). I glossed over what to do with non-dracut-based initr*, as I've not used anything other than dracut and direct no-initr*, and dracut's only very recently. However, I'd be quite surprised if others didn't have something similar to dracuts rd.break options, etc, and you certainly should be able to tell whether the initr* is running or not, based on the early-boot console output. Of course, whether you're /familiar/ enough with that output or not to tell what's initr* and what's not, is an entirely different question, but if you know well what one looks like, the other should be enough different that you can tell, if you look closely. >From the initrd, it should be possible to mount something else other than the old raid as rootfs, and by that time, you'll have the kernel populated /dev tree to work with as well as possibly the udev populated disk/by-* trees, so finding the right one to mount shouldn't be an issue -- no worries about kernel device order not matching grub device order, because you're past grub and using the kernel already, by that point. That was definitely one of the things I tested on my dracut-based initr*, that from within the initr* (using for instance rd.break=pre-mount to drop to a shell before the mount), I could find and mount backup root filesystems, should it be necessary. >From within the initrd, you should be able to mount using label, uuid or device, any of the three, provided of course that udev has populated the disk/by-label and by-uuid trees, and I could certainly mount with either label or device (I didn't try uuid), using my dracut-based initramfs, here. So really, you shouldn't need the noinitrd option for that. Tho as long as your selected rootfs doesn't /require/ an initr* to boot (as my multi- device btrfs rootfs seems to here), you should be able to boot to it with an appropriate kernel commandline root= option, with or without the initr*. > c2RAID6 ~ # fdisk -l > > Disk /dev/sda: 128.0 GB [snip] > > Device Boot Start End Blocks Id System > /dev/sda1 2048 250069679 125033816 83 Linux > > Disk /dev/sdb: 500.1 GB > > Device Boot Start End Blocks Id System > /dev/sdb1 * 63 112454 56196 83 Linux > /dev/sdb2 112455 8514449 4200997+ 82 Linux swap > /dev/sdb3 8594775 976773167 484089196+ fd Linux raid > > Disk /dev/sdc: 500.1 GB > > Device Boot Start End Blocks Id System > /dev/sdc1 * 63 112454 56196 83 Linux > /dev/sdc2 112455 8514449 4200997+ 82 Linux swap > /dev/sdc3 8594775 976773167 484089196+ fd Linux raid > > Disk /dev/sde: 500.1 GB > > Device Boot Start End Blocks Id System > /dev/sde1 2048 8594774 4296363+ 83 Linux > /dev/sde3 8594775 976773167 484089196+ fd Linux raid > > Disk /dev/sdf: 500.1 GB > > Device Boot Start End Blocks Id System > /dev/sdf1 2048 8594774 4296363+ 83 Linux > /dev/sdf3 8595456 976773167 484088856 fd Linux raid > > Disk /dev/sdd: 500.1 GB > > Device Boot Start End Blocks Id System > /dev/sdd1 * 63 112454 56196 83 Linux > /dev/sdd2 112455 8514449 4200997+ 82 Linux swap > /dev/sdd3 8594775 976773167 484089196+ fd Linux raid > > Disk /dev/md3: 1487.1 GB > > c2RAID6 ~ # -- Duncan - List replies preferred. No HTML msgs. "Every nonfree program has a lord, a master -- and if you use the program, he is your master." Richard Stallman