Date: Wed, 21 Jan 2004 14:51:28 -0700 From: Shawn Ostapuk <flagg@slumber.org> To: freebsd-fs@freebsd.org Subject: Vinum / 5.2 cant mount/growfs/newfs terabyte Message-ID: <20040121215128.GB32719@slumber.org>
next in thread | raw e-mail | index | archive | help
Well its been quite a few months, and FreeBSD 5.2 got released today so I quickly went to see if my problem has been fixed yet, specially since the release notes were quite hopeful: "The sizes of some members of the statfs structure have changed from 32 bits to 64 bits in order to better support multi-terabyte filesystems." I was hoping that was the fix...but it wasnt... And now my problem... I use vinum and a number of IDE drives on two boxes, primary server and backup. Around 5 or 6 months ago the size grew to be needing a terabyte or more -- so I did the usual and added a new drive to the vinum configuration, went to growfs (which has saved me more time than you can imagine, thanks freebsd) and no luck, newfs, no luck, mount the old one without modifying? no luck...it just breaks after making it larger than 1 tb. I've tried this one two boxes, various drives configs. I can rearranage them all i want and they all work fine, but the second i make it larger than 1 terabyte, everything fails on it. bash-2.05b# vinum list 9 drives: D vinumdrive1 State: up /dev/ad1s1e A: 0/156327 MB (0%) D vinumdrive2 State: up /dev/ad2s1e A: 0/76316 MB (0%) D vinumdrive3 State: up /dev/ad4s1e A: 0/117239 MB (0%) D vinumdrive4 State: up /dev/ad5s1e A: 0/114470 MB (0%) D vinumdrive5 State: up /dev/ad6s1e A: 0/76292 MB (0%) D vinumdrive6 State: up /dev/ad7s1e A: 0/76292 MB (0%) D vinumdrive7 State: up /dev/ad8s1e A: 0/156327 MB (0%) D vinumdrive8 State: up /dev/ad9s1e A: 0/78159 MB (0%) D vinumdrive9 State: up /dev/ad11s1e A: 0/286102 MB (0%) 1 volumes: V pr0n State: up Plexes: 1 Size: 1110 GB 1 plexes: P vinum0.p0 C State: up Subdisks: 9 Size: 1110 GB 9 subdisks: S vinum0.p0.s1 State: up D: vinumdrive1 Size: 152 GB S vinum0.p0.s2 State: up D: vinumdrive2 Size: 74 GB S vinum0.p0.s3 State: up D: vinumdrive3 Size: 114 GB S vinum0.p0.s4 State: up D: vinumdrive4 Size: 111 GB S vinum0.p0.s5 State: up D: vinumdrive5 Size: 74 GB S vinum0.p0.s6 State: up D: vinumdrive6 Size: 74 GB S vinum0.p0.s7 State: up D: vinumdrive7 Size: 152 GB S vinum0.p0.s8 State: up D: vinumdrive8 Size: 76 GB S vinum0.p0.s9 State: up D: vinumdrive9 Size: 279 GB (no its not really porn stupid name i madeup too long ago to change =) All the drives work. In fact i've used two vinum configs at the same time using all the drives because of this terabyte limit :( Now with the above setup, vinum thinks everything is fine and looks fine as far as I can tell. And i've heard of vinum being used in greater than 1tb situations making me think there is no problem with it. But the second i try to do anything... bash-2.05b# mount /dev/vinum/pr0n mount: /dev/vinum/pr0n: unknown special file or file system ---- bash-2.05b# growfs /dev/vinum/pr0n growfs: rdfs: read error: 128: Invalid argument ---- bash-2.05b# newfs /dev/vinum/pr0n /dev/vinum/pr0n: 1137530.4MB (2329662200 sectors) block size 16384, fragment size 2048 using 6191 cylinder groups of 183.77MB, 11761 blks, 23552 inodes. newfs: can't read old UFS1 superblock: read error from block device: Invalid argument ---- No, it is NOT using ufs1, it is ufs2 -- in fact that code it hits in newfs only comes up when dealing with ufs2 partitions. The second i remove any drives that make it past the 1TB barrior i can mount/newfs it again no problems. Again, in any order and any drives on two different machines with a whole new set of drives. So thoughts anyone? is no one else in the world using vinum to concat some IDE drives past 1tb? i've heard of mount correctly mounting ufs2 partitions off hardware raids that are larger than 1tb no problem... it just seems to be some sort of problem with the two together...i'll be enternally grateful for anyone who can help me past this barrior, it is getting more and more difficult to work around separate partitions, etc -- and i see this becoming more and more common now that 300 gig drives are $299 retail... I should note also i've had roughly a dozen responses from my last post to questions recently and again to -fs/-questions about 6 months back who have the same problem wanting to know if i've found a way around it. Hopefully someone knows whats up :) Thanks, Shawn.
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?20040121215128.GB32719>