bug-parted
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: Concern for my data on large partition


From: Space Ship Traveller
Subject: Re: Concern for my data on large partition
Date: Wed, 19 Aug 2009 04:43:27 +1200

Dear Bill

It is great that you are experimenting with file systems and large partitions. This is an area where things start to get quite complex and there is no single best practise solution. Linux is a great tool but is by no means the only option. Some higher end NAS products and SAN products may provide a better and easier to administer solution. 3ware make good cards, but if performance is not a high priority, Linux software raid is also fairly capable.

I noticed a few things:
- msdos partition format is actually limited to 2TB partitions. So you might want to try using GPT (use mklabel gpt).
- fdisk doesn't support large partitions. you should be exclusively using parted at this point.

There is lots of information about creating large partitions on linux:

http://www.google.com/search?client=safari&rls=en-us&q=gpt+partitions+linux&ie=UTF-8&oe=UTF-8

Alternatively you might want to consider using LVM. Given the size of the volume, it might make things easier to manage.

http://www.google.com/search?hl=en&client=safari&rls=en-us&q=large+partitions+linux+LVM&aq=f&oq=&aqi=



(I'm sorry if the following is offensive, but I feel it must be said - this is supposed to be constructive criticism!)

I noticed that you are working for a lab and that you are the "Sr. Technical Advisor". As a consultant who deals with the deployment of various storage technologies on a regular basis, the story below is personally frightening to me..

You mention that you've already lost data, and that you've decided to move to RAID6 because of this... RAID is not a backup solution. It never was and never will be. Using RAID6 is not going to improve the situation very much. Some kind of frequent reliable off-line backup is the bare minimum in my mind for actual data backup. If data is important (and it sounds like you have a fair amount of it at ~5.5TB) you need to have a policy in place to manage it. To not have this in place when dealing with important data might be construed as negligence, especially in the event of data loss. How much trouble is it if you loose the data? Does it represent man-years or man-hours?

The [apparent] lack of experience could potentially cause a lot of problems down the road. For your own peace of mind, you might want to seriously consider hiring a consultant who specialises in storage if what you are putting together is intended for production - even it is to simply confirm you are on the right track both reliability and performance wise.

Other things you might want to consider:
Are you using a UPS to protect the system including hard-drives?
Is the unit located in a secure cabinet in a secure room? [is the data confidential?]
Is there suitable air conditioning and environmental monitoring?
Is there suitable fire protection?
Is it possible to have an online backup to a different building/lab on campus?

I do wish you the best of luck but "thar be the dragons".

Kind regards,
Samuel

On 19/08/2009, at 3:56 AM, Sherman, William R wrote:

Hello,

I have a largish (~5.5TB) partition onto which I thought I had successfully
created a filesystem -- and have been using it, but recently I've been
given cause for concern that there are grave issues that have the potential
for the loss of a large amount of data.

I was made aware of my potentially dangerous situation after rebooting
my machine and finding that not only did my large partition not mount,
but that the  partition was listed by the OS as only begin about 1.5TB.

I managed to convince the system that the partition was okay, and it
mounts again, but I get this error (which I also get when rebooting):
 "Aug 14 19:07:33 angelico kernel: I/O error in filesystem ("sdb1") meta-data dev sdb1 block 0x2ae7bffff       ("xfs_read_buf") error 5 buf count 512"

I'm convinced that this error is related to the incongruous partition
size issue that I am seeing.


For some background information, here are some details about my system
and what I learned when first creating the partition:

I am running Fedora-11 with a 3ware 9650SE RAID controller card with
8 Seagate 1TB drives configured as a ~5.8TB RAID-6 system.  Using the
RAID bios controls, I configured the system to have two logical units.
One unit is for the OS (or actually OSes) and is only 96GB, the other
is for all the data, and is (or is supposed to be) ~5.7TB.

The 96GB unit shows up as drive "/dev/sda" and I've partitioned it
into four partitions to hold multiple OSes (though thus far, I've
only put Fedora-11 on the machine).  The 2nd logical unit is referred
to as "/dev/sdb", and I've only created one partition to consume
the entire unit.

Of course the reason I've created the RAID-6 system to for the protection
of my data -- got burned once with RAID-5 when I was under the gun to get
some data off a system with one bad drive -- a 2nd drive went bad, and
all was lost.  So, putting my data in peril is something I'm not
interested in (if that isn't a tautology).

When creating the partitions, I first tried to use the partition manager
that Fedora 11 provides, and it reported that it had successfully created
a "5623675 MB" partition.  But after the system came up, "/proc/partitions"
disagreed, reporting:
  # cat /proc/partitions
  major minor  #blocks  name

  8        0  100663295 sda
  8        1   33554432 sda1
  8        2   33554432 sda2
  8        3   12582912 sda3
  8        4          1 sda4
  8        5   20964352 sda5
  8       16 5758648320 sdb
  8       17 1463676507 sdb1

So somewhere the partition size got messed up.

I then went to the partitioning tool I'm familiar with: fdisk.  But
when creating the partition, it only allowed me to create a partition
of 267349 cylinders.  If I tried anything larger, it reported "Value
out of range."  The man page for fdisk didn't report anything about
partition size limitations, so not sure why that's failing.

I then tried the "cfdisk" program, and it seemed to accept my larger
partition size, but after writing the partition table and quitting,
"/proc/partitions" reported the same values as above -- no change.

I then tried "parted", but didn't find an option that would allow
for the "xfs" filesystem (or even ext3).

So I moved on to "sfdisk" which also seemed to accept my desired
partition size, but again after writing the partition table and
quitting, there was no change in the values of "/proc/partitions".

So I decided to give "parted" another try.  And it seemed to work,
though now when I look back at my notes there was something that I
should have been concerned about in the output.

The essential commands I gave to "parted" are:
  # parted /dev/sdb
  (parted) rm 1
  (parted) mkpart primary 0 -0
  (parted) print
  (parted) quit

At this point, I checked "/proc/partitions" and was happy to see
the results I had been shooting for:
  # cat /proc/partitions
  major minor  #blocks  name

  8        0  100663295 sda
  8        1   33554432 sda1
  8        2   33554432 sda2
  8        3   12582912 sda3
  8        4          1 sda4
  8        5   20964352 sda5
  8       16 5758648320 sdb
  8       17 5758648320 sdb1

From there I was able to create an 'xfs' filesystem and mount it:
  % df
  Filesystem           1K-blocks      Used Available Use% Mounted on
  /dev/sda1             33027952   6217352  25132880  20% /
  tmpfs                  5076716      1820   5074896   1% /dev/shm
  /dev/sr0                   388       388         0 100% /media/Bluebirds
  /dev/sdb1            5758517248      4320 5758512928   1% /mnt

And I proceeded happily along until after a reboot, I got the above
'meta-data' error, and "/proc/partitions" was again reporting the
partition to be only 1463676507 1K blocks.

So thinking my data might be lost anyway (and fortunately still having
the original data available), I did an experiment.  I tried just
doing a re-partition with "parted", and then mounting it -- and
that worked!
  # parted /dev/sdb
  (parted) print
  (parted) rm 1
  (parted) mkpart primary 0 -0
  (parted) print
  (parted) quit
  # cat /proc/partitions
  [...]
  8       17 5758648320 sdb1
  # mount /raid (renamed from previous)

And it worked, and the files were still there!

But then I looked back and noticed a discrepancy I'd missed before.
That final 'print' command to "parted" responded with:
  Model: AMCC 9650SE-8LP DISK (scsi)
  Disk /dev/sdb: 5897GB
  Sector size (logical/physical): 512B/512B
  Partition Table: msdos

  Number  Start  End     Size    Type     File system  Flags
  1      512B   1499GB  1499GB  primary  xfs

That "1499GB" isn't right.  In fact, I compared the two 'print'
outputs (before removing and recreating the partition, and after),
and they were identical.  So "parted" reported that it had created
the exact same partition as I started with -- but for some reason
unknown to me, it still causes "/proc/partitions" to recognize
the full disk as the partition, and the mount proceedure works
(whereas it failed before).

So, I'm at a loss for what the problem is.  I looked at the data
from the 3Ware RAID system, and it is reported as expected.

So, I'm hoping someone more familiar with "parted", and even
more familiar with Linux partitions might be able to shed some
light into the situation and illuminate my mind.

My appologies for the length of this email, but I figured all
this would be requested (and useful) in determining the situation.

  Thank you in advance,
  Bill

--
Bill Sherman
Sr. Technical Advisor
Advanced Visualization Lab
Indiana University
address@hidden








_______________________________________________
bug-parted mailing list
address@hidden
http://lists.gnu.org/mailman/listinfo/bug-parted


reply via email to

[Prev in Thread] Current Thread [Next in Thread]