May I ask what gmailfs is? This is the first I've heard of it. On 10/10/05, Matt Alexander wrote: > I've basically accomplished what I wanted by using the loop device. Here's > a brief tutorial: > > I have 4 remote filesystems mounted. One is using sshfs, one is using > Samba from a Windows server, and 2 are using NFS from two different NetApp > filers. > > I then create a 1G file on each mount point and also one on my local file > system... > > dd if=/dev/zero of=/sshfs/matt0 bs=1M count=1024 > dd if=/dev/zero of=/samba/matt1 bs=1M count=1024 > dd if=/dev/zero of=/nfs1/matt2 bs=1M count=1024 > dd if=/dev/zero of=/nfs2/matt3 bs=1M count=1024 > dd if=/dev/zero of=/home/matt/matt4 bs=1M count=1024 > > I then create a loop device for each file... > > losetup /dev/loop0 /sshfs/matt0 > losetup /dev/loop1 /samba/matt1 > losetup /dev/loop2 /nfs1/matt2 > losetup /dev/loop3 /nfs2/matt3 > losetup /dev/loop4 /home/matt/matt4 > > I then create a RAID5 array from the loop devices... > > mdadm --create /dev/md0 --level=5 --raid-devices=5 /dev/loop0 /dev/loop1 > /dev/loop2 /dev/loop3 /dev/loop4 > > I then create a filesystem and mount it... > > mkfs.ext3 /dev/md0 > mount /dev/md0 /mnt/matt > > The performance actually isn't that bad. Copying a 2G file to /mnt/matt > took about 6 minutes. > My next step is to try it over the 'net using gmailfs to create my giant > filesystem in the sky. > ~M > > > > On 10/10/05, Dan Lund wrote: > > The RAID subsystem takes care of all of the backend issues like > > latency to disk and so forth. > > nbd is merely a block device, and from my experience if a disk even > > has a single timeout, the disk will be written as faulty by the > > subsystem and be set to [_]. > > raid6 is: > http://www.synetic.net/Tech-Support/Education/RAID6.htm > > Essentially it's data and parity striped across the array, and that > > itself has a parity. > > > > You can lose multiple disks with a raid6 setup and not lose data, I > > use it and I absolutely love it. (mine is raid6+1) > > > > On 10/10/05, Joseph Sinclair < plug-discuss@stcaz.net > wrote: > > > two questions: > > > 1) How does nbd deal with the differential latency issue? If latency > differs by too much a RAID system will end up with stripes on different > "disks" out of order, and things get REALLY messed up at that point. > > > 2) What is RAID 6? > > > > > > Dan Lund wrote: > > > > I've done work like this with the network block device as an > > > > experiment in several different ways. > > > > To put it in a nutshell I had a machine exporting a couple of nbd > > > > (network block devices), and I accepted them on another. They showed > > > > up as /dev/nbd/0, /dev/nbd/1, etc. > > > > I then made a raidtab that took them and set them into a RAID5 and had > > > > a hotspare. > > > > I've tested it with RAID1/5/5+1/6/6+1, made it failover, hot-added > > > > "drives", etc. > > > > > > > > It was pretty decent in throughput, and I was about ready to put > > > > together a turnkey solution for my work as an expandable disk > > > > subsystem. (on it's own gig backplane) I made sure it was on it's own > > > > gig backplane because the nbd devices are solely dependant on the > > > > network. If it so much as blips, your disks go away. > > > > RAID, as far as I know, only works on block devices. > > > > You could always check out PVFS, or Coda if your looking for something > > > > on the filesystem layer. I have far more faith in nbd though....... > > > > > > > > --Dan > > > > > > > > > > > > On 10/9/05, Matt Alexander wrote: > > > > > > > >>I'm wondering if anyone knows if this is possible... > > > >> > > > >> Take multiple remote filesystems such as NFS, gmailfs, Samba, sshfs, > and > > > >>layer a filesystem over the top to create one namespace. Ideally it > would > > > >>provide some fault tolerance/redundancy and improved performance by > using > > > >>the concept of RAID over the multiple connections. > > > >> > > > >> In reality, this new filesystem layer wouldn't care if the > filesystems are > > > >>remote or not. You could have... > > > >> > > > >> /mynfsmount > > > >> /mygmailfsmount > > > >> /myothergmailfsmount > > > >> /mysshfsmount > > > >> > > > >> ...and then a new mount point of... > > > >> > > > >> /myreallycoolmount > > > >> > > > >> ...and when you put files here, they're striped/mirrored over all the > > > >>previous mounts. > > > >> > > > >> Is this currently possible? If not, then perhaps I'll see if I can > make it > > > >>happen in my minuscule free time. I know there are a ton of potential > > > >>problems with this, but it'd be a fun project nonetheless. > > > >> Thanks, > > > >> ~M > > > --------------------------------------------------- > PLUG-discuss mailing list - > PLUG-discuss@lists.plug.phoenix.az.us > To subscribe, unsubscribe, or to change you mail settings: > http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss > > -- To exercise power costs effort and demands courage. That is why so many fail to assert rights to which they are perfectly entitled - because a right is a kind of power but they are too lazy or too cowardly to exercise it. The virtues which cloak these faults are called patience and forbearance. Friedrich Nietzsche --------------------------------------------------- PLUG-discuss mailing list - PLUG-discuss@lists.plug.phoenix.az.us To subscribe, unsubscribe, or to change you mail settings: http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss