On Tue, Jun 05, 2007 at 03:23:50PM -0700, Michael Nishimoto wrote:
> David Chinner wrote:
> >On Wed, May 30, 2007 at 09:49:38AM -0700, Michael Nishimoto wrote:
> > > Hello,
> > >
> > > Has anyone done any work or had thoughts on changes required
> > > to reduce the total memory footprint of high extent xfs files?
> >Yes, it could, but that's a pretty major overhaul of the extent
> >interface which currently assumes everywhere that the entire
> >extent tree is in core.
> >Can you describe the problem you are seeing that leads you to
> >ask this question? What's the problem you need to solve?
> I realize that this work won't be trivial which is why I asked if anyone
> has thought about all relevant issues.
> When using NFS over XFS, slowly growing files (can be ascii log files)
> tend to fragment quite a bit.
Oh, that problem.
The issue is that allocation beyond EOF (the normal way we prevent
fragmentation in this case) gets truncated off on file close.
Even NFS request is processed by doing:
And so XFS truncates the allocation beyond EOF on close. Hence
the next write requires a new allocation and that results in
a non-contiguous file because the adjacent blocks have already
- NFS server open file cache to avoid the close.
- add detection to XFS to determine if the called is
an NFS thread and don't truncate on close.
- use preallocation.
- preallocation on the file once will result in the
XFS_DIFLAG_PREALLOC being set on the inode and it
won't truncate on close.
- append only flag will work in the same way as the
prealloc flag w.r.t preventing truncation on close.
- run xfs_fsr
Note - i don't think extent size hints alone will help as they
don't prevent EOF truncation on close.
> One system had several hundred files
> which required more than one page to store the extents.
I don't consider that a problem as such. We'll always get some
level of fragmentation if we don't preallocate.
> Quite a few
> files had extent counts greater than 10k, and one file had 120k extents.
you should run xfs_fsr occassionally....
> Besides the memory consumption, latency to return the first byte of the
> file can get noticeable.
Yes, that too :/
However, I think we should be trying to fix the root cause of this
worst case fragmentation rather than trying to make the rest of the
filesystem accommodate an extreme corner case efficiently. i.e.
let's look at the test cases and determine what piece of logic we
need to add or remove to prevent this cause of fragmentation.
SGI Australian Software Group