JFFS3 memory consumption
Josh Boyer
jdub at us.ibm.com
Wed Jan 26 12:06:44 EST 2005
On Wed, 2005-01-26 at 09:57, Artem B. Bityuckiy wrote:
> On Wed, 26 Jan 2005, David Woodhouse wrote:
>
> > On Wed, 2005-01-26 at 08:44 +0000, Artem B. Bityuckiy wrote:
> > > K1 is decreased by the Ferenc Havasi's "summary" patch well. Would be
> > > nice to decrease K2 either. Any Ideas how to do so?
> >
> > For a start, we can increase the maximum node size. Limiting it to
> > PAGE_SIZE means we have a larger number of nodes than we'd otherwise
> > have. If we quadruple that, we'll probably end up with about a third as
> > many nodes.
> I thought about to have several 4K chunks within one node. Thus we will
> keep reading fast but will have few nodes number. Didi you mean this?
If you use larger "chunks" you'll get better overall compression. E.g.
a 64KiB node should compress better than a 4KiB node. This is partially
how squashfs achieves better compression that cramfs.
So the end result is better compression and fewer nodes.
josh
More information about the linux-mtd
mailing list