Re: [PATCH 4/7] ppc64 - Specify amount of kernel memory at boot time

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, 2006-02-23 at 17:19 +0000, Mel Gorman wrote:
> On Thu, 23 Feb 2006, Dave Hansen wrote:
> >> +/* Initialise the size of each zone in a node */
> >> +void __init zone_sizes_init(unsigned int nid,
> >> +		unsigned long kernelcore_pages,
> >> +		unsigned long *zones_size)
> >
> > Minor nit territory: set_zone_sizes(), maybe?
> >
> 
> In this case, the choice of name is to match an x86 function that does 
> something very similar. If one had read through the x86 code and then saw 
> this function, it would set their expectations of what the code is 
> intended to do.

x86 is bad.  Try to do better. :)

> per-node. A node goes no ZONE_EASYRCLM pages if it is not large enough to 
> contain kernelcore_pages. That means that on a system with 2 nodes, 
> kernelcore=512MB will results in 1024MB of ZONE_DMA in total.
> 
> > Also, how do we want to distribute kernelcore memory over each node?
> > The way it is coded up for now, it will all be sliced out of the first
> > node.  I'm not sure that's a good thing.
> 
> It gets set in every node.

They hypervisor has a memory allocator which it uses to piece out memory
to various LPARs.  When things like partition memory resizing or reboots
occur, that memory gets allocated and freed and so forth.

These machines are _also_ NUMA.  Memory can effectively get lumped so
that you can't always get memory on a single NUMA node.  Because of all
of the other actions of other partitions, these conditions are always
changing.  The end result is that the amount of memory which each NUMA
node has *in* *a* *single* *partition* can theoretically change between
reboots.

OK, back to the hapless system admin using kernelcore.  They have a
4-node system with 2GB of RAM in each node for 8GB total.  They use
kernelcore=1GB.  They end up with 4x1GB ZONE_DMA and 4x1GB
ZONE_EASYRCLM.  Perfect.  You can safely remove 4GB of RAM.

Now, imagine that the machine has been heavily used for a while, there
is only 1 node's memory available, but CPUs are available in the same
places as before.  So, you start up your partition again have 8GB of
memory in one node.  Same kernelcore=1GB option.  You get 1x7GB ZONE_DMA
and 1x1GB ZONE_EASYRCLM.  I'd argue this is going to be a bit of a
surprise to the poor admin.

> zones_size[] is what free_area_init() expects to receive so there is not a 
> lot of room to fiddle with it's meaning without causing more trouble.

It is just passed in there as an argument.  If you can think of a way to
make it more understandable, change it in your architecture, and send
the patch for the main one.

> > One other thing, I want to _know_ that variables being compared are in
> > the same units.  When one is called "pages_" something and the other is
> > something "_size", I don't _know_.
> 
> chunk_num_pages ?

No. :)  The words "chunks" and "clumps" have a bit of a stigma, just
like the number "3".  Ask Matt Dobson.

num_pages_WHAT?  node_num_pages?  silly_num_pages?

Chunk is pretty meaningless.

> yep. get_zholes_size() could be split into two functions 
> find_start_easyrclm_pfn() and get_nid_zholes_size(). Would that be pretty 
> clear-cut?

I think so.

-- Dave

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

[Index of Archives]     [Kernel Newbies]     [Netfilter]     [Bugtraq]     [Photo]     [Stuff]     [Gimp]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Video 4 Linux]     [Linux for the blind]     [Linux Resources]
  Powered by Linux