On Thu, 2006-07-06 at 02:56 -0700, Andrew Morton wrote:
> On Thu, 06 Jul 2006 05:28:35 -0400
> Shailabh Nagar <[email protected]> wrote:
>
> > On systems with a large number of cpus, with even a modest rate of
> > tasks exiting per cpu, the volume of taskstats data sent on thread exit
> > can overflow a userspace listener's buffers.
> >
> > One approach to avoiding overflow is to allow listeners to get data for
> > a limited and specific set of cpus. By scaling the number of listeners
> > and/or the cpus they monitor, userspace can handle the statistical data
> > overload more gracefully.
> >
> > In this patch, each listener registers to listen to a specific set of
> > cpus by specifying a cpumask. The interest is recorded per-cpu. When
> > a task exits on a cpu, its taskstats data is unicast to each listener
> > interested in that cpu.
> >
> > Thanks to Andrew Morton for pointing out the various scalability and
> > general concerns of previous attempts and for suggesting this design.
> >
> > ...
> >
> > --- linux-2.6.17-mm3equiv.orig/include/linux/taskstats.h 2006-06-30 19:03:40.000000000 -0400
> > +++ linux-2.6.17-mm3equiv/include/linux/taskstats.h 2006-07-06 02:38:28.000000000 -0400
>
> Your email client performs space-stuffing. Fortunately "sed -e 's/^ / /'"
> is easy.
>
> > #include <linux/taskstats_kern.h>
> > #include <linux/delayacct.h>
> > +#include <linux/cpumask.h>
> > +#include <linux/percpu.h>
> > #include <net/genetlink.h>
> > #include <asm/atomic.h>
>
> Like that.
Sorry. First the text breaking and now patch mangling....I'll switch
mail clients.
>
> >
> > +static int add_del_listener(pid_t pid, cpumask_t *maskp, int isadd)
> > +{
> > + struct listener *s;
> > + struct listener_list *listeners;
> > + unsigned int cpu;
> > + cpumask_t mask;
> > + struct list_head *p;
> > +
> > + memcpy(&mask, maskp, sizeof(cpumask_t));
>
> mask = *maskp; ?
Yes.
>
> > + if (!cpus_subset(mask, cpu_possible_map))
> > + return -EINVAL;
> > +
> > + if (isadd == REGISTER) {
> > + for_each_cpu_mask(cpu, mask) {
> > + s = kmalloc_node(sizeof(struct listener), GFP_KERNEL,
> > + cpu_to_node(cpu));
> > + if (!s)
> > + return -ENOMEM;
> > + s->pid = pid;
> > + INIT_LIST_HEAD(&s->list);
> > +
> > + listeners = &per_cpu(listener_array, cpu);
> > + down_write(&listeners->sem);
> > + list_add(&s->list, &listeners->list);
> > + up_write(&listeners->sem);
> > + }
> > + } else {
> > + for_each_cpu_mask(cpu, mask) {
> > + struct list_head *tmp;
> > +
> > + listeners = &per_cpu(listener_array, cpu);
> > + down_write(&listeners->sem);
> > + list_for_each_safe(p, tmp, &listeners->list) {
> > + s = list_entry(p, struct listener, list);
> > + if (s->pid == pid) {
> > + list_del(&s->list);
> > + kfree(s);
> > + break;
> > + }
> > + }
> > + up_write(&listeners->sem);
> > + }
> > + }
> > + return 0;
> > +}
>
> You might choose to handle the ENOMEM situation here by backing out and not
> leaving things half-installed. I suspect that's just a simple `goto'.
Will do.
>
> > -static int taskstats_send_stats(struct sk_buff *skb, struct genl_info *info)
> > +static int taskstats_user_cmd(struct sk_buff *skb, struct genl_info *info)
> > {
> > int rc = 0;
> > struct sk_buff *rep_skb;
> > @@ -210,6 +302,25 @@ static int taskstats_send_stats(struct s
> > void *reply;
> > size_t size;
> > struct nlattr *na;
> > + cpumask_t mask;
>
> When counting add_del_listener(), that's two cpumasks on the stack. How
> big can these get? 256 bytes? Is it possible to get by with just the one?
For 1024 cpus, 128 bytes. All the cpumask functions seem to need the
mask (though they internally use the address). Will try to see if I can
reduce.
>
> > +
> > + if (info->attrs[TASKSTATS_CMD_ATTR_REGISTER_CPUMASK]) {
> > + na = info->attrs[TASKSTATS_CMD_ATTR_REGISTER_CPUMASK];
> > + if (nla_len(na) > TASKSTATS_CPUMASK_MAXLEN)
> > + return -E2BIG;
> > + cpulist_parse((char *)nla_data(na), mask);
>
> Best check the return value from this function.
Oops.
>
> > + rc = add_del_listener(info->snd_pid, &mask, REGISTER);
> > + return rc;
> > + }
> > +
> > + if (info->attrs[TASKSTATS_CMD_ATTR_DEREGISTER_CPUMASK]) {
> > + na = info->attrs[TASKSTATS_CMD_ATTR_DEREGISTER_CPUMASK];
> > + if (nla_len(na) > TASKSTATS_CPUMASK_MAXLEN)
> > + return -E2BIG;
> > + cpulist_parse((char *)nla_data(na), mask);
> > + rc = add_del_listener(info->snd_pid, &mask, DEREGISTER);
> > + return rc;
> > + }
> >
> > ...
> >
> > +void taskstats_exit_alloc(struct taskstats **ptidstats, unsigned int *mycpu)
> > +{
> > + struct listener_list *listeners;
> > + struct taskstats *tmp;
> > + /*
> > + * This is the cpu on which the task is exiting currently and will
> > + * be the one for which the exit event is sent, even if the cpu
> > + * on which this function is running changes later.
> > + */
> > + *mycpu = raw_smp_processor_id();
> > +
> > + *ptidstats = NULL;
> > + tmp = kmem_cache_zalloc(taskstats_cache, SLAB_KERNEL);
> > + if (!tmp)
> > + return;
> > +
> > + listeners = &per_cpu(listener_array, *mycpu);
> > + down_read(&listeners->sem);
> > + if (!list_empty(&listeners->list)) {
> > + *ptidstats = tmp;
> > + tmp = NULL;
> > + }
> > + up_read(&listeners->sem);
> > + if (tmp)
> > + kfree(tmp);
>
> kfree(NULL) is legal.
Aha. Didn't remember that.
>
>
> Looks good to me. Does it work?
Yes, though ts only been run on a single cpu so far but the
blocking of sends without a registered listener,
registration/deregistration etc. is all working (confirmed by printing
out the registered list of pids etc.). Stability is also fine so far.
--Shailabh
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]