This patch removes all cpuset-specific knowlege from the container
system, replacing it with a generic API that can be used by multiple
subsystems. Cpusets is adapted to be a container subsystem.

Signed-off-by: Paul Menage <[EMAIL PROTECTED]>

---
 Documentation/containers.txt |  224 ++++++++++++++--
 include/linux/container.h    |   70 ++++-
 include/linux/cpuset.h       |   16 -
 include/linux/mempolicy.h    |   12 
 include/linux/sched.h        |    2 
 kernel/container.c           |  589 ++++++++++++++++++++++++++++++++++---------
 kernel/cpuset.c              |  168 ++++++++----
 mm/mempolicy.c               |    2 
 8 files changed, 852 insertions(+), 231 deletions(-)

Index: container-2.6.19-rc5/include/linux/container.h
===================================================================
--- container-2.6.19-rc5.orig/include/linux/container.h
+++ container-2.6.19-rc5/include/linux/container.h
@@ -14,8 +14,6 @@
 
 #ifdef CONFIG_CONTAINERS
 
-extern int number_of_containers;       /* How many containers are defined in 
system? */
-
 extern int container_init_early(void);
 extern int container_init(void);
 extern void container_init_smp(void);
@@ -30,6 +28,13 @@ extern void container_unlock(void);
 extern void container_manage_lock(void);
 extern void container_manage_unlock(void);
 
+struct containerfs_root;
+
+/* Per-subsystem/per-container state maintained by the system. */
+struct container_subsys_state {
+       struct container *container;
+};
+
 struct container {
        unsigned long flags;            /* "unsigned long" so bitops work */
 
@@ -46,11 +51,15 @@ struct container {
        struct list_head children;      /* my children */
 
        struct container *parent;       /* my parent */
-       struct dentry *dentry;          /* container fs entry */
+       struct dentry *dentry;          /* container fs entry */
 
-#ifdef CONFIG_CPUSETS
-       struct cpuset *cpuset;
-#endif
+       /* Private pointers for each registered subsystem */
+       struct container_subsys_state *subsys[CONFIG_MAX_CONTAINER_SUBSYS];
+
+       int hierarchy;
+
+       struct containerfs_root *root;
+       struct container *top_container;
 };
 
 /* struct cftype:
@@ -85,6 +94,55 @@ int container_add_file(struct container 
 int container_is_removed(const struct container *cont);
 void container_set_release_agent_path(const char *path);
 
+/* Container subsystem type. See Documentation/containers.txt for details */
+
+struct container_subsys {
+       int (*create)(struct container_subsys *ss,
+                     struct container *cont);
+       void (*destroy)(struct container_subsys *ss, struct container *cont);
+       int (*can_attach)(struct container_subsys *ss,
+                         struct container *cont, struct task_struct *tsk);
+       void (*attach)(struct container_subsys *ss, struct container *cont,
+                       struct container *old_cont, struct task_struct *tsk);
+       void (*post_attach)(struct container_subsys *ss,
+                           struct container *cont,
+                           struct container *old_cont,
+                           struct task_struct *tsk);
+       int (*populate)(struct container_subsys *ss,
+                       struct container *cont);
+
+       int subsys_id;
+#define MAX_CONTAINER_TYPE_NAMELEN 32
+       const char *name;
+
+       /* Protected by RCU */
+       int hierarchy;
+
+       struct list_head sibling;
+};
+
+int container_register_subsys(struct container_subsys *subsys);
+
+static inline struct container_subsys_state *container_subsys_state(
+       struct container *cont,
+       struct container_subsys *ss)
+{
+       return cont->subsys[ss->subsys_id];
+}
+
+static inline struct container* task_container(struct task_struct *task,
+                                              struct container_subsys *ss)
+{
+       return rcu_dereference(task->container[ss->hierarchy]);
+}
+
+static inline struct container_subsys_state *task_subsys_state(
+       struct task_struct *task,
+       struct container_subsys *ss)
+{
+       return container_subsys_state(task_container(task, ss), ss);
+}
+
 #else /* !CONFIG_CONTAINERS */
 
 static inline int container_init_early(void) { return 0; }
Index: container-2.6.19-rc5/include/linux/cpuset.h
===================================================================
--- container-2.6.19-rc5.orig/include/linux/cpuset.h
+++ container-2.6.19-rc5/include/linux/cpuset.h
@@ -60,16 +60,7 @@ static inline int cpuset_do_slab_mem_spr
 
 extern void cpuset_track_online_nodes(void);
 
-extern int cpuset_can_attach_task(struct container *cont,
-                                 struct task_struct *tsk);
-extern void cpuset_attach_task(struct container *cont,
-                               struct task_struct *tsk);
-extern void cpuset_post_attach_task(struct container *cont,
-                                   struct container *oldcont,
-                                   struct task_struct *tsk);
-extern int cpuset_populate_dir(struct container *cont);
-extern int cpuset_create(struct container *cont);
-extern void cpuset_destroy(struct container *cont);
+extern int current_cpuset_is_being_rebound(void);
 
 #else /* !CONFIG_CPUSETS */
 
@@ -131,6 +122,11 @@ static inline int cpuset_do_slab_mem_spr
 
 static inline void cpuset_track_online_nodes(void) {}
 
+static inline int current_cpuset_is_being_rebound(void)
+{
+       return 0;
+}
+
 #endif /* !CONFIG_CPUSETS */
 
 #endif /* _LINUX_CPUSET_H */
Index: container-2.6.19-rc5/kernel/container.c
===================================================================
--- container-2.6.19-rc5.orig/kernel/container.c
+++ container-2.6.19-rc5/kernel/container.c
@@ -55,7 +55,6 @@
 #include <linux/time.h>
 #include <linux/backing-dev.h>
 #include <linux/sort.h>
-#include <linux/cpuset.h>
 
 #include <asm/uaccess.h>
 #include <asm/atomic.h>
@@ -63,12 +62,27 @@
 
 #define CONTAINER_SUPER_MAGIC          0x27e0eb
 
-/*
- * Tracks how many containers are currently defined in system.
- * When there is only one container (the root container) we can
- * short circuit some hooks.
- */
-int number_of_containers __read_mostly;
+static struct container_subsys *subsys[CONFIG_MAX_CONTAINER_SUBSYS];
+static int subsys_count = 0;
+
+struct containerfs_root {
+       struct super_block *sb;
+       unsigned long subsys_bits;
+       struct list_head subsys_list;
+       struct container top_container;
+       /*
+        * Tracks how many containers are currently defined in system.
+        * When there is only one container (the root container) we can
+        * short circuit some hooks.
+        */
+       int number_of_containers;
+       struct vfsmount *pin_mount;
+};
+
+/* The set of hierarchies in use. Hierachy 0 is reserved for the
+ * subsystems that are otherwise unattached. */
+
+static struct containerfs_root rootnode[CONFIG_MAX_CONTAINER_HIERARCHIES];
 
 /* bits in struct container flags field */
 typedef enum {
@@ -87,11 +101,8 @@ static inline int notify_on_release(cons
        return test_bit(CONT_NOTIFY_ON_RELEASE, &cont->flags);
 }
 
-static struct container top_container = {
-       .count = ATOMIC_INIT(0),
-       .sibling = LIST_HEAD_INIT(top_container.sibling),
-       .children = LIST_HEAD_INIT(top_container.children),
-};
+#define for_each_subsys(_hierarchy, _ss) list_for_each_entry(_ss, 
&rootnode[_hierarchy].subsys_list, sibling)
+
 
 /* The path to use for release notifications. No locking between
  * setting and use - so if userspace updates this while subcontainers
@@ -105,9 +116,6 @@ void container_set_release_agent_path(co
        container_manage_unlock();
 }
 
-static struct vfsmount *container_mount;
-static struct super_block *container_sb;
-
 /*
  * We have two global container mutexes below.  They can nest.
  * It is ok to first take manage_mutex, then nest callback_mutex.  We also
@@ -202,15 +210,18 @@ static DEFINE_MUTEX(callback_mutex);
 
 static int container_mkdir(struct inode *dir, struct dentry *dentry, int mode);
 static int container_rmdir(struct inode *unused_dir, struct dentry *dentry);
+static int container_populate_dir(struct container *cont);
+static struct inode_operations container_dir_inode_operations;
+struct file_operations proc_containerstats_operations;
 
 static struct backing_dev_info container_backing_dev_info = {
        .ra_pages = 0,          /* No readahead */
        .capabilities   = BDI_CAP_NO_ACCT_DIRTY | BDI_CAP_NO_WRITEBACK,
 };
 
-static struct inode *container_new_inode(mode_t mode)
+static struct inode *container_new_inode(mode_t mode, struct super_block *sb)
 {
-       struct inode *inode = new_inode(container_sb);
+       struct inode *inode = new_inode(sb);
 
        if (inode) {
                inode->i_mode = mode;
@@ -282,32 +293,93 @@ static void container_d_remove_dir(struc
        remove_dir(dentry);
 }
 
+/*
+ * Release the last use of a hierarchy.  Will never be called when
+ * there are active subcontainers since each subcontainer bumps the
+ * value of sb->s_active.
+ */
+
+static void container_put_super(struct super_block *sb) {
+
+       struct containerfs_root *root = sb->s_fs_info;
+       int hierarchy = root->top_container.hierarchy;
+       int i;
+       struct container *cont = &root->top_container;
+       struct container *dummytop = &rootnode[0].top_container;
+       struct task_struct *g, *p;
+
+       root->sb = NULL;
+       sb->s_fs_info = NULL;
+
+       mutex_lock(&manage_mutex);
+
+       BUG_ON(root->number_of_containers != 1);
+       BUG_ON(!list_empty(&cont->children));
+       BUG_ON(!list_empty(&cont->sibling));
+       BUG_ON(!root->subsys_bits);
+
+       mutex_lock(&callback_mutex);
+
+       /* Remove all tasks from this container hierarchy */
+       read_lock(&tasklist_lock);
+       do_each_thread(g, p) {
+               task_lock(p);
+               BUG_ON(!p->container[hierarchy]);
+               BUG_ON(p->container[hierarchy] != cont);
+               rcu_assign_pointer(p->container[hierarchy], NULL);
+               task_unlock(p);
+       } while_each_thread(g, p);
+       read_unlock(&tasklist_lock);
+       atomic_set(&cont->count, 1);
+
+       /* Remove all subsystems from this hierarchy */
+       for (i = 0; i < subsys_count; i++) {
+               if (root->subsys_bits & (1 << i)) {
+                       struct container_subsys *ss = subsys[i];
+                       BUG_ON(cont->subsys[i] != dummytop->subsys[i]);
+                       BUG_ON(cont->subsys[i]->container != cont);
+                       dummytop->subsys[i]->container = dummytop;
+                       cont->subsys[i] = NULL;
+                       rcu_assign_pointer(subsys[i]->hierarchy, 0);
+                       list_del(&ss->sibling);
+               } else {
+                       BUG_ON(cont->subsys[i]);
+               }
+       }
+       root->subsys_bits = 0;
+       mutex_unlock(&callback_mutex);
+       synchronize_rcu();
+
+       mutex_unlock(&manage_mutex);
+}
+
 static struct super_operations container_ops = {
        .statfs = simple_statfs,
        .drop_inode = generic_delete_inode,
+       .put_super = container_put_super,
 };
 
-static int container_fill_super(struct super_block *sb, void *unused_data,
-                                                       int unused_silent)
+static int container_fill_super(struct super_block *sb, void *options,
+                               int unused_silent)
 {
        struct inode *inode;
        struct dentry *root;
+       struct containerfs_root *hroot = options;
 
        sb->s_blocksize = PAGE_CACHE_SIZE;
        sb->s_blocksize_bits = PAGE_CACHE_SHIFT;
        sb->s_magic = CONTAINER_SUPER_MAGIC;
        sb->s_op = &container_ops;
-       container_sb = sb;
 
-       inode = container_new_inode(S_IFDIR | S_IRUGO | S_IXUGO | S_IWUSR);
-       if (inode) {
-               inode->i_op = &simple_dir_inode_operations;
-               inode->i_fop = &simple_dir_operations;
-               /* directories start off with i_nlink == 2 (for "." entry) */
-               inode->i_nlink++;
-       } else {
+       inode = container_new_inode(S_IFDIR | S_IRUGO | S_IXUGO | S_IWUSR, sb);
+       if (!inode)
                return -ENOMEM;
-       }
+
+       inode->i_op = &simple_dir_inode_operations;
+       inode->i_fop = &simple_dir_operations;
+       inode->i_op = &container_dir_inode_operations;
+       /* directories start off with i_nlink == 2 (for "." entry) */
+       inc_nlink(inode);
 
        root = d_alloc_root(inode);
        if (!root) {
@@ -315,6 +387,12 @@ static int container_fill_super(struct s
                return -ENOMEM;
        }
        sb->s_root = root;
+       root->d_fsdata = &hroot->top_container;
+       hroot->top_container.dentry = root;
+
+       sb->s_fs_info = hroot;
+       hroot->sb = sb;
+
        return 0;
 }
 
@@ -322,7 +400,131 @@ static int container_get_sb(struct file_
                         int flags, const char *unused_dev_name,
                         void *data, struct vfsmount *mnt)
 {
-       return get_sb_single(fs_type, flags, data, container_fill_super, mnt);
+       int i;
+       struct container_subsys *ss;
+       char *token, *o = data ?: "all";
+       unsigned long subsys_bits = 0;
+       int ret = 0;
+       struct containerfs_root *root = NULL;
+       int hierarchy;
+
+       mutex_lock(&manage_mutex);
+
+       /* First find the desired set of resource controllers */
+       while ((token = strsep(&o, ",")) != NULL) {
+               if (!*token) {
+                       ret = -EINVAL;
+                       goto out_unlock;
+               }
+               if (!strcmp(token, "all")) {
+                       subsys_bits = (1 << subsys_count) - 1;
+               } else {
+                       for (i = 0; i < subsys_count; i++) {
+                               ss = subsys[i];
+                               if (!strcmp(token, ss->name)) {
+                                       subsys_bits |= 1 << i;
+                                       break;
+                               }
+                       }
+                       if (i == subsys_count) {
+                               ret = -ENOENT;
+                               goto out_unlock;
+                       }
+               }
+       }
+
+       /* See if we already have a hierarchy containing this set */
+
+       for (i = 1; i < CONFIG_MAX_CONTAINER_HIERARCHIES; i++) {
+               root = &rootnode[i];
+               /* We match - use this hieracrchy */
+               if (root->subsys_bits == subsys_bits) break;
+               /* We clash - fail */
+               if (root->subsys_bits & subsys_bits) {
+                       ret = -EBUSY;
+                       goto out_unlock;
+               }
+       }
+
+       if (i == CONFIG_MAX_CONTAINER_HIERARCHIES) {
+               /* No existing hierarchy matched this set - but we
+                * know that all the subsystems are free */
+               for (i = 1; i < CONFIG_MAX_CONTAINER_HIERARCHIES; i++) {
+                       root = &rootnode[i];
+                       if (!root->sb && !root->subsys_bits) break;
+               }
+       }
+
+       if (i == CONFIG_MAX_CONTAINER_HIERARCHIES) {
+               ret = -ENOSPC;
+               goto out_unlock;
+       }
+
+       hierarchy = i;
+
+       if (!root->sb) {
+               /* We need a new superblock for this container combination */
+               struct container *cont = &root->top_container;
+               struct container_subsys *ss;
+               struct task_struct *p, *g;
+               struct container *dummytop = &rootnode[0].top_container;
+
+               BUG_ON(root->subsys_bits);
+               root->subsys_bits = subsys_bits;
+               ret = get_sb_nodev(fs_type, flags, root,
+                                  container_fill_super, mnt);
+               if (ret)
+                       goto out_unlock;
+
+               BUG_ON(!list_empty(&cont->sibling));
+               BUG_ON(!list_empty(&cont->children));
+               BUG_ON(root->number_of_containers != 1);
+
+               mutex_lock(&callback_mutex);
+
+               /* Add all tasks into this container hierarchy */
+               atomic_set(&cont->count, 1);
+               read_lock(&tasklist_lock);
+               do_each_thread(g, p) {
+                       task_lock(p);
+                       BUG_ON(p->container[hierarchy]);
+                       rcu_assign_pointer(p->container[hierarchy], cont);
+                       if (!(p->flags & PF_EXITING)) {
+                               atomic_inc(&cont->count);
+                       }
+                       task_unlock(p);
+               } while_each_thread(g, p);
+               read_unlock(&tasklist_lock);
+
+               /* Move all the relevant subsystems into the hierarchy. */
+               for (i = 0; i < subsys_count; i++) {
+                       if (!(subsys_bits & (1 << i))) continue;
+
+                       ss = subsys[i];
+
+                       BUG_ON(cont->subsys[i]);
+                       BUG_ON(dummytop->subsys[i]->container != dummytop);
+                       cont->subsys[i] = dummytop->subsys[i];
+                       cont->subsys[i]->container = cont;
+                       list_add(&ss->sibling, &root->subsys_list);
+                       rcu_assign_pointer(subsys[i]->hierarchy,
+                                          hierarchy);
+               }
+               mutex_unlock(&callback_mutex);
+               synchronize_rcu();
+
+               container_populate_dir(cont);
+
+       } else {
+               /* Reuse the existing superblock */
+               ret = simple_set_mnt(mnt, root->sb);
+               if (!ret)
+                       atomic_inc(&root->sb->s_active);
+       }
+
+ out_unlock:
+       mutex_unlock(&manage_mutex);
+       return ret;
 }
 
 static struct file_system_type container_fs_type = {
@@ -501,6 +703,8 @@ static int attach_task(struct container 
        struct task_struct *tsk;
        struct container *oldcont;
        int retval = 0;
+       struct container_subsys *ss;
+       int h = cont->hierarchy;
 
        if (sscanf(pidbuf, "%d", &pid) != 1)
                return -EIO;
@@ -527,18 +731,20 @@ static int attach_task(struct container 
                get_task_struct(tsk);
        }
 
-#ifdef CONFIG_CPUSETS
-       retval = cpuset_can_attach_task(cont, tsk);
-#endif
-       if (retval) {
-               put_task_struct(tsk);
-               return retval;
+       for_each_subsys(h, ss) {
+               if (ss->can_attach) {
+                       retval = ss->can_attach(ss, cont, tsk);
+                       if (retval) {
+                               put_task_struct(tsk);
+                               return retval;
+                       }
+               }
        }
 
        mutex_lock(&callback_mutex);
 
        task_lock(tsk);
-       oldcont = tsk->container;
+       oldcont = tsk->container[h];
        if (!oldcont) {
                task_unlock(tsk);
                mutex_unlock(&callback_mutex);
@@ -546,18 +752,22 @@ static int attach_task(struct container 
                return -ESRCH;
        }
        atomic_inc(&cont->count);
-       rcu_assign_pointer(tsk->container, cont);
+       rcu_assign_pointer(tsk->container[h], cont);
        task_unlock(tsk);
 
-#ifdef CONFIG_CPUSETS
-       cpuset_attach_task(cont, tsk);
-#endif
+       for_each_subsys(h, ss) {
+               if (ss->attach) {
+                       ss->attach(ss, cont, oldcont, tsk);
+               }
+       }
 
        mutex_unlock(&callback_mutex);
 
-#ifdef CONFIG_CPUSETS
-       cpuset_post_attach_task(cont, oldcont, tsk);
-#endif
+       for_each_subsys(h, ss) {
+               if (ss->post_attach) {
+                       ss->post_attach(ss, cont, oldcont, tsk);
+               }
+       }
 
        put_task_struct(tsk);
        synchronize_rcu();
@@ -780,7 +990,7 @@ static struct inode_operations container
        .rename = container_rename,
 };
 
-static int container_create_file(struct dentry *dentry, int mode)
+static int container_create_file(struct dentry *dentry, int mode, struct 
super_block *sb)
 {
        struct inode *inode;
 
@@ -789,7 +999,7 @@ static int container_create_file(struct 
        if (dentry->d_inode)
                return -EEXIST;
 
-       inode = container_new_inode(mode);
+       inode = container_new_inode(mode, sb);
        if (!inode)
                return -ENOMEM;
 
@@ -798,7 +1008,7 @@ static int container_create_file(struct 
                inode->i_fop = &simple_dir_operations;
 
                /* start off with i_nlink == 2 (for "." entry) */
-               inode->i_nlink++;
+               inc_nlink(inode);
        } else if (S_ISREG(mode)) {
                inode->i_size = 0;
                inode->i_fop = &container_file_operations;
@@ -828,10 +1038,10 @@ static int container_create_dir(struct c
        dentry = container_get_dentry(parent, name);
        if (IS_ERR(dentry))
                return PTR_ERR(dentry);
-       error = container_create_file(dentry, S_IFDIR | mode);
+       error = container_create_file(dentry, S_IFDIR | mode, cont->root->sb);
        if (!error) {
                dentry->d_fsdata = cont;
-               parent->d_inode->i_nlink++;
+               inc_nlink(parent->d_inode);
                cont->dentry = dentry;
        }
        dput(dentry);
@@ -848,7 +1058,7 @@ int container_add_file(struct container 
        mutex_lock(&dir->d_inode->i_mutex);
        dentry = container_get_dentry(dir, cft->name);
        if (!IS_ERR(dentry)) {
-               error = container_create_file(dentry, 0644 | S_IFREG);
+               error = container_create_file(dentry, 0644 | S_IFREG, 
cont->root->sb);
                if (!error)
                        dentry->d_fsdata = (void *)cft;
                dput(dentry);
@@ -894,7 +1104,7 @@ static int pid_array_load(pid_t *pidarra
        read_lock(&tasklist_lock);
 
        do_each_thread(g, p) {
-               if (p->container == cont) {
+               if (p->container[cont->hierarchy] == cont) {
                        pidarray[n++] = p->pid;
                        if (unlikely(n == npids))
                                goto array_full;
@@ -1037,18 +1247,21 @@ static struct cftype cft_release_agent =
 static int container_populate_dir(struct container *cont)
 {
        int err;
+       struct container_subsys *ss;
 
        if ((err = container_add_file(cont, &cft_notify_on_release)) < 0)
                return err;
        if ((err = container_add_file(cont, &cft_tasks)) < 0)
                return err;
-       if ((cont == &top_container) &&
+       if ((cont == cont->top_container) &&
            (err = container_add_file(cont, &cft_release_agent)) < 0)
                return err;
-#ifdef CONFIG_CPUSETS
-       if ((err = cpuset_populate_dir(cont)) < 0)
-               return err;
-#endif
+
+       for_each_subsys(cont->hierarchy, ss) {
+               if (ss->populate && (err = ss->populate(ss, cont)) < 0)
+                       return err;
+       }
+
        return 0;
 }
 
@@ -1064,13 +1277,24 @@ static int container_populate_dir(struct
 static long container_create(struct container *parent, const char *name, int 
mode)
 {
        struct container *cont;
-       int err;
+       struct containerfs_root *root = parent->root;
+       int err = 0;
+       struct container_subsys *ss;
+       struct super_block *sb = root->sb;
 
        cont = kmalloc(sizeof(*cont), GFP_KERNEL);
        if (!cont)
                return -ENOMEM;
 
+       /* Grab a reference on the superblock so the hierarchy doesn't
+        * get deleted on unmount if there are child containers.  This
+        * can be done outside manage_mutex, since the sb can't
+        * disappear while someone has an open control file on the
+        * fs */
+       atomic_inc(&sb->s_active);
+
        mutex_lock(&manage_mutex);
+
        cont->flags = 0;
        if (notify_on_release(parent))
                set_bit(CONT_NOTIFY_ON_RELEASE, &cont->flags);
@@ -1079,16 +1303,18 @@ static long container_create(struct cont
        INIT_LIST_HEAD(&cont->children);
 
        cont->parent = parent;
+       cont->root = parent->root;
+       cont->hierarchy = parent->hierarchy;
 
-#ifdef CONFIG_CPUSETS
-       err = cpuset_create(cont);
-       if (err)
-               goto err_unlock_free;
-#endif
+       for_each_subsys(cont->hierarchy, ss) {
+               err = ss->create(ss, cont);
+               if (err) goto err_destroy;
+               cont->subsys[ss->subsys_id]->container = cont;
+       }
 
        mutex_lock(&callback_mutex);
        list_add(&cont->sibling, &cont->parent->children);
-       number_of_containers++;
+       root->number_of_containers++;
        mutex_unlock(&callback_mutex);
 
        err = container_create_dir(cont, name, mode);
@@ -1107,15 +1333,22 @@ static long container_create(struct cont
        return 0;
 
  err_remove:
-#ifdef CONFIG_CPUSETS
-       cpuset_destroy(cont);
-#endif
+
        mutex_lock(&callback_mutex);
        list_del(&cont->sibling);
-       number_of_containers--;
+       root->number_of_containers--;
        mutex_unlock(&callback_mutex);
- err_unlock_free:
+
+ err_destroy:
+
+       for_each_subsys(cont->hierarchy, ss) {
+               ss->destroy(ss, cont);
+       }
+
        mutex_unlock(&manage_mutex);
+
+       deactivate_super(sb);
+
        kfree(cont);
        return err;
 }
@@ -1145,6 +1378,9 @@ static int container_rmdir(struct inode 
        struct dentry *d;
        struct container *parent;
        char *pathbuf = NULL;
+       struct container_subsys *ss;
+       struct super_block *sb;
+       struct containerfs_root *root;
 
        /* the vfs holds both inode->i_mutex already */
 
@@ -1158,6 +1394,8 @@ static int container_rmdir(struct inode 
                return -EBUSY;
        }
        parent = cont->parent;
+       root = cont->root;
+       sb = root->sb;
        mutex_lock(&callback_mutex);
        set_bit(CONT_REMOVED, &cont->flags);
        list_del(&cont->sibling);       /* delete my sibling from 
parent->children */
@@ -1165,67 +1403,115 @@ static int container_rmdir(struct inode 
        d = dget(cont->dentry);
        cont->dentry = NULL;
        spin_unlock(&d->d_lock);
+
+       for_each_subsys(root->top_container.hierarchy, ss) {
+               ss->destroy(ss, cont);
+       }
        container_d_remove_dir(d);
        dput(d);
-       number_of_containers--;
+       root->number_of_containers--;
        mutex_unlock(&callback_mutex);
-#ifdef CONFIG_CPUSETS
-       cpuset_destroy(cont);
-#endif
+
        if (list_empty(&parent->children))
                check_for_release(parent, &pathbuf);
+
        mutex_unlock(&manage_mutex);
+       /* Drop the active superblock reference that we took when we
+        * created the container */
+       deactivate_super(sb);
        container_release_agent(pathbuf);
        return 0;
 }
 
-/*
- * container_init_early - probably not needed yet, but will be needed
- * once cpusets are hooked into this code
- */
+
+/**
+ * container_init_early - initialize containers at system boot
+ *
+ * Description: Initialize the container housekeeping structures
+ **/
 
 int __init container_init_early(void)
 {
-       struct task_struct *tsk = current;
+       int i;
+
+       for (i = 0; i < CONFIG_MAX_CONTAINER_HIERARCHIES; i++) {
+               struct containerfs_root *root = &rootnode[i];
+               struct container *cont = &root->top_container;
+               INIT_LIST_HEAD(&root->subsys_list);
+               root->number_of_containers = 1;
+
+               cont->root = root;
+               cont->hierarchy = i;
+               INIT_LIST_HEAD(&cont->sibling);
+               INIT_LIST_HEAD(&cont->children);
+               cont->top_container = cont;
+               atomic_set(&cont->count, 1);
+       }
+       init_task.container[0] = &rootnode[0].top_container;
+       atomic_inc(&init_task.container[0]->count);
 
-       tsk->container = &top_container;
        return 0;
 }
 
 /**
- * container_init - initialize containers at system boot
- *
- * Description: Initialize top_container and the container internal file 
system,
+ * container_init - register container filesystem and /proc file
  **/
 
 int __init container_init(void)
 {
-       struct dentry *root;
        int err;
-
-       init_task.container = &top_container;
+       struct proc_dir_entry *entry;
 
        err = register_filesystem(&container_fs_type);
        if (err < 0)
                goto out;
-       container_mount = kern_mount(&container_fs_type);
-       if (IS_ERR(container_mount)) {
-               printk(KERN_ERR "container: could not mount!\n");
-               err = PTR_ERR(container_mount);
-               container_mount = NULL;
-               goto out;
-       }
-       root = container_mount->mnt_sb->s_root;
-       root->d_fsdata = &top_container;
-       root->d_inode->i_nlink++;
-       top_container.dentry = root;
-       root->d_inode->i_op = &container_dir_inode_operations;
-       number_of_containers = 1;
-       err = container_populate_dir(&top_container);
+
+       entry = create_proc_entry("containers", 0, NULL);
+       if (entry)
+               entry->proc_fops = &proc_containerstats_operations;
+
 out:
        return err;
 }
 
+int container_register_subsys(struct container_subsys *new_subsys) {
+       int retval = 0;
+       int i;
+       struct container *dummytop = &rootnode[0].top_container;
+
+       BUG_ON(new_subsys->hierarchy);
+       mutex_lock(&manage_mutex);
+       if (subsys_count == CONFIG_MAX_CONTAINER_SUBSYS) {
+               retval = -ENOSPC;
+               goto out;
+       }
+       if (!new_subsys->name ||
+           (strlen(new_subsys->name) > MAX_CONTAINER_TYPE_NAMELEN) ||
+           !new_subsys->create || !new_subsys->destroy) {
+               retval = -EINVAL;
+               goto out;
+       }
+       for (i = 0; i < subsys_count; i++) {
+               if (!strcmp(subsys[i]->name, new_subsys->name)) {
+                       retval = -EEXIST;
+                       goto out;
+               }
+       }
+
+       new_subsys->subsys_id = subsys_count;
+       retval = new_subsys->create(new_subsys, dummytop);
+       if (retval) {
+               new_subsys->subsys_id = -1;
+               goto out;
+       }
+       dummytop->subsys[subsys_count]->container = dummytop;
+       subsys[subsys_count++] = new_subsys;
+
+ out:
+       mutex_unlock(&manage_mutex);
+       return retval;
+}
+
 /**
  * container_fork - attach newly forked task to its parents container.
  * @tsk: pointer to task_struct of forking parent process.
@@ -1246,9 +1532,14 @@ out:
 
 void container_fork(struct task_struct *child)
 {
+       int i;
        task_lock(current);
-       child->container = current->container;
-       atomic_inc(&child->container->count);
+       for (i = 0; i < CONFIG_MAX_CONTAINER_HIERARCHIES; i++) {
+               struct container *cont = current->container[i];
+               if (!cont) continue;
+               child->container[i] = cont;
+               atomic_inc(&cont->count);
+       }
        task_unlock(current);
 }
 
@@ -1314,20 +1605,22 @@ void container_fork(struct task_struct *
 void container_exit(struct task_struct *tsk)
 {
        struct container *cont;
-
-       cont = tsk->container;
-       tsk->container = &top_container;        /* the_top_container_hack - see 
above */
-
-       if (notify_on_release(cont)) {
-               char *pathbuf = NULL;
-
-               mutex_lock(&manage_mutex);
-               if (atomic_dec_and_test(&cont->count))
-                       check_for_release(cont, &pathbuf);
-               mutex_unlock(&manage_mutex);
-               container_release_agent(pathbuf);
-       } else {
-               atomic_dec(&cont->count);
+       int i;
+       for (i = 0; i < CONFIG_MAX_CONTAINER_HIERARCHIES; i++) {
+               cont = tsk->container[i];
+               if (!cont) continue;
+               /* the_top_container_hack - see above */
+               tsk->container[i] = cont->top_container;
+               if (notify_on_release(cont)) {
+                       char *pathbuf = NULL;
+                       mutex_lock(&manage_mutex);
+                       if (atomic_dec_and_test(&cont->count))
+                               check_for_release(cont, &pathbuf);
+                       mutex_unlock(&manage_mutex);
+                       container_release_agent(pathbuf);
+               } else {
+                       atomic_dec(&cont->count);
+               }
        }
 }
 
@@ -1387,12 +1680,15 @@ void container_manage_unlock(void)
  *    the_top_container_hack in container_exit(), which sets an exiting tasks
  *    container to top_container.
  */
+
+/* TODO: Use a proper seq_file iterator */
 static int proc_container_show(struct seq_file *m, void *v)
 {
        struct pid *pid;
        struct task_struct *tsk;
        char *buf;
        int retval;
+       int i;
 
        retval = -ENOMEM;
        buf = kmalloc(PAGE_SIZE, GFP_KERNEL);
@@ -1405,14 +1701,25 @@ static int proc_container_show(struct se
        if (!tsk)
                goto out_free;
 
-       retval = -EINVAL;
+       retval = 0;
+
        mutex_lock(&manage_mutex);
 
-       retval = container_path(tsk->container, buf, PAGE_SIZE);
-       if (retval < 0)
-               goto out_unlock;
-       seq_puts(m, buf);
-       seq_putc(m, '\n');
+       for (i = 1; i < CONFIG_MAX_CONTAINER_HIERARCHIES; i++) {
+               struct containerfs_root *root = &rootnode[i];
+               struct container_subsys *ss;
+               int count = 0;
+               if (!root->subsys_bits) continue;
+               for_each_subsys(i, ss) {
+                       seq_printf(m, "%s%s", count++ ? "," : "", ss->name);
+               }
+               seq_putc(m, ':');
+               retval = container_path(tsk->container[i], buf, PAGE_SIZE);
+               if (retval < 0)
+                       goto out_unlock;
+               seq_puts(m, buf);
+               seq_putc(m, '\n');
+       }
 out_unlock:
        mutex_unlock(&manage_mutex);
        put_task_struct(tsk);
@@ -1434,3 +1741,47 @@ struct file_operations proc_container_op
        .llseek         = seq_lseek,
        .release        = single_release,
 };
+
+static int proc_containerstats_show(struct seq_file *m, void *v)
+{
+       int i;
+       mutex_lock(&manage_mutex);
+       seq_puts(m, "Hierarchies:\n");
+       for (i = 0; i < CONFIG_MAX_CONTAINER_HIERARCHIES; i++) {
+               struct containerfs_root *root = &rootnode[i];
+               struct container_subsys *ss;
+               int first = 1;
+               seq_printf(m, "%d: topcount=%d bits=%lx containers=%d (",
+                          i, atomic_read(&root->top_container.count),
+                          root->subsys_bits, root->number_of_containers);
+               for_each_subsys(i, ss) {
+                       seq_printf(m, "%s%s", first ? "" : ", ", ss->name);
+                       first = false;
+               }
+               seq_putc(m, ')');
+               if (root->sb) {
+                       seq_printf(m, " s_active=%d", 
atomic_read(&root->sb->s_active));
+               }
+               seq_putc(m, '\n');
+       }
+       seq_puts(m, "Subsystems:\n");
+       for (i = 0; i < subsys_count; i++) {
+               struct container_subsys *ss = subsys[i];
+               seq_printf(m, "%d: name=%s hierarchy=%d\n",
+                          i, ss->name, ss->hierarchy);
+       }
+       mutex_unlock(&manage_mutex);
+       return 0;
+}
+
+static int containerstats_open(struct inode *inode, struct file *file)
+{
+       return single_open(file, proc_containerstats_show, 0);
+}
+
+struct file_operations proc_containerstats_operations = {
+       .open = containerstats_open,
+       .read = seq_read,
+       .llseek = seq_lseek,
+       .release = single_release,
+};
Index: container-2.6.19-rc5/kernel/cpuset.c
===================================================================
--- container-2.6.19-rc5.orig/kernel/cpuset.c
+++ container-2.6.19-rc5/kernel/cpuset.c
@@ -5,6 +5,7 @@
  *
  *  Copyright (C) 2003 BULL SA.
  *  Copyright (C) 2004-2006 Silicon Graphics, Inc.
+ *  Copyright (C) 2006 Google, Inc
  *
  *  Portions derived from Patrick Mochel's sysfs code.
  *  sysfs is Copyright (c) 2001-3 Patrick Mochel
@@ -12,6 +13,7 @@
  *  2003-10-10 Written by Simon Derr.
  *  2003-10-22 Updates by Stephen Hemminger.
  *  2004 May-July Rework by Paul Jackson.
+ *  2006 Rework by Paul Menage to use generic containers
  *
  *  This file is subject to the terms and conditions of the GNU General Public
  *  License.  See the file COPYING in the main directory of the Linux
@@ -61,6 +63,10 @@
  */
 int number_of_cpusets __read_mostly;
 
+/* Retrieve the cpuset from a container */
+static struct container_subsys cpuset_subsys;
+struct cpuset;
+
 /* See "Frequency meter" comments, below. */
 
 struct fmeter {
@@ -71,11 +77,12 @@ struct fmeter {
 };
 
 struct cpuset {
+       struct container_subsys_state css;
+
        unsigned long flags;            /* "unsigned long" so bitops work */
        cpumask_t cpus_allowed;         /* CPUs allowed to tasks in cpuset */
        nodemask_t mems_allowed;        /* Memory Nodes allowed to tasks */
 
-       struct container *container;    /* Task container */
        struct cpuset *parent;          /* my parent */
 
        /*
@@ -87,6 +94,26 @@ struct cpuset {
        struct fmeter fmeter;           /* memory_pressure filter */
 };
 
+/* Update the cpuset for a container */
+static inline void set_container_cs(struct container *cont, struct cpuset *cs)
+{
+       cont->subsys[cpuset_subsys.subsys_id] = &cs->css;
+}
+
+/* Retrieve the cpuset for a container */
+static inline struct cpuset *container_cs(struct container *cont)
+{
+       return container_of(container_subsys_state(cont, &cpuset_subsys),
+                           struct cpuset, css);
+}
+
+/* Retrieve the cpuset for a task */
+static inline struct cpuset *task_cs(struct task_struct *task)
+{
+       return container_cs(task_container(task, &cpuset_subsys));
+}
+
+
 /* bits in struct cpuset flags field */
 typedef enum {
        CS_CPU_EXCLUSIVE,
@@ -162,7 +189,7 @@ static int cpuset_get_sb(struct file_sys
        if (container_fs) {
                ret = container_fs->get_sb(container_fs, flags,
                                           unused_dev_name,
-                                          data, mnt);
+                                          "cpuset", mnt);
                put_filesystem(container_fs);
        }
        return ret;
@@ -270,20 +297,19 @@ void cpuset_update_task_memory_state(voi
        struct task_struct *tsk = current;
        struct cpuset *cs;
 
-       if (tsk->container->cpuset == &top_cpuset) {
+       if (task_cs(tsk) == &top_cpuset) {
                /* Don't need rcu for top_cpuset.  It's never freed. */
                my_cpusets_mem_gen = top_cpuset.mems_generation;
        } else {
                rcu_read_lock();
-               cs = rcu_dereference(tsk->container->cpuset);
-               my_cpusets_mem_gen = cs->mems_generation;
+               my_cpusets_mem_gen = task_cs(current)->mems_generation;
                rcu_read_unlock();
        }
 
        if (my_cpusets_mem_gen != tsk->cpuset_mems_generation) {
                container_lock();
                task_lock(tsk);
-               cs = tsk->container->cpuset; /* Maybe changed when task not 
locked */
+               cs = task_cs(tsk); /* Maybe changed when task not locked */
                guarantee_online_mems(cs, &tsk->mems_allowed);
                tsk->cpuset_mems_generation = cs->mems_generation;
                if (is_spread_page(cs))
@@ -342,9 +368,8 @@ static int validate_change(const struct 
        struct cpuset *c, *par;
 
        /* Each of our child cpusets must be a subset of us */
-       list_for_each_entry(cont, &cur->container->children, sibling) {
-               c = cont->cpuset;
-               if (!is_cpuset_subset(c, trial))
+       list_for_each_entry(cont, &cur->css.container->children, sibling) {
+               if (!is_cpuset_subset(container_cs(cont), trial))
                        return -EBUSY;
        }
 
@@ -357,8 +382,8 @@ static int validate_change(const struct 
                return -EACCES;
 
        /* If either I or some sibling (!= me) is exclusive, we can't overlap */
-       list_for_each_entry(cont, &par->container->children, sibling) {
-               c = cont->cpuset;
+       list_for_each_entry(cont, &par->css.container->children, sibling) {
+               c = container_cs(cont);
                if ((is_cpu_exclusive(trial) || is_cpu_exclusive(c)) &&
                    c != cur &&
                    cpus_intersects(trial->cpus_allowed, c->cpus_allowed))
@@ -400,8 +425,8 @@ static void update_cpu_domains(struct cp
         * children
         */
        pspan = par->cpus_allowed;
-       list_for_each_entry(cont, &par->container->children, sibling) {
-               c = cont->cpuset;
+       list_for_each_entry(cont, &par->css.container->children, sibling) {
+               c = container_cs(cont);
                if (is_cpu_exclusive(c))
                        cpus_andnot(pspan, pspan, c->cpus_allowed);
        }
@@ -418,8 +443,8 @@ static void update_cpu_domains(struct cp
                 * Get all cpus from current cpuset's cpus_allowed not part
                 * of exclusive children
                 */
-               list_for_each_entry(cont, &cur->container->children, sibling) {
-                       c = cont->cpuset;
+               list_for_each_entry(cont, &cur->css.container->children, 
sibling) {
+                       c = container_cs(cont);
                        if (is_cpu_exclusive(c))
                                cpus_andnot(cspan, cspan, c->cpus_allowed);
                }
@@ -507,7 +532,7 @@ static void cpuset_migrate_mm(struct mm_
        do_migrate_pages(mm, from, to, MPOL_MF_MOVE_ALL);
 
        container_lock();
-       guarantee_online_mems(tsk->container->cpuset, &tsk->mems_allowed);
+       guarantee_online_mems(task_cs(tsk),&tsk->mems_allowed);
        container_unlock();
 }
 
@@ -525,6 +550,8 @@ static void cpuset_migrate_mm(struct mm_
  * their mempolicies to the cpusets new mems_allowed.
  */
 
+static void *cpuset_being_rebound;
+
 static int update_nodemask(struct cpuset *cs, char *buf)
 {
        struct cpuset trialcs;
@@ -542,7 +569,7 @@ static int update_nodemask(struct cpuset
                return -EACCES;
 
        trialcs = *cs;
-       cont = cs->container;
+       cont = cs->css.container;
        retval = nodelist_parse(buf, trialcs.mems_allowed);
        if (retval < 0)
                goto done;
@@ -565,7 +592,7 @@ static int update_nodemask(struct cpuset
        cs->mems_generation = cpuset_mems_generation++;
        container_unlock();
 
-       set_cpuset_being_rebound(cs);           /* causes mpol_copy() rebind */
+       cpuset_being_rebound = cs;              /* causes mpol_copy() rebind */
 
        fudge = 10;                             /* spare mmarray[] slots */
        fudge += cpus_weight(cs->cpus_allowed); /* imagine one fork-bomb/cpu */
@@ -579,13 +606,13 @@ static int update_nodemask(struct cpuset
         * enough mmarray[] w/o using GFP_ATOMIC.
         */
        while (1) {
-               ntasks = atomic_read(&cs->container->count);  /* guess */
+               ntasks = atomic_read(&cs->css.container->count);  /* guess */
                ntasks += fudge;
                mmarray = kmalloc(ntasks * sizeof(*mmarray), GFP_KERNEL);
                if (!mmarray)
                        goto done;
                write_lock_irq(&tasklist_lock);         /* block fork */
-               if (atomic_read(&cs->container->count) <= ntasks)
+               if (atomic_read(&cs->css.container->count) <= ntasks)
                        break;                          /* got enough */
                write_unlock_irq(&tasklist_lock);       /* try again */
                kfree(mmarray);
@@ -602,7 +629,7 @@ static int update_nodemask(struct cpuset
                                "Cpuset mempolicy rebind incomplete.\n");
                        continue;
                }
-               if (p->container != cont)
+               if (task_cs(p) != cs)
                        continue;
                mm = get_task_mm(p);
                if (!mm)
@@ -636,12 +663,17 @@ static int update_nodemask(struct cpuset
 
        /* We're done rebinding vma's to this cpusets new mems_allowed. */
        kfree(mmarray);
-       set_cpuset_being_rebound(NULL);
+       cpuset_being_rebound = NULL;
        retval = 0;
 done:
        return retval;
 }
 
+int current_cpuset_is_being_rebound(void)
+{
+       return task_cs(current) == cpuset_being_rebound;
+}
+
 /*
  * Call with manage_mutex held.
  */
@@ -795,9 +827,10 @@ static int fmeter_getrate(struct fmeter 
        return val;
 }
 
-int cpuset_can_attach_task(struct container *cont, struct task_struct *tsk)
+int cpuset_can_attach(struct container_subsys *ss,
+                     struct container *cont, struct task_struct *tsk)
 {
-       struct cpuset *cs = cont->cpuset;
+       struct cpuset *cs = container_cs(cont);
 
        if (cpus_empty(cs->cpus_allowed) || nodes_empty(cs->mems_allowed))
                return -ENOSPC;
@@ -805,22 +838,23 @@ int cpuset_can_attach_task(struct contai
        return security_task_setscheduler(tsk, 0, NULL);
 }
 
-void cpuset_attach_task(struct container *cont, struct task_struct *tsk)
+void cpuset_attach(struct container_subsys *ss, struct container *cont,
+                  struct container *old_cont, struct task_struct *tsk)
 {
        cpumask_t cpus;
-       struct cpuset *cs = cont->cpuset;
-       guarantee_online_cpus(cs, &cpus);
+       guarantee_online_cpus(container_cs(cont), &cpus);
        set_cpus_allowed(tsk, cpus);
 }
 
-void cpuset_post_attach_task(struct container *cont,
-                            struct container *oldcont,
-                            struct task_struct *tsk)
+void cpuset_post_attach(struct container_subsys *ss,
+                       struct container *cont,
+                       struct container *oldcont,
+                       struct task_struct *tsk)
 {
        nodemask_t from, to;
        struct mm_struct *mm;
-       struct cpuset *cs = cont->cpuset;
-       struct cpuset *oldcs = oldcont->cpuset;
+       struct cpuset *cs = container_cs(cont);
+       struct cpuset *oldcs = container_cs(oldcont);
 
        from = oldcs->mems_allowed;
        to = cs->mems_allowed;
@@ -854,7 +888,7 @@ static ssize_t cpuset_common_file_write(
                                        const char __user *userbuf,
                                        size_t nbytes, loff_t *unused_ppos)
 {
-       struct cpuset *cs = cont->cpuset;
+       struct cpuset *cs = container_cs(cont);
        cpuset_filetype_t type = cft->private;
        char *buffer;
        int retval = 0;
@@ -964,7 +998,7 @@ static ssize_t cpuset_common_file_read(s
                                       char __user *buf,
                                       size_t nbytes, loff_t *ppos)
 {
-       struct cpuset *cs = cont->cpuset;
+       struct cpuset *cs = container_cs(cont);
        cpuset_filetype_t type = cft->private;
        char *page;
        ssize_t retval = 0;
@@ -1083,7 +1117,7 @@ static struct cftype cft_spread_slab = {
        .private = FILE_SPREAD_SLAB,
 };
 
-int cpuset_populate_dir(struct container *cont)
+int cpuset_populate(struct container_subsys *ss, struct container *cont)
 {
        int err;
 
@@ -1118,11 +1152,19 @@ int cpuset_populate_dir(struct container
  *     Must be called with the mutex on the parent inode held
  */
 
-int cpuset_create(struct container *cont)
+int cpuset_create(struct container_subsys *ss, struct container *cont)
 {
        struct cpuset *cs;
-       struct cpuset *parent = cont->parent->cpuset;
+       struct cpuset *parent;
 
+       if (!cont->parent) {
+               /* This is early initialization for the top container */
+               set_container_cs(cont, &top_cpuset);
+               top_cpuset.css.container = cont;
+               top_cpuset.mems_generation = cpuset_mems_generation++;
+               return 0;
+       }
+       parent = container_cs(cont->parent);
        cs = kmalloc(sizeof(*cs), GFP_KERNEL);
        if (!cs)
                return -ENOMEM;
@@ -1139,12 +1181,19 @@ int cpuset_create(struct container *cont
        fmeter_init(&cs->fmeter);
 
        cs->parent = parent;
-       cont->cpuset = cs;
-       cs->container = cont;
+       set_container_cs(cont, cs);
+       cs->css.container = cont;
        number_of_cpusets++;
        return 0;
 }
 
+void cpuset_reparent(struct container_subsys *ss, struct container *cont,
+                    void *state)
+{
+       struct cpuset *cs = state;
+       cs->css.container = cont;
+}
+
 /*
  * Locking note on the strange update_flag() call below:
  *
@@ -1156,9 +1205,9 @@ int cpuset_create(struct container *cont
  * nesting would risk an ABBA deadlock.
  */
 
-void cpuset_destroy(struct container *cont)
+void cpuset_destroy(struct container_subsys *ss, struct container *cont)
 {
-       struct cpuset *cs = cont->cpuset;
+       struct cpuset *cs = container_cs(cont);
 
        cpuset_update_task_memory_state();
        if (is_cpu_exclusive(cs)) {
@@ -1168,6 +1217,17 @@ void cpuset_destroy(struct container *co
        number_of_cpusets--;
 }
 
+static struct container_subsys cpuset_subsys = {
+       .name = "cpuset",
+       .create = cpuset_create,
+       .destroy  = cpuset_destroy,
+       .can_attach = cpuset_can_attach,
+       .attach = cpuset_attach,
+       .post_attach = cpuset_post_attach,
+       .populate = cpuset_populate,
+       .subsys_id = -1,
+};
+
 /*
  * cpuset_init_early - just enough so that the calls to
  * cpuset_update_task_memory_state() in early init code
@@ -1176,13 +1236,13 @@ void cpuset_destroy(struct container *co
 
 int __init cpuset_init_early(void)
 {
-       struct container *cont = current->container;
-       cont->cpuset = &top_cpuset;
-       top_cpuset.container = cont;
-       cont->cpuset->mems_generation = cpuset_mems_generation++;
+       if (container_register_subsys(&cpuset_subsys) < 0)
+               panic("Couldn't register cpuset subsystem");
+       top_cpuset.mems_generation = cpuset_mems_generation++;
        return 0;
 }
 
+
 /**
  * cpuset_init - initialize cpusets at system boot
  *
@@ -1192,6 +1252,7 @@ int __init cpuset_init_early(void)
 int __init cpuset_init(void)
 {
        int err = 0;
+
        top_cpuset.cpus_allowed = CPU_MASK_ALL;
        top_cpuset.mems_allowed = NODE_MASK_ALL;
 
@@ -1234,7 +1295,7 @@ static void guarantee_online_cpus_mems_i
        struct cpuset *c;
 
        /* Each of our child cpusets mems must be online */
-       list_for_each_entry(c, &cur->container->children, sibling) {
+       list_for_each_entry(c, &cur->css.container->children, sibling) {
                c = container_cs(cont);
                guarantee_online_cpus_mems_in_subtree(c);
                if (!cpus_empty(c->cpus_allowed))
@@ -1336,7 +1397,7 @@ cpumask_t cpuset_cpus_allowed(struct tas
 
        container_lock();
        task_lock(tsk);
-       guarantee_online_cpus(tsk->container->cpuset, &mask);
+       guarantee_online_cpus(task_cs(tsk), &mask);
        task_unlock(tsk);
        container_unlock();
 
@@ -1364,7 +1425,7 @@ nodemask_t cpuset_mems_allowed(struct ta
 
        container_lock();
        task_lock(tsk);
-       guarantee_online_mems(tsk->container->cpuset, &mask);
+       guarantee_online_mems(task_cs(tsk), &mask);
        task_unlock(tsk);
        container_unlock();
 
@@ -1469,7 +1530,7 @@ int __cpuset_zone_allowed(struct zone *z
        container_lock();
 
        task_lock(current);
-       cs = nearest_exclusive_ancestor(current->container->cpuset);
+       cs = nearest_exclusive_ancestor(task_cs(current));
        task_unlock(current);
 
        allowed = node_isset(node, cs->mems_allowed);
@@ -1537,7 +1598,7 @@ int cpuset_excl_nodes_overlap(const stru
                task_unlock(current);
                goto done;
        }
-       cs1 = nearest_exclusive_ancestor(current->container->cpuset);
+       cs1 = nearest_exclusive_ancestor(task_cs(current));
        task_unlock(current);
 
        task_lock((struct task_struct *)p);
@@ -1545,7 +1606,7 @@ int cpuset_excl_nodes_overlap(const stru
                task_unlock((struct task_struct *)p);
                goto done;
        }
-       cs2 = nearest_exclusive_ancestor(p->container->cpuset);
+       cs2 = nearest_exclusive_ancestor(task_cs((struct task_struct *)p));
        task_unlock((struct task_struct *)p);
 
        overlap = nodes_intersects(cs1->mems_allowed, cs2->mems_allowed);
@@ -1581,11 +1642,8 @@ int cpuset_memory_pressure_enabled __rea
 
 void __cpuset_memory_pressure_bump(void)
 {
-       struct cpuset *cs;
-
        task_lock(current);
-       cs = current->container->cpuset;
-       fmeter_markevent(&cs->fmeter);
+       fmeter_markevent(&task_cs(current)->fmeter);
        task_unlock(current);
 }
 
Index: container-2.6.19-rc5/Documentation/containers.txt
===================================================================
--- container-2.6.19-rc5.orig/Documentation/containers.txt
+++ container-2.6.19-rc5/Documentation/containers.txt
@@ -17,12 +17,16 @@ CONTENTS:
   1.2 Why are containers needed ?
   1.3 How are containers implemented ?
   1.4 What does notify_on_release do ?
-  1.5 How do I use containers ?
+  1.5 What do the xxx_enabled files do ?
+  1.6 How do I use containers ?
 2. Usage Examples and Syntax
   2.1 Basic Usage
   2.2 Attaching processes
-3. Questions
-4. Contact
+3. Kernel API
+  3.1 Overview
+  3.2 Synchronization
+  3.3 Subsystem API
+4. Questions
 
 1. Containers
 ==========
@@ -31,13 +35,17 @@ CONTENTS:
 ----------------------
 
 Containers provide a mechanism for aggregating sets of tasks, and all
-their children, into hierarchical groups.
-
-Each task has a pointer to a container.  Multiple tasks may reference
-the same container. User level code may create and destroy containers
-by name in the container virtual file system, specify and query to
-which container a task is assigned, and list the task pids assigned to
-a container.
+their children, into hierarchical groups. A container associates a set
+of tasks with a set of parameters for one or more "subsystems"
+(typically resource controllers).
+
+At any one time there may be up to CONFIG_MAX_CONTAINER_HIERARCHIES
+active hierachies of tasks.  Each task has a pointer to a container in
+each active hierarchy.  Multiple tasks may reference the same
+container. User level code may create and destroy containers by name
+in an instance of the container virtual file system, specify and query
+to which container a task is assigned, and list the task pids assigned
+to a container.
 
 On their own, the only use for containers is for simple job
 tracking. The intention is that other subsystems, such as cpusets (see
@@ -67,27 +75,43 @@ desired.
 
 Containers extends the kernel as follows:
 
- - Each task in the system is attached to a container, via a pointer
-   in the task structure to a reference counted container structure.
- - The hierarchy of containers can be mounted at /dev/container (or
-   elsewhere), for browsing and manipulation from user space.
+ - Each task in the system has set of reference-counted container
+   pointers, one for each active hierarchy
+ - A container hierarchy filesystem can be mounted  for browsing and
+   manipulation from user space.
  - You can list all the tasks (by pid) attached to any container.
 
 The implementation of containers requires a few, simple hooks
 into the rest of the kernel, none in performance critical paths:
 
- - in init/main.c, to initialize the root container at system boot.
- - in fork and exit, to attach and detach a task from its container.
+ - in init/main.c, to initialize the root containers at system boot.
+ - in fork and exit, to attach and detach a task from its containers.
 
 In addition a new file system, of type "container" may be mounted,
-typically at /dev/container, to enable browsing and modifying the containers
-presently known to the kernel.  No new system calls are added for
-containers - all support for querying and modifying containers is via
-this container file system.
-
-Each task under /proc has an added file named 'container', displaying
-the container name, as the path relative to the root of the container file
-system.
+typically at /dev/container, to enable browsing and modifying the
+containers presently known to the kernel.  When mounting a container
+hierarchy, you may specify a comma-separated list of subsystems to
+mount as the filesystem mount options.  By default, mounting the
+container filesystem attempts to mount a hierarchy containing all
+registered subsystems.
+
+If an active hierarchy with exactly the same set of subsystems already
+exists, it will be reused for the new mount. If no existing hierarchy
+matches, and any of the requested subsystems are in use in an existing
+hierarchy, the mount will fail with -EBUSY. Otherwise, a new hierarchy
+is created, associated with the requested subsystems.
+
+When a container filesystem is unmounted, if there are any
+subcontainers created below the top-level container, that hierarchy
+will remain active even though unmounted; if there are no
+subcontainers then the hierarchy will be deactivated.
+
+No new system calls are added for containers - all support for
+querying and modifying containers is via this container file system.
+
+Each task under /proc has an added file named 'container' displaying,
+for each active hierarchy, the subsystem names and the container name
+as the path relative to the root of the container file system.
 
 Each container is represented by a directory in the container file system
 containing the following files describing that container:
@@ -129,7 +153,18 @@ The default value of notify_on_release i
 boot is disabled (0).  The default value of other containers at creation
 is the current value of their parents notify_on_release setting.
 
-1.5 How do I use containers ?
+1.5 What do the xxx_enabled files do ?
+--------------------------------------
+
+In the top-level container directory there are a series of
+<subsys>_enabled files, one for each registered subsystem.  Each of
+these files contains 0 or 1 to indicate whether the named container
+subsystem is enabled, and can only be modified when there are no
+subcontainers.  Disabled container subsystems don't get new instances
+created when a subcontainer is created; the subsystem-specific state
+is simply inherited from the parent container.
+
+1.6 How do I use containers ?
 --------------------------
 
 To start a new job that is to be contained within a container, the steps are:
@@ -214,8 +249,145 @@ If you have several tasks to attach, you
        ...
 # /bin/echo PIDn > tasks
 
+3. Kernel API
+=============
+
+3.1 Overview
+------------
+
+Each kernel subsystem that wants to hook into the generic container
+system needs to create a container_subsys object. This contains
+various methods, which are callbacks from the container system, along
+with a subsystem id which will be assigned by the container system.
+
+Other fields in the container_subsys object include:
+
+- subsys_id: a unique array index for the subsystem, indicating which
+  entry in container->subsys[] this subsystem should be
+  managing. Initialized by container_register_subsys(); prior to this
+  it should be initialized to -1
+
+- top_container: the subsystem-specific state representing the root
+  container in the system. This should always exist, even when the
+  subsystem isn't attached to a hierarchy.
+
+- hierarchy: an index indicating which hierarchy, if any, this
+  subsystem is currently attached to. If this is -1, then the
+  subsystem is not attached to any hierarchy, and all tasks should be
+  considered to be members of the subsystem's top_container. It should
+  be initialized to -1.
+
+- name: should be initialized to a unique subsystem name prior to
+  calling container_register_subsystem. Should be no longer than
+  MAX_CONTAINER_TYPE_NAMELEN
+
+Each container object created by the system has an array of pointers,
+indexed by subsystem id; this pointer is entirely managed by the
+subsystem; the generic container code will never touch this pointer.
+
+3.2 Synchronization
+-------------------
+
+There are two global mutexes used by the container system. The first
+is the manage_mutex, which should be taken by anything that wants to
+modify a container; The second if the callback_mutex, which should be
+taken by holders of the manage_mutex at the point when they actually
+make changes, and by callbacks from lower-level subsystems that want
+to ensure that no container changes occur.  Note that memory
+allocations cannot be made while holding callback_mutex.
+
+The callback_mutex nests inside the manage_mutex.
+
+In general, the pattern of use is:
+
+1) take manage_mutex
+2) verify that the change is valid and do any necessary allocations\
+3) take callback_mutex
+4) make changes
+5) release callback_mutex
+6) release manage_mutex
+
+See kernel/container.c for more details.
+
+Subsystems can take/release the manage_mutex via the functions
+container_manage_lock()/container_manage_unlock(), and can
+take/release the callback_mutex via the functions
+container_lock()/container_unlock().
+
+Accessing a task's container pointer may be done in the following ways:
+- while holding manage_mutex
+- while holding callback_mutex
+- while holding the task's alloc_lock (via task_lock())
+- inside an rcu_read_lock() section via rcu_dereference()
+
+3.3 Subsystem API
+--------------------------
+
+Each subsystem should call container_register_subsys() with a pointer
+to its subsystem object. This will store the new subsystem id in the
+subsystem subsys_id field and return 0, or a negative error.  There's
+currently no facility for deregestering a subsystem nor for
+registering a subsystem after any containers (other than the default
+"top_container") have been created.
+
+Each subsystem may export the following methods. The only mandatory
+methods are create/destroy. Any others that are null are presumed to
+be successful no-ops.
+
+int create(struct container *cont)
+LL=manage_mutex
+
+The subsystem should appropriately initialize its subsystem pointer
+for the passed container, returning 0 on success or a negative error
+code. Typically this will involve allocating a new per-container
+structure and storing a reference to it in the container, but there's
+nothing to stop a subsystem having multiple containers with pointers
+to the same subsystem object.  Note that this will be called during
+container_register_subsys() to initialize this subsystem on the root
+container.
+
+void destroy(struct container *cont)
+LL=manage_mutex
+
+The container system is about to destroy the passed container; the
+subsystem should do any necessary cleanup
+
+int can_attach(struct container_subsys *ss, struct container *cont,
+              struct task_struct *task)
+LL=manage_mutex
+
+Called prior to moving a task into a container; if the subsystem
+returns an error, this will abort the attach operation.  Note that
+this isn't called on a fork.
+
+void attach(struct container_subsys *ss, struct container *cont,
+           struct container *old_cont, struct task_struct *task)
+LL=manage_mutex & callback_mutex
+
+Called during the attach operation.  The subsystem should do any
+necessary work that can be accomplished without memory allocations or
+sleeping.
+
+void post_attach(struct container_subsys *ss, struct container *cont,
+                struct container *old_cont, struct task_struct *task)
+LL=manage_mutex
+
+Called after the task has been attached to the container, to allow any
+post-attachment activity that requires memory allocations or blocking.
+
+int populate(struct container_subsys *ss, struct container *cont)
+LL=none
+
+Called after creation of a container to allow a subsystem to populate
+the container directory with file entries.  The subsystem should make
+calls to container_add_file() with objects of type cftype (see
+include/linux/container.h for details).  Called during
+container_register_subsys() to populate the root container.  Note that
+although this method can return an error code, the error code is
+currently not always handled well.
+
 
-3. Questions
+4. Questions
 ============
 
 Q: what's up with this '/bin/echo' ?
Index: container-2.6.19-rc5/include/linux/mempolicy.h
===================================================================
--- container-2.6.19-rc5.orig/include/linux/mempolicy.h
+++ container-2.6.19-rc5/include/linux/mempolicy.h
@@ -148,14 +148,6 @@ extern void mpol_rebind_task(struct task
                                        const nodemask_t *new);
 extern void mpol_rebind_mm(struct mm_struct *mm, nodemask_t *new);
 extern void mpol_fix_fork_child_flag(struct task_struct *p);
-#define set_cpuset_being_rebound(x) (cpuset_being_rebound = (x))
-
-#ifdef CONFIG_CPUSETS
-#define current_cpuset_is_being_rebound() \
- (cpuset_being_rebound == current->container->cpuset)
-#else
-#define current_cpuset_is_being_rebound() 0
-#endif
 
 extern struct mempolicy default_policy;
 extern struct zonelist *huge_zonelist(struct vm_area_struct *vma,
@@ -173,8 +165,6 @@ static inline void check_highest_zone(en
 int do_migrate_pages(struct mm_struct *mm,
        const nodemask_t *from_nodes, const nodemask_t *to_nodes, int flags);
 
-extern void *cpuset_being_rebound;     /* Trigger mpol_copy vma rebind */
-
 #else
 
 struct mempolicy {};
@@ -253,8 +243,6 @@ static inline void mpol_fix_fork_child_f
 {
 }
 
-#define set_cpuset_being_rebound(x) do {} while (0)
-
 static inline struct zonelist *huge_zonelist(struct vm_area_struct *vma,
                unsigned long addr)
 {
Index: container-2.6.19-rc5/include/linux/sched.h
===================================================================
--- container-2.6.19-rc5.orig/include/linux/sched.h
+++ container-2.6.19-rc5/include/linux/sched.h
@@ -1005,7 +1005,7 @@ struct task_struct {
        int cpuset_mem_spread_rotor;
 #endif
 #ifdef CONFIG_CONTAINERS
-       struct container *container;
+       struct container *container[CONFIG_MAX_CONTAINER_HIERARCHIES];
 #endif
        struct robust_list_head __user *robust_list;
 #ifdef CONFIG_COMPAT
Index: container-2.6.19-rc5/mm/mempolicy.c
===================================================================
--- container-2.6.19-rc5.orig/mm/mempolicy.c
+++ container-2.6.19-rc5/mm/mempolicy.c
@@ -1307,7 +1307,6 @@ EXPORT_SYMBOL(alloc_pages_current);
  * keeps mempolicies cpuset relative after its cpuset moves.  See
  * further kernel/cpuset.c update_nodemask().
  */
-void *cpuset_being_rebound;
 
 /* Slow path of a mempolicy copy */
 struct mempolicy *__mpol_copy(struct mempolicy *old)
@@ -1906,4 +1905,3 @@ out:
                m->version = (vma != priv->tail_vma) ? vma->vm_start : 0;
        return 0;
 }
-

--
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to