From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-yw0-f193.google.com ([209.85.161.193]:35128 "EHLO mail-yw0-f193.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726473AbeGKJAy (ORCPT ); Wed, 11 Jul 2018 05:00:54 -0400 Received: by mail-yw0-f193.google.com with SMTP id t18-v6so8875119ywg.2 for ; Wed, 11 Jul 2018 01:57:38 -0700 (PDT) MIME-Version: 1.0 In-Reply-To: <20180710100217.12866-10-jack@suse.cz> References: <20180710100217.12866-1-jack@suse.cz> <20180710100217.12866-10-jack@suse.cz> From: Amir Goldstein Date: Wed, 11 Jul 2018 11:57:37 +0300 Message-ID: Subject: Re: [PATCH 09/10] audit: Allocate fsnotify mark independently of chunk To: Jan Kara Cc: Linux Audit , Paul Moore , linux-fsdevel , Al Viro , Richard Guy Briggs Content-Type: text/plain; charset="UTF-8" Sender: linux-fsdevel-owner@vger.kernel.org List-ID: On Tue, Jul 10, 2018 at 1:02 PM, Jan Kara wrote: > Allocate fsnotify mark independently instead of embedding it inside > chunk. This will allow us to just replace chunk attached to mark when > growing / shrinking chunk instead of replacing mark attached to inode > which is a more complex operation. > > Signed-off-by: Jan Kara Ack. Thanks for separating this patch. Amir. > --- > kernel/audit_tree.c | 59 ++++++++++++++++++++++++++++++++++++++++------------- > 1 file changed, 45 insertions(+), 14 deletions(-) > > diff --git a/kernel/audit_tree.c b/kernel/audit_tree.c > index bce3b04a365d..aec9b27a20ff 100644 > --- a/kernel/audit_tree.c > +++ b/kernel/audit_tree.c > @@ -25,7 +25,7 @@ struct audit_tree { > struct audit_chunk { > struct list_head hash; > unsigned long key; > - struct fsnotify_mark mark; > + struct fsnotify_mark *mark; > struct list_head trees; /* with root here */ > int dead; > int count; > @@ -38,6 +38,11 @@ struct audit_chunk { > } owners[]; > }; > > +struct audit_tree_mark { > + struct fsnotify_mark fsn_mark; > + struct audit_chunk *chunk; > +}; > + > static LIST_HEAD(tree_list); > static LIST_HEAD(prune_list); > static struct task_struct *prune_thread; > @@ -73,6 +78,7 @@ static struct task_struct *prune_thread; > */ > > static struct fsnotify_group *audit_tree_group; > +static struct kmem_cache *audit_tree_mark_cachep __read_mostly; > > static struct audit_tree *alloc_tree(const char *s) > { > @@ -142,10 +148,28 @@ static void audit_mark_put_chunk(struct audit_chunk *chunk) > call_rcu(&chunk->head, __put_chunk); > } > > +static inline struct audit_tree_mark *AUDIT_M(struct fsnotify_mark *entry) > +{ > + return container_of(entry, struct audit_tree_mark, fsn_mark); > +} > + > static void audit_tree_destroy_watch(struct fsnotify_mark *entry) > { > - struct audit_chunk *chunk = container_of(entry, struct audit_chunk, mark); > + struct audit_chunk *chunk = AUDIT_M(entry)->chunk; > audit_mark_put_chunk(chunk); > + kmem_cache_free(audit_tree_mark_cachep, entry); > +} > + > +static struct fsnotify_mark *alloc_fsnotify_mark(void) > +{ > + struct audit_tree_mark *mark; > + > + mark = kmem_cache_zalloc(audit_tree_mark_cachep, GFP_KERNEL); > + if (!mark) > + return NULL; > + fsnotify_init_mark(&mark->fsn_mark, audit_tree_group); > + mark->fsn_mark.mask = FS_IN_IGNORED; > + return &mark->fsn_mark; > } > > static struct audit_chunk *alloc_chunk(int count) > @@ -159,6 +183,13 @@ static struct audit_chunk *alloc_chunk(int count) > if (!chunk) > return NULL; > > + chunk->mark = alloc_fsnotify_mark(); > + if (!chunk->mark) { > + kfree(chunk); > + return NULL; > + } > + AUDIT_M(chunk->mark)->chunk = chunk; > + > INIT_LIST_HEAD(&chunk->hash); > INIT_LIST_HEAD(&chunk->trees); > chunk->count = count; > @@ -167,8 +198,6 @@ static struct audit_chunk *alloc_chunk(int count) > INIT_LIST_HEAD(&chunk->owners[i].list); > chunk->owners[i].index = i; > } > - fsnotify_init_mark(&chunk->mark, audit_tree_group); > - chunk->mark.mask = FS_IN_IGNORED; > return chunk; > } > > @@ -278,7 +307,7 @@ static void replace_chunk(struct audit_chunk *new, struct audit_chunk *old, > static void untag_chunk(struct node *p) > { > struct audit_chunk *chunk = find_chunk(p); > - struct fsnotify_mark *entry = &chunk->mark; > + struct fsnotify_mark *entry = chunk->mark; > struct audit_chunk *new = NULL; > struct audit_tree *owner; > int size = chunk->count - 1; > @@ -298,7 +327,7 @@ static void untag_chunk(struct node *p) > if (chunk->dead || !(entry->flags & FSNOTIFY_MARK_FLAG_ATTACHED)) { > mutex_unlock(&entry->group->mark_mutex); > if (new) > - fsnotify_put_mark(&new->mark); > + fsnotify_put_mark(new->mark); > goto out; > } > > @@ -322,9 +351,9 @@ static void untag_chunk(struct node *p) > if (!new) > goto Fallback; > > - if (fsnotify_add_mark_locked(&new->mark, entry->connector->obj, > + if (fsnotify_add_mark_locked(new->mark, entry->connector->obj, > FSNOTIFY_OBJ_TYPE_INODE, 1)) { > - fsnotify_put_mark(&new->mark); > + fsnotify_put_mark(new->mark); > goto Fallback; > } > > @@ -344,7 +373,7 @@ static void untag_chunk(struct node *p) > fsnotify_detach_mark(entry); > mutex_unlock(&entry->group->mark_mutex); > fsnotify_free_mark(entry); > - fsnotify_put_mark(&new->mark); /* drop initial reference */ > + fsnotify_put_mark(new->mark); /* drop initial reference */ > goto out; > > Fallback: > @@ -375,7 +404,7 @@ static int create_chunk(struct inode *inode, struct audit_tree *tree) > return -ENOMEM; > } > > - entry = &chunk->mark; > + entry = chunk->mark; > if (fsnotify_add_inode_mark_locked(entry, inode, 0)) { > mutex_unlock(&audit_tree_group->mark_mutex); > fsnotify_put_mark(entry); > @@ -426,7 +455,7 @@ static int tag_chunk(struct inode *inode, struct audit_tree *tree) > if (!old_entry) > return create_chunk(inode, tree); > > - old = container_of(old_entry, struct audit_chunk, mark); > + old = AUDIT_M(old_entry)->chunk; > > /* are we already there? */ > spin_lock(&hash_lock); > @@ -447,7 +476,7 @@ static int tag_chunk(struct inode *inode, struct audit_tree *tree) > return -ENOMEM; > } > > - chunk_entry = &chunk->mark; > + chunk_entry = chunk->mark; > > /* > * mark_mutex protects mark from getting detached and thus also from > @@ -457,7 +486,7 @@ static int tag_chunk(struct inode *inode, struct audit_tree *tree) > /* old_entry is being shot, lets just lie */ > mutex_unlock(&audit_tree_group->mark_mutex); > fsnotify_put_mark(old_entry); > - fsnotify_put_mark(&chunk->mark); > + fsnotify_put_mark(chunk->mark); > return -ENOENT; > } > > @@ -1009,7 +1038,7 @@ static int audit_tree_handle_event(struct fsnotify_group *group, > > static void audit_tree_freeing_mark(struct fsnotify_mark *entry, struct fsnotify_group *group) > { > - struct audit_chunk *chunk = container_of(entry, struct audit_chunk, mark); > + struct audit_chunk *chunk = AUDIT_M(entry)->chunk; > > evict_chunk(chunk); > > @@ -1030,6 +1059,8 @@ static int __init audit_tree_init(void) > { > int i; > > + audit_tree_mark_cachep = KMEM_CACHE(audit_tree_mark, SLAB_PANIC); > + > audit_tree_group = fsnotify_alloc_group(&audit_tree_ops); > if (IS_ERR(audit_tree_group)) > audit_panic("cannot initialize fsnotify group for rectree watches"); > -- > 2.16.4 >