linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH v2] kernel/ucounts: expose count of inotify watches in use
@ 2019-02-01 20:39 Albert Vaca Cintora
  2019-02-22 17:58 ` Albert Vaca Cintora
  2019-04-25 20:06 ` Andrew Morton
  0 siblings, 2 replies; 5+ messages in thread
From: Albert Vaca Cintora @ 2019-02-01 20:39 UTC (permalink / raw)
  To: albertvaka, akpm, rdunlap, mingo, jack, ebiederm, nsaenzjulienne,
	linux-kernel

Adds a readonly 'current_inotify_watches' entry to the user sysctl table.
The handler for this entry is a custom function that ends up calling
proc_dointvec. Said sysctl table already contains 'max_inotify_watches'
and it gets mounted under /proc/sys/user/.

Inotify watches are a finite resource, in a similar way to available file
descriptors. The motivation for this patch is to be able to set up
monitoring and alerting before an application starts failing because
it runs out of inotify watches.

Signed-off-by: Albert Vaca Cintora <albertvaka@gmail.com>
Acked-by: Jan Kara <jack@suse.cz>
Reviewed-by: Nicolas Saenz Julienne <nsaenzjulienne@suse.de>
---
 kernel/ucount.c | 29 +++++++++++++++++++++++++++++
 1 file changed, 29 insertions(+)

diff --git a/kernel/ucount.c b/kernel/ucount.c
index f48d1b6376a4..d8b11e53f098 100644
--- a/kernel/ucount.c
+++ b/kernel/ucount.c
@@ -57,6 +57,11 @@ static struct ctl_table_root set_root = {
 	.permissions = set_permissions,
 };
 
+#ifdef CONFIG_INOTIFY_USER
+int proc_read_inotify_watches(struct ctl_table *table, int write,
+		     void __user *buffer, size_t *lenp, loff_t *ppos);
+#endif
+
 static int zero = 0;
 static int int_max = INT_MAX;
 #define UCOUNT_ENTRY(name)				\
@@ -79,6 +84,12 @@ static struct ctl_table user_table[] = {
 #ifdef CONFIG_INOTIFY_USER
 	UCOUNT_ENTRY("max_inotify_instances"),
 	UCOUNT_ENTRY("max_inotify_watches"),
+	{
+		.procname	= "current_inotify_watches",
+		.maxlen		= sizeof(int),
+		.mode		= 0444,
+		.proc_handler	= proc_read_inotify_watches,
+	},
 #endif
 	{ }
 };
@@ -226,6 +237,24 @@ void dec_ucount(struct ucounts *ucounts, enum ucount_type type)
 	put_ucounts(ucounts);
 }
 
+#ifdef CONFIG_INOTIFY_USER
+int proc_read_inotify_watches(struct ctl_table *table, int write,
+		     void __user *buffer, size_t *lenp, loff_t *ppos)
+{
+	struct ucounts *ucounts;
+	struct ctl_table fake_table;
+	int count;
+
+	ucounts = get_ucounts(current_user_ns(), current_euid());
+	count = atomic_read(&ucounts->ucount[UCOUNT_INOTIFY_WATCHES]);
+	put_ucounts(ucounts);
+
+	fake_table.data = &count;
+	fake_table.maxlen = sizeof(count);
+	return proc_dointvec(&fake_table, write, buffer, lenp, ppos);
+}
+#endif
+
 static __init int user_namespace_sysctl_init(void)
 {
 #ifdef CONFIG_SYSCTL
-- 
2.20.1


^ permalink raw reply related	[flat|nested] 5+ messages in thread

* Re: [PATCH v2] kernel/ucounts: expose count of inotify watches in use
  2019-02-01 20:39 [PATCH v2] kernel/ucounts: expose count of inotify watches in use Albert Vaca Cintora
@ 2019-02-22 17:58 ` Albert Vaca Cintora
  2019-04-25 15:58   ` Matthias Brugger
  2019-04-25 20:06 ` Andrew Morton
  1 sibling, 1 reply; 5+ messages in thread
From: Albert Vaca Cintora @ 2019-02-22 17:58 UTC (permalink / raw)
  To: Albert Vaca Cintora, akpm, rdunlap, mingo, Jan Kara, ebiederm,
	Nicolas Saenz Julienne, linux-kernel, mbrugger

On Fri, Feb 1, 2019 at 9:42 PM Albert Vaca Cintora <albertvaka@gmail.com> wrote:
>
> Adds a readonly 'current_inotify_watches' entry to the user sysctl table.
> The handler for this entry is a custom function that ends up calling
> proc_dointvec. Said sysctl table already contains 'max_inotify_watches'
> and it gets mounted under /proc/sys/user/.
>
> Inotify watches are a finite resource, in a similar way to available file
> descriptors. The motivation for this patch is to be able to set up
> monitoring and alerting before an application starts failing because
> it runs out of inotify watches.
>
> Signed-off-by: Albert Vaca Cintora <albertvaka@gmail.com>
> Acked-by: Jan Kara <jack@suse.cz>
> Reviewed-by: Nicolas Saenz Julienne <nsaenzjulienne@suse.de>

Friendly ping. Any comments on this?

> ---
>  kernel/ucount.c | 29 +++++++++++++++++++++++++++++
>  1 file changed, 29 insertions(+)
>
> diff --git a/kernel/ucount.c b/kernel/ucount.c
> index f48d1b6376a4..d8b11e53f098 100644
> --- a/kernel/ucount.c
> +++ b/kernel/ucount.c
> @@ -57,6 +57,11 @@ static struct ctl_table_root set_root = {
>         .permissions = set_permissions,
>  };
>
> +#ifdef CONFIG_INOTIFY_USER
> +int proc_read_inotify_watches(struct ctl_table *table, int write,
> +                    void __user *buffer, size_t *lenp, loff_t *ppos);
> +#endif
> +
>  static int zero = 0;
>  static int int_max = INT_MAX;
>  #define UCOUNT_ENTRY(name)                             \
> @@ -79,6 +84,12 @@ static struct ctl_table user_table[] = {
>  #ifdef CONFIG_INOTIFY_USER
>         UCOUNT_ENTRY("max_inotify_instances"),
>         UCOUNT_ENTRY("max_inotify_watches"),
> +       {
> +               .procname       = "current_inotify_watches",
> +               .maxlen         = sizeof(int),
> +               .mode           = 0444,
> +               .proc_handler   = proc_read_inotify_watches,
> +       },
>  #endif
>         { }
>  };
> @@ -226,6 +237,24 @@ void dec_ucount(struct ucounts *ucounts, enum ucount_type type)
>         put_ucounts(ucounts);
>  }
>
> +#ifdef CONFIG_INOTIFY_USER
> +int proc_read_inotify_watches(struct ctl_table *table, int write,
> +                    void __user *buffer, size_t *lenp, loff_t *ppos)
> +{
> +       struct ucounts *ucounts;
> +       struct ctl_table fake_table;
> +       int count;
> +
> +       ucounts = get_ucounts(current_user_ns(), current_euid());
> +       count = atomic_read(&ucounts->ucount[UCOUNT_INOTIFY_WATCHES]);
> +       put_ucounts(ucounts);
> +
> +       fake_table.data = &count;
> +       fake_table.maxlen = sizeof(count);
> +       return proc_dointvec(&fake_table, write, buffer, lenp, ppos);
> +}
> +#endif
> +
>  static __init int user_namespace_sysctl_init(void)
>  {
>  #ifdef CONFIG_SYSCTL
> --
> 2.20.1
>

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [PATCH v2] kernel/ucounts: expose count of inotify watches in use
  2019-02-22 17:58 ` Albert Vaca Cintora
@ 2019-04-25 15:58   ` Matthias Brugger
  0 siblings, 0 replies; 5+ messages in thread
From: Matthias Brugger @ 2019-04-25 15:58 UTC (permalink / raw)
  To: Albert Vaca Cintora, akpm, rdunlap, mingo, Jan Kara, ebiederm,
	Nicolas Saenz Julienne, linux-kernel



On 22/02/2019 18:58, Albert Vaca Cintora wrote:
> On Fri, Feb 1, 2019 at 9:42 PM Albert Vaca Cintora <albertvaka@gmail.com> wrote:
>>
>> Adds a readonly 'current_inotify_watches' entry to the user sysctl table.
>> The handler for this entry is a custom function that ends up calling
>> proc_dointvec. Said sysctl table already contains 'max_inotify_watches'
>> and it gets mounted under /proc/sys/user/.
>>
>> Inotify watches are a finite resource, in a similar way to available file
>> descriptors. The motivation for this patch is to be able to set up
>> monitoring and alerting before an application starts failing because
>> it runs out of inotify watches.
>>
>> Signed-off-by: Albert Vaca Cintora <albertvaka@gmail.com>
>> Acked-by: Jan Kara <jack@suse.cz>
>> Reviewed-by: Nicolas Saenz Julienne <nsaenzjulienne@suse.de>
> 
> Friendly ping. Any comments on this?
> 

Any comments on this? Just to make it clear, Albert found this problem while
working on montitoring software, so it fixes a real problem out there.

Regards,
Matthias

>> ---
>>  kernel/ucount.c | 29 +++++++++++++++++++++++++++++
>>  1 file changed, 29 insertions(+)
>>
>> diff --git a/kernel/ucount.c b/kernel/ucount.c
>> index f48d1b6376a4..d8b11e53f098 100644
>> --- a/kernel/ucount.c
>> +++ b/kernel/ucount.c
>> @@ -57,6 +57,11 @@ static struct ctl_table_root set_root = {
>>         .permissions = set_permissions,
>>  };
>>
>> +#ifdef CONFIG_INOTIFY_USER
>> +int proc_read_inotify_watches(struct ctl_table *table, int write,
>> +                    void __user *buffer, size_t *lenp, loff_t *ppos);
>> +#endif
>> +
>>  static int zero = 0;
>>  static int int_max = INT_MAX;
>>  #define UCOUNT_ENTRY(name)                             \
>> @@ -79,6 +84,12 @@ static struct ctl_table user_table[] = {
>>  #ifdef CONFIG_INOTIFY_USER
>>         UCOUNT_ENTRY("max_inotify_instances"),
>>         UCOUNT_ENTRY("max_inotify_watches"),
>> +       {
>> +               .procname       = "current_inotify_watches",
>> +               .maxlen         = sizeof(int),
>> +               .mode           = 0444,
>> +               .proc_handler   = proc_read_inotify_watches,
>> +       },
>>  #endif
>>         { }
>>  };
>> @@ -226,6 +237,24 @@ void dec_ucount(struct ucounts *ucounts, enum ucount_type type)
>>         put_ucounts(ucounts);
>>  }
>>
>> +#ifdef CONFIG_INOTIFY_USER
>> +int proc_read_inotify_watches(struct ctl_table *table, int write,
>> +                    void __user *buffer, size_t *lenp, loff_t *ppos)
>> +{
>> +       struct ucounts *ucounts;
>> +       struct ctl_table fake_table;
>> +       int count;
>> +
>> +       ucounts = get_ucounts(current_user_ns(), current_euid());
>> +       count = atomic_read(&ucounts->ucount[UCOUNT_INOTIFY_WATCHES]);
>> +       put_ucounts(ucounts);
>> +
>> +       fake_table.data = &count;
>> +       fake_table.maxlen = sizeof(count);
>> +       return proc_dointvec(&fake_table, write, buffer, lenp, ppos);
>> +}
>> +#endif
>> +
>>  static __init int user_namespace_sysctl_init(void)
>>  {
>>  #ifdef CONFIG_SYSCTL
>> --
>> 2.20.1
>>
> 

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [PATCH v2] kernel/ucounts: expose count of inotify watches in use
  2019-02-01 20:39 [PATCH v2] kernel/ucounts: expose count of inotify watches in use Albert Vaca Cintora
  2019-02-22 17:58 ` Albert Vaca Cintora
@ 2019-04-25 20:06 ` Andrew Morton
  2019-04-25 21:39   ` Albert Vaca Cintora
  1 sibling, 1 reply; 5+ messages in thread
From: Andrew Morton @ 2019-04-25 20:06 UTC (permalink / raw)
  To: Albert Vaca Cintora
  Cc: rdunlap, mingo, jack, ebiederm, nsaenzjulienne, linux-kernel,
	Matthias Brugger

On Fri,  1 Feb 2019 21:39:59 +0100 Albert Vaca Cintora <albertvaka@gmail.com> wrote:

> Adds a readonly 'current_inotify_watches' entry to the user sysctl table.
> The handler for this entry is a custom function that ends up calling
> proc_dointvec. Said sysctl table already contains 'max_inotify_watches'
> and it gets mounted under /proc/sys/user/.
> 
> Inotify watches are a finite resource, in a similar way to available file
> descriptors. The motivation for this patch is to be able to set up
> monitoring and alerting before an application starts failing because
> it runs out of inotify watches.

Matthias said "Albert found this problem while working on montitoring
software, so it fixes a real problem out there", so please include full
details of the problem which you encountered so that we are better able
to understand the value of the patch.

>
> ...
>
>  kernel/ucount.c | 29 +++++++++++++++++++++++++++++

Documentation, please.  Documentation/filesystems/inotify.txt and/or
Documentation/filesystems/proc.txt.

Also, max_inotify_instances (at least) also appears to be undocumented,
so it would be good to address this as well while you're in there.

> 
> diff --git a/kernel/ucount.c b/kernel/ucount.c
> index f48d1b6376a4..d8b11e53f098 100644
> --- a/kernel/ucount.c
> +++ b/kernel/ucount.c
> @@ -57,6 +57,11 @@ static struct ctl_table_root set_root = {
>  	.permissions = set_permissions,
>  };
>  
> +#ifdef CONFIG_INOTIFY_USER
> +int proc_read_inotify_watches(struct ctl_table *table, int write,
> +		     void __user *buffer, size_t *lenp, loff_t *ppos);
> +#endif

The ifdefs aren't really needed.  And this should be in a header file
if it is indeed to be non-static.

But it should be static, in which case the ifdef will be needed to
prevent a warning.  It's kinda irksome and perhaps it would be better
to move proc_read_inotify_watches() to be ahead of user_table[].


>  static int zero = 0;
>  static int int_max = INT_MAX;
>  #define UCOUNT_ENTRY(name)				\
> @@ -79,6 +84,12 @@ static struct ctl_table user_table[] = {
>  #ifdef CONFIG_INOTIFY_USER
>  	UCOUNT_ENTRY("max_inotify_instances"),
>  	UCOUNT_ENTRY("max_inotify_watches"),
> +	{
> +		.procname	= "current_inotify_watches",
> +		.maxlen		= sizeof(int),
> +		.mode		= 0444,
> +		.proc_handler	= proc_read_inotify_watches,
> +	},
>  #endif
>  	{ }
>  };
> @@ -226,6 +237,24 @@ void dec_ucount(struct ucounts *ucounts, enum ucount_type type)
>  	put_ucounts(ucounts);
>  }
>  
> +#ifdef CONFIG_INOTIFY_USER
> +int proc_read_inotify_watches(struct ctl_table *table, int write,
> +		     void __user *buffer, size_t *lenp, loff_t *ppos)
> +{
> +	struct ucounts *ucounts;
> +	struct ctl_table fake_table;
> +	int count;
> +
> +	ucounts = get_ucounts(current_user_ns(), current_euid());

get_ucounts() can return NULL.  The kernel will crash.

> +	count = atomic_read(&ucounts->ucount[UCOUNT_INOTIFY_WATCHES]);
> +	put_ucounts(ucounts);
> +
> +	fake_table.data = &count;
> +	fake_table.maxlen = sizeof(count);
> +	return proc_dointvec(&fake_table, write, buffer, lenp, ppos);
> +}
> +#endif


^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [PATCH v2] kernel/ucounts: expose count of inotify watches in use
  2019-04-25 20:06 ` Andrew Morton
@ 2019-04-25 21:39   ` Albert Vaca Cintora
  0 siblings, 0 replies; 5+ messages in thread
From: Albert Vaca Cintora @ 2019-04-25 21:39 UTC (permalink / raw)
  To: Andrew Morton
  Cc: rdunlap, mingo, Jan Kara, ebiederm, Nicolas Saenz Julienne,
	linux-kernel, Matthias Brugger

On Thu, Apr 25, 2019 at 10:07 PM Andrew Morton
<akpm@linux-foundation.org> wrote:
>
> On Fri,  1 Feb 2019 21:39:59 +0100 Albert Vaca Cintora <albertvaka@gmail.com> wrote:
>
> > Adds a readonly 'current_inotify_watches' entry to the user sysctl table.
> > The handler for this entry is a custom function that ends up calling
> > proc_dointvec. Said sysctl table already contains 'max_inotify_watches'
> > and it gets mounted under /proc/sys/user/.
> >
> > Inotify watches are a finite resource, in a similar way to available file
> > descriptors. The motivation for this patch is to be able to set up
> > monitoring and alerting before an application starts failing because
> > it runs out of inotify watches.
>
> Matthias said "Albert found this problem while working on montitoring
> software, so it fixes a real problem out there", so please include full
> details of the problem which you encountered so that we are better able
> to understand the value of the patch.
>

This is an important metric to track as a sysadmin. Currently,
monitoring software have to workaround the lack of a single metric by
iterating all the file descriptors for all processes and checking
which ones are inotify watches [1].

If this seems enough justification to you, please say so and I will
submit a patch v3 with the requested implementation changes.

[1] https://github.com/prometheus/node_exporter/blob/master/text_collector_examples/inotify-instances

Albert

^ permalink raw reply	[flat|nested] 5+ messages in thread

end of thread, other threads:[~2019-04-25 21:40 UTC | newest]

Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-02-01 20:39 [PATCH v2] kernel/ucounts: expose count of inotify watches in use Albert Vaca Cintora
2019-02-22 17:58 ` Albert Vaca Cintora
2019-04-25 15:58   ` Matthias Brugger
2019-04-25 20:06 ` Andrew Morton
2019-04-25 21:39   ` Albert Vaca Cintora

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).