From: Xiaoguang Wang <wangxg.fnst@cn.fujitsu.com>
To: ltp@lists.linux.it
Subject: [LTP] [PATCH v2] inotify: Add test for inotify mark destruction race
Date: Thu, 14 Apr 2016 16:14:25 +0800 [thread overview]
Message-ID: <570F5161.7060806@cn.fujitsu.com> (raw)
In-Reply-To: <20160414081516.GA2753@quack2.suse.cz>
hello,
On 04/14/2016 04:15 PM, Jan Kara wrote:
> Hello,
>
> On Thu 14-04-16 10:06:59, Xiaoguang Wang wrote:
>> On 08/25/2015 07:29 PM, Cyril Hrubis wrote:
>>> Hi!
>>>> Interesting, probably SRCU is much slower with this older kernel. From my
>>>> experiments 100 iterations isn't quite reliable to trigger the oops in my
>>>> testing instance. But 400 seem to be good enough.
>>>
>>> I've changed the nuber of iterations to 400 and pushed it to git,
>>> thanks.
>>>
>>
>> In upstream kernel v4.6-rc3-17-g1c74a7f and RHEL7.2GA, I sometimes get such
>> error:
>> ---------------------------------------------------------------------------
>> inotify06 1 TBROK : inotify06.c:104: inotify_init failed: errno=EMFILE(24): Too many open files
>> inotify06 2 TBROK : inotify06.c:104: Remaining cases broken
>> ---------------------------------------------------------------------------
>> But look at the inotify06.c, inotify_fd is closed every iteration.
>> For normal file descriptors, "close(fd) succeeds" does not mean related kernel
>> resources have been released immediately(processes may still reference fd).
>>
>> Then inotify_fd also has similar behavior? Even close(inotify_fd) returns,
>> that does not mean the number of current inotify instances have decreased one
>> immediately, then later inotify_init() calls may exceeds the /proc/sys/fs/inotify/max_user_instances and
>> return EMFILE error? I had added some debug code in kernel, it seems that close(inotify_fd)
>> does not make sure current inotify instances decreases one immediately.
>>
>> So I'd like to know this is expected behavior for inotify? If yes, we can
>> echo 400 > /proc/sys/fs/inotify/max_user_instances to avoid EMFILE error.
>> If not, this is a kernel bug?
>
> Interesting, I've never seen this. Number of inotify instances is maintaned
> immediately - i.e., it is dropped as soon as the last descriptor pointing to
> the instance is closed. So I'm not sure how what you describe can happen.
> How do you reproduce the issue?
I just call ./inotify06 directly, and about 50% chance, it'll fail(return EMFILE).
Regards,
Xiaoguang Wang
>
> Honza
>
next prev parent reply other threads:[~2016-04-14 8:14 UTC|newest]
Thread overview: 13+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-08-10 14:04 [LTP] [PATCH v2] inotify: Add test for inotify mark destruction race Jan Kara
2015-08-11 14:14 ` Cyril Hrubis
[not found] ` <20150811142035.GD2659@quack.suse.cz>
2015-08-25 9:29 ` Cyril Hrubis
[not found] ` <20150825103803.GA15280@quack.suse.cz>
2015-08-25 11:29 ` Cyril Hrubis
2016-04-14 2:06 ` Xiaoguang Wang
2016-04-14 8:15 ` Jan Kara
2016-04-14 8:14 ` Xiaoguang Wang [this message]
2016-04-14 8:46 ` Jan Kara
2016-04-18 3:37 ` Xiaoguang Wang
2016-04-19 13:05 ` Jan Kara
2016-04-26 10:42 ` Jan Kara
2016-04-27 4:48 ` Xiaoguang Wang
2016-04-27 7:58 ` Jan Kara
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=570F5161.7060806@cn.fujitsu.com \
--to=wangxg.fnst@cn.fujitsu.com \
--cc=ltp@lists.linux.it \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox