From: Joseph Qi <jiangqi903@gmail.com>
To: Changwei Ge <ge.changwei@h3c.com>, Larry Chen <lchen@suse.com>,
"mark@fasheh.com" <mark@fasheh.com>,
"jlbec@evilplan.org" <jlbec@evilplan.org>
Cc: "linux-kernel@vger.kernel.org" <linux-kernel@vger.kernel.org>,
"ocfs2-devel@oss.oracle.com" <ocfs2-devel@oss.oracle.com>,
Andrew Morton <akpm@linux-foundation.org>
Subject: Re: [Ocfs2-devel] [PATCH V3] ocfs2: fix dead lock caused by ocfs2_defrag_extent
Date: Fri, 2 Nov 2018 09:18:05 +0800 [thread overview]
Message-ID: <011125b5-6343-6f4e-b420-3f152f395980@gmail.com> (raw)
In-Reply-To: <63ADC13FD55D6546B7DECE290D39E37301277DE3FE@H3CMLB12-EX.srv.huawei-3com.com>
On 18/11/1 20:34, Changwei Ge wrote:
> Hello Joseph,
>
> On 2018/11/1 20:16, Joseph Qi wrote:
>>
>>
>> On 18/11/1 19:52, Changwei Ge wrote:
>>> Hello Joseph,
>>>
>>> On 2018/11/1 17:01, Joseph Qi wrote:
>>>> Hi Larry,
>>>>
>>>> On 18/11/1 15:14, Larry Chen wrote:
>>>>> ocfs2_defrag_extent may fall into deadlock.
>>>>>
>>>>> ocfs2_ioctl_move_extents
>>>>> ocfs2_ioctl_move_extents
>>>>> ocfs2_move_extents
>>>>> ocfs2_defrag_extent
>>>>> ocfs2_lock_allocators_move_extents
>>>>>
>>>>> ocfs2_reserve_clusters
>>>>> inode_lock GLOBAL_BITMAP_SYSTEM_INODE
>>>>>
>>>>> __ocfs2_flush_truncate_log
>>>>> inode_lock GLOBAL_BITMAP_SYSTEM_INODE
>>>>>
>>>>> As backtrace shows above, ocfs2_reserve_clusters() will call inode_lock
>>>>> against the global bitmap if local allocator has not sufficient cluters.
>>>>> Once global bitmap could meet the demand, ocfs2_reserve_cluster will
>>>>> return success with global bitmap locked.
>>>>>
>>>>> After ocfs2_reserve_cluster(), if truncate log is full,
>>>>> __ocfs2_flush_truncate_log() will definitely fall into deadlock because it
>>>>> needs to inode_lock global bitmap, which has already been locked.
>>>>>
>>>>> To fix this bug, we could remove from ocfs2_lock_allocators_move_extents()
>>>>> the code which intends to lock global allocator, and put the removed code
>>>>> after __ocfs2_flush_truncate_log().
>>>>>
>>>>> ocfs2_lock_allocators_move_extents() is referred by 2 places, one is here,
>>>>> the other does not need the data allocator context, which means this patch
>>>>> does not affect the caller so far.
>>>>>
>>>>> Change log:
>>>>> 1. Correct the function comment.
>>>>> 2. Remove unused argument from ocfs2_lock_meta_allocator_move_extents.
>>>>>
>>>>> Signed-off-by: Larry Chen <lchen@suse.com>
>>>>> ---
>>>>> fs/ocfs2/move_extents.c | 47 ++++++++++++++++++++++++++---------------------
>>>>> 1 file changed, 26 insertions(+), 21 deletions(-)
>>>>>
>>>
>>>> IMO, here clusters_to_move is only for data_ac, since we change this
>>>> function to only handle meta_ac, I'm afraid clusters_to_move related
>>>> logic has to been changed correspondingly.
>>>
>>> I think we can't remove *clusters_to_move* from here as clusters can be reserved latter outsides this function, but we
>>> still have to reserve metadata(extents) in advance.
>>> So we need that argument.
>>>
>> I was not saying just remove it.
>> IIUC, clusters_to_move is for reserving data clusters (for meta_ac, we
>
> Um...
> *cluster_to_move* is not only used for reserving data clusters.
> It is also an effecting input for calculating if existed extents still have enough free records for later
> tree operation like attaching clusters to extents.
>
> Please refer to below code:
> 175 unsigned int max_recs_needed = 2 * extents_to_split + clusters_to_move;
>
IC. It is a bit odd that calculate it here but do the real reserve out.
>
>
>> mostly talk about blocks). Since we have moved data cluster reserve
>> logic out of ocfs2_lock_allocators_move_extents() now, then left
>> clusters_to_move related logic here is odd.
>
> Like my preceding elaboration, it is used for telling if we need more extents.
> Anyway, I think we must keep *cluster_to_move* here as it used to. :-)
>
> Thanks,
> Changwei
>
>
>
>
>>
>>>>> u32 extents_to_split,
>>>>> struct ocfs2_alloc_context **meta_ac,
>>>>> - struct ocfs2_alloc_context **data_ac,
>>>>> int extra_blocks,
>>>>> int *credits)
>>>>> {
>>>>> @@ -192,13 +188,6 @@ static int ocfs2_lock_allocators_move_extents(struct inode *inode,
>>>>> goto out;
>>>>> }
>>>>>
>>>>> - if (data_ac) {
>>>>> - ret = ocfs2_reserve_clusters(osb, clusters_to_move, data_ac);
>>>>> - if (ret) {
>>>>> - mlog_errno(ret);
>>>>> - goto out;
>>>>> - }
>>>>> - }
>>>>>
>>>>> *credits += ocfs2_calc_extend_credits(osb->sb, et->et_root_el);
>>>>>
>>>>> @@ -257,10 +246,10 @@ static int ocfs2_defrag_extent(struct ocfs2_move_extents_context *context,
>>>>> }
>>>>> }
>>>>>
>>>>> - ret = ocfs2_lock_allocators_move_extents(inode, &context->et, *len, 1,
>>>>> - &context->meta_ac,
>>>>> - &context->data_ac,
>>>>> - extra_blocks, &credits);
>>>>> + ret = ocfs2_lock_meta_allocator_move_extents(inode, &context->et,
>>>>> + *len, 1,
>>>>> + &context->meta_ac,
>>>>> + extra_blocks, &credits);
>>>>> if (ret) {
>>>>> mlog_errno(ret);
>>>>> goto out;
>>>>> @@ -283,6 +272,21 @@ static int ocfs2_defrag_extent(struct ocfs2_move_extents_context *context,
>>>>> }
>>>>> }
>>>>>
>>>>> + /*
>>>>> + * Make sure ocfs2_reserve_cluster is called after
>>>>> + * __ocfs2_flush_truncate_log, otherwise, dead lock may happen.
>>>>> + *
>>>>> + * If ocfs2_reserve_cluster is called
>>>>> + * before __ocfs2_flush_truncate_log, dead lock on global bitmap
>>>>> + * may happen.
>>>>> + *
>>>>> + */
>>>>> + ret = ocfs2_reserve_clusters(osb, *len, &context->data_ac);
>>>>> + if (ret) {
>>>>> + mlog_errno(ret);
>>>>> + goto out_unlock_mutex;
>>>>> + }
>>>>> +
>>>>> handle = ocfs2_start_trans(osb, credits);
>>>>> if (IS_ERR(handle)) {
>>>>> ret = PTR_ERR(handle);
>>>>> @@ -600,9 +604,10 @@ static int ocfs2_move_extent(struct ocfs2_move_extents_context *context,
>>>>> }
>>>>> }
>>>>>
>>>>> - ret = ocfs2_lock_allocators_move_extents(inode, &context->et, len, 1,
>>>>> - &context->meta_ac,
>>>>> - NULL, extra_blocks, &credits);
>>>>> + ret = ocfs2_lock_meta_allocator_move_extents(inode, &context->et,
>>>>> + len, 1,
>>>>> + &context->meta_ac,
>>>>> + extra_blocks, &credits);
>>>>> if (ret) {
>>>>> mlog_errno(ret);
>>>>> goto out;
>>>>>
>>>>
>>>> _______________________________________________
>>>> Ocfs2-devel mailing list
>>>> Ocfs2-devel@oss.oracle.com
>>>> https://oss.oracle.com/mailman/listinfo/ocfs2-devel
>>>>
>>
next prev parent reply other threads:[~2018-11-02 1:18 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-11-01 7:14 [PATCH V3] ocfs2: fix dead lock caused by ocfs2_defrag_extent Larry Chen
2018-11-01 8:58 ` [Ocfs2-devel] " Joseph Qi
2018-11-01 11:52 ` Changwei Ge
2018-11-01 12:15 ` Joseph Qi
2018-11-01 12:34 ` Changwei Ge
2018-11-02 1:18 ` Joseph Qi [this message]
2018-11-01 12:39 ` Larry Chen
2018-11-01 12:48 ` Changwei Ge
2018-11-02 0:53 ` Changwei Ge
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=011125b5-6343-6f4e-b420-3f152f395980@gmail.com \
--to=jiangqi903@gmail.com \
--cc=akpm@linux-foundation.org \
--cc=ge.changwei@h3c.com \
--cc=jlbec@evilplan.org \
--cc=lchen@suse.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mark@fasheh.com \
--cc=ocfs2-devel@oss.oracle.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).