* [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse
[not found] <567E3C07.2000305@huawei.com>
@ 2015-12-28 7:44 ` xuejiufei
2015-12-30 2:52 ` Junxiao Bi
0 siblings, 1 reply; 8+ messages in thread
From: xuejiufei @ 2015-12-28 7:44 UTC (permalink / raw)
To: ocfs2-devel
We have found that migration source will trigger a BUG that the
refcount of mle is already zero before put when the target is
down during migration. The situation is as follows:
dlm_migrate_lockres
dlm_add_migration_mle
dlm_mark_lockres_migrating
dlm_get_mle_inuse
<<<<<< Now the refcount of the mle is 2.
dlm_send_one_lockres and wait for the target to become the
new master.
<<<<<< o2hb detect the target down and clean the migration
mle. Now the refcount is 1.
dlm_migrate_lockres woken, and put the mle twice when found
the target goes down which trigger the BUG with the following
message:
"ERROR: bad mle: ".
Signed-off-by: Jiufei Xue <xuejiufei@huawei.com>
Reviewed-by: Joseph Qi <joseph.qi@huawei.com>
---
fs/ocfs2/dlm/dlmmaster.c | 26 +++++++++++++++-----------
1 file changed, 15 insertions(+), 11 deletions(-)
diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
index 936e11b..b713140 100644
--- a/fs/ocfs2/dlm/dlmmaster.c
+++ b/fs/ocfs2/dlm/dlmmaster.c
@@ -2519,6 +2519,11 @@ static int dlm_migrate_lockres(struct dlm_ctxt *dlm,
spin_lock(&dlm->master_lock);
ret = dlm_add_migration_mle(dlm, res, mle, &oldmle, name,
namelen, target, dlm->node_num);
+ /* get an extra reference on the mle.
+ * otherwise the assert_master from the new
+ * master will destroy this.
+ */
+ dlm_get_mle_inuse(mle);
spin_unlock(&dlm->master_lock);
spin_unlock(&dlm->spinlock);
@@ -2554,6 +2559,7 @@ fail:
if (mle_added) {
dlm_mle_detach_hb_events(dlm, mle);
dlm_put_mle(mle);
+ dlm_put_mle_inuse(mle);
} else if (mle) {
kmem_cache_free(dlm_mle_cache, mle);
mle = NULL;
@@ -2571,17 +2577,6 @@ fail:
* ensure that all assert_master work is flushed. */
flush_workqueue(dlm->dlm_worker);
- /* get an extra reference on the mle.
- * otherwise the assert_master from the new
- * master will destroy this.
- * also, make sure that all callers of dlm_get_mle
- * take both dlm->spinlock and dlm->master_lock */
- spin_lock(&dlm->spinlock);
- spin_lock(&dlm->master_lock);
- dlm_get_mle_inuse(mle);
- spin_unlock(&dlm->master_lock);
- spin_unlock(&dlm->spinlock);
-
/* notify new node and send all lock state */
/* call send_one_lockres with migration flag.
* this serves as notice to the target node that a
@@ -3312,6 +3307,15 @@ top:
mle->new_master != dead_node)
continue;
+ if (mle->new_master == dead_node && mle->inuse) {
+ mlog(ML_NOTICE, "%s: target %u died during "
+ "migration from %u, the MLE is "
+ "still keep used, ignore it!\n",
+ dlm->name, dead_node,
+ mle->master);
+ continue;
+ }
+
/* If we have reached this point, this mle needs to be
* removed from the list and freed. */
dlm_clean_migration_mle(dlm, mle);
--
1.8.4.3
^ permalink raw reply related [flat|nested] 8+ messages in thread
* [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse
2015-12-28 7:44 ` [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse xuejiufei
@ 2015-12-30 2:52 ` Junxiao Bi
2015-12-30 9:56 ` xuejiufei
0 siblings, 1 reply; 8+ messages in thread
From: Junxiao Bi @ 2015-12-30 2:52 UTC (permalink / raw)
To: ocfs2-devel
Hi Jiufei,
When target node down, mle is cleared from
dlm_do_local_recovery_cleanup()->dlm_clean_master_list()->dlm_clean_migration_mle()?
mle->woken is set to 1 in dlm_clean_migration_mle(), so the code to
detect target node down(if (dlm_is_node_dead(dlm, target))) will never
be run in dlm_migrate_lockres()?
2621 while (1) {
2622 ret = wait_event_interruptible_timeout(mle->wq,
2623 (atomic_read(&mle->woken)
== 1),
2624 msecs_to_jiffies(5000));
2625
2626 if (ret >= 0) {
2627 if (atomic_read(&mle->woken) == 1 ||
2628 res->owner == target)
2629 break;
2630
2631 mlog(0, "%s:%.*s: timed out during
migration\n",
2632 dlm->name, res->lockname.len,
res->lockname.name);
2633 /* avoid hang during shutdown when
migrating lockres
2634 * to a node which also goes down */
2635 if (dlm_is_node_dead(dlm, target)) {
2636 mlog(0, "%s:%.*s: expected migration "
2637 "target %u is no longer up,
restarting\n",
2638 dlm->name, res->lockname.len,
2639 res->lockname.name, target);
2640 ret = -EINVAL;
2641 /* migration failed, detach and
clean up mle */
2642 dlm_mle_detach_hb_events(dlm, mle);
2643 dlm_put_mle(mle);
2644 dlm_put_mle_inuse(mle);
2645 spin_lock(&res->spinlock);
2646 res->state &= ~DLM_LOCK_RES_MIGRATING;
2647 wake = 1;
2648 spin_unlock(&res->spinlock);
2649 goto leave;
2650 }
2651 } else
2652 mlog(0, "%s:%.*s: caught signal during
migration\n",
2653 dlm->name, res->lockname.len,
res->lockname.name);
2654 }
Thanks,
Junxiao.
On 12/28/2015 03:44 PM, xuejiufei wrote:
> We have found that migration source will trigger a BUG that the
> refcount of mle is already zero before put when the target is
> down during migration. The situation is as follows:
>
> dlm_migrate_lockres
> dlm_add_migration_mle
> dlm_mark_lockres_migrating
> dlm_get_mle_inuse
> <<<<<< Now the refcount of the mle is 2.
> dlm_send_one_lockres and wait for the target to become the
> new master.
> <<<<<< o2hb detect the target down and clean the migration
> mle. Now the refcount is 1.
>
> dlm_migrate_lockres woken, and put the mle twice when found
> the target goes down which trigger the BUG with the following
> message:
> "ERROR: bad mle: ".
>
> Signed-off-by: Jiufei Xue <xuejiufei@huawei.com>
> Reviewed-by: Joseph Qi <joseph.qi@huawei.com>
> ---
> fs/ocfs2/dlm/dlmmaster.c | 26 +++++++++++++++-----------
> 1 file changed, 15 insertions(+), 11 deletions(-)
>
> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
> index 936e11b..b713140 100644
> --- a/fs/ocfs2/dlm/dlmmaster.c
> +++ b/fs/ocfs2/dlm/dlmmaster.c
> @@ -2519,6 +2519,11 @@ static int dlm_migrate_lockres(struct dlm_ctxt *dlm,
> spin_lock(&dlm->master_lock);
> ret = dlm_add_migration_mle(dlm, res, mle, &oldmle, name,
> namelen, target, dlm->node_num);
> + /* get an extra reference on the mle.
> + * otherwise the assert_master from the new
> + * master will destroy this.
> + */
> + dlm_get_mle_inuse(mle);
> spin_unlock(&dlm->master_lock);
> spin_unlock(&dlm->spinlock);
>
> @@ -2554,6 +2559,7 @@ fail:
> if (mle_added) {
> dlm_mle_detach_hb_events(dlm, mle);
> dlm_put_mle(mle);
> + dlm_put_mle_inuse(mle);
> } else if (mle) {
> kmem_cache_free(dlm_mle_cache, mle);
> mle = NULL;
> @@ -2571,17 +2577,6 @@ fail:
> * ensure that all assert_master work is flushed. */
> flush_workqueue(dlm->dlm_worker);
>
> - /* get an extra reference on the mle.
> - * otherwise the assert_master from the new
> - * master will destroy this.
> - * also, make sure that all callers of dlm_get_mle
> - * take both dlm->spinlock and dlm->master_lock */
> - spin_lock(&dlm->spinlock);
> - spin_lock(&dlm->master_lock);
> - dlm_get_mle_inuse(mle);
> - spin_unlock(&dlm->master_lock);
> - spin_unlock(&dlm->spinlock);
> -
> /* notify new node and send all lock state */
> /* call send_one_lockres with migration flag.
> * this serves as notice to the target node that a
> @@ -3312,6 +3307,15 @@ top:
> mle->new_master != dead_node)
> continue;
>
> + if (mle->new_master == dead_node && mle->inuse) {
> + mlog(ML_NOTICE, "%s: target %u died during "
> + "migration from %u, the MLE is "
> + "still keep used, ignore it!\n",
> + dlm->name, dead_node,
> + mle->master);
> + continue;
> + }
> +
> /* If we have reached this point, this mle needs to be
> * removed from the list and freed. */
> dlm_clean_migration_mle(dlm, mle);
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse
2015-12-30 2:52 ` Junxiao Bi
@ 2015-12-30 9:56 ` xuejiufei
2015-12-31 3:05 ` Junxiao Bi
0 siblings, 1 reply; 8+ messages in thread
From: xuejiufei @ 2015-12-30 9:56 UTC (permalink / raw)
To: ocfs2-devel
Hi Junxiao,
You are right. But it may happen that mle->woken is set to 1 in
dlm_clean_migration_mle() just after atomic_read() in
dlm_migrate_lockres(). Actually we trigger this BUG when dlm_send_one_lockres()
return error.
And I think dlm_migrate_lockres() should not set owner to target and return 0
when mle->woken is set to 1 in dlm_clean_migration_mle(). This is another problem?
Thanks
Jiufei.
On 2015/12/30 10:52, Junxiao Bi wrote:
> Hi Jiufei,
>
> When target node down, mle is cleared from
> dlm_do_local_recovery_cleanup()->dlm_clean_master_list()->dlm_clean_migration_mle()?
> mle->woken is set to 1 in dlm_clean_migration_mle(), so the code to
> detect target node down(if (dlm_is_node_dead(dlm, target))) will never
> be run in dlm_migrate_lockres()?
>
>
> 2621 while (1) {
> 2622 ret = wait_event_interruptible_timeout(mle->wq,
> 2623 (atomic_read(&mle->woken)
> == 1),
> 2624 msecs_to_jiffies(5000));
> 2625
> 2626 if (ret >= 0) {
> 2627 if (atomic_read(&mle->woken) == 1 ||
> 2628 res->owner == target)
> 2629 break;
> 2630
> 2631 mlog(0, "%s:%.*s: timed out during
> migration\n",
> 2632 dlm->name, res->lockname.len,
> res->lockname.name);
> 2633 /* avoid hang during shutdown when
> migrating lockres
> 2634 * to a node which also goes down */
> 2635 if (dlm_is_node_dead(dlm, target)) {
> 2636 mlog(0, "%s:%.*s: expected migration "
> 2637 "target %u is no longer up,
> restarting\n",
> 2638 dlm->name, res->lockname.len,
> 2639 res->lockname.name, target);
> 2640 ret = -EINVAL;
> 2641 /* migration failed, detach and
> clean up mle */
> 2642 dlm_mle_detach_hb_events(dlm, mle);
> 2643 dlm_put_mle(mle);
> 2644 dlm_put_mle_inuse(mle);
> 2645 spin_lock(&res->spinlock);
> 2646 res->state &= ~DLM_LOCK_RES_MIGRATING;
> 2647 wake = 1;
> 2648 spin_unlock(&res->spinlock);
> 2649 goto leave;
> 2650 }
> 2651 } else
> 2652 mlog(0, "%s:%.*s: caught signal during
> migration\n",
> 2653 dlm->name, res->lockname.len,
> res->lockname.name);
> 2654 }
>
>
> Thanks,
> Junxiao.
> On 12/28/2015 03:44 PM, xuejiufei wrote:
>> We have found that migration source will trigger a BUG that the
>> refcount of mle is already zero before put when the target is
>> down during migration. The situation is as follows:
>>
>> dlm_migrate_lockres
>> dlm_add_migration_mle
>> dlm_mark_lockres_migrating
>> dlm_get_mle_inuse
>> <<<<<< Now the refcount of the mle is 2.
>> dlm_send_one_lockres and wait for the target to become the
>> new master.
>> <<<<<< o2hb detect the target down and clean the migration
>> mle. Now the refcount is 1.
>>
>> dlm_migrate_lockres woken, and put the mle twice when found
>> the target goes down which trigger the BUG with the following
>> message:
>> "ERROR: bad mle: ".
>>
>> Signed-off-by: Jiufei Xue <xuejiufei@huawei.com>
>> Reviewed-by: Joseph Qi <joseph.qi@huawei.com>
>> ---
>> fs/ocfs2/dlm/dlmmaster.c | 26 +++++++++++++++-----------
>> 1 file changed, 15 insertions(+), 11 deletions(-)
>>
>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>> index 936e11b..b713140 100644
>> --- a/fs/ocfs2/dlm/dlmmaster.c
>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>> @@ -2519,6 +2519,11 @@ static int dlm_migrate_lockres(struct dlm_ctxt *dlm,
>> spin_lock(&dlm->master_lock);
>> ret = dlm_add_migration_mle(dlm, res, mle, &oldmle, name,
>> namelen, target, dlm->node_num);
>> + /* get an extra reference on the mle.
>> + * otherwise the assert_master from the new
>> + * master will destroy this.
>> + */
>> + dlm_get_mle_inuse(mle);
>> spin_unlock(&dlm->master_lock);
>> spin_unlock(&dlm->spinlock);
>>
>> @@ -2554,6 +2559,7 @@ fail:
>> if (mle_added) {
>> dlm_mle_detach_hb_events(dlm, mle);
>> dlm_put_mle(mle);
>> + dlm_put_mle_inuse(mle);
>> } else if (mle) {
>> kmem_cache_free(dlm_mle_cache, mle);
>> mle = NULL;
>> @@ -2571,17 +2577,6 @@ fail:
>> * ensure that all assert_master work is flushed. */
>> flush_workqueue(dlm->dlm_worker);
>>
>> - /* get an extra reference on the mle.
>> - * otherwise the assert_master from the new
>> - * master will destroy this.
>> - * also, make sure that all callers of dlm_get_mle
>> - * take both dlm->spinlock and dlm->master_lock */
>> - spin_lock(&dlm->spinlock);
>> - spin_lock(&dlm->master_lock);
>> - dlm_get_mle_inuse(mle);
>> - spin_unlock(&dlm->master_lock);
>> - spin_unlock(&dlm->spinlock);
>> -
>> /* notify new node and send all lock state */
>> /* call send_one_lockres with migration flag.
>> * this serves as notice to the target node that a
>> @@ -3312,6 +3307,15 @@ top:
>> mle->new_master != dead_node)
>> continue;
>>
>> + if (mle->new_master == dead_node && mle->inuse) {
>> + mlog(ML_NOTICE, "%s: target %u died during "
>> + "migration from %u, the MLE is "
>> + "still keep used, ignore it!\n",
>> + dlm->name, dead_node,
>> + mle->master);
>> + continue;
>> + }
>> +
>> /* If we have reached this point, this mle needs to be
>> * removed from the list and freed. */
>> dlm_clean_migration_mle(dlm, mle);
>>
>
>
> .
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse
2015-12-30 9:56 ` xuejiufei
@ 2015-12-31 3:05 ` Junxiao Bi
2015-12-31 7:15 ` xuejiufei
0 siblings, 1 reply; 8+ messages in thread
From: Junxiao Bi @ 2015-12-31 3:05 UTC (permalink / raw)
To: ocfs2-devel
On 12/30/2015 05:56 PM, xuejiufei wrote:
> Hi Junxiao,
> You are right. But it may happen that mle->woken is set to 1 in
> dlm_clean_migration_mle() just after atomic_read() in
> dlm_migrate_lockres(). Actually we trigger this BUG when dlm_send_one_lockres()
> return error.
Yes, that's possible because of that 5s timeout wakeup, I think this
timeout is useless and can be removed?
>
> And I think dlm_migrate_lockres() should not set owner to target and return 0
> when mle->woken is set to 1 in dlm_clean_migration_mle(). This is another problem?
Yes, it should be fixed, or the lockres owner will be set to a down node
wrongly. Can the following fix these two issue?
diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
index 84f2f8079466..d0380ea62340 100644
--- a/fs/ocfs2/dlm/dlmmaster.c
+++ b/fs/ocfs2/dlm/dlmmaster.c
@@ -2618,39 +2618,22 @@ fail:
/* wait for new node to assert master */
- while (1) {
- ret = wait_event_interruptible_timeout(mle->wq,
- (atomic_read(&mle->woken) == 1),
- msecs_to_jiffies(5000));
-
- if (ret >= 0) {
- if (atomic_read(&mle->woken) == 1 ||
- res->owner == target)
- break;
-
- mlog(0, "%s:%.*s: timed out during migration\n",
- dlm->name, res->lockname.len, res->lockname.name);
- /* avoid hang during shutdown when migrating lockres
- * to a node which also goes down */
- if (dlm_is_node_dead(dlm, target)) {
- mlog(0, "%s:%.*s: expected migration "
- "target %u is no longer up, restarting\n",
- dlm->name, res->lockname.len,
- res->lockname.name, target);
- ret = -EINVAL;
- /* migration failed, detach and clean up mle */
- dlm_mle_detach_hb_events(dlm, mle);
- dlm_put_mle(mle);
- dlm_put_mle_inuse(mle);
- spin_lock(&res->spinlock);
- res->state &= ~DLM_LOCK_RES_MIGRATING;
- wake = 1;
- spin_unlock(&res->spinlock);
- goto leave;
- }
- } else
- mlog(0, "%s:%.*s: caught signal during migration\n",
- dlm->name, res->lockname.len, res->lockname.name);
+ wait_event(mle->wq, ((atomic_read(&mle->woken) == 1) ||
+ (atomic_read(&mle->woken) == 2)));
+
+ /* migrate target down */
+ if (atomic_read(&mle->woken) == 2) {
+ mlog(0, "%s:%.*s: expected migration "
+ "target %u is no longer up, restarting\n",
+ dlm->name, res->lockname.len,
+ res->lockname.name, target);
+ ret = -EINVAL;
+ dlm_put_mle_inuse(mle);
+ spin_lock(&res->spinlock);
+ res->state &= ~DLM_LOCK_RES_MIGRATING;
+ wake = 1;
+ spin_unlock(&res->spinlock);
+ goto leave;
}
/* all done, set the owner, clear the flag */
@@ -3227,7 +3210,7 @@ static void dlm_clean_migration_mle(struct
dlm_ctxt *dlm,
spin_lock(&mle->spinlock);
__dlm_unlink_mle(dlm, mle);
- atomic_set(&mle->woken, 1);
+ atomic_set(&mle->woken, 2);
spin_unlock(&mle->spinlock);
wake_up(&mle->wq);
Thanks,
Junxiao.
>
> Thanks
> Jiufei.
>
> On 2015/12/30 10:52, Junxiao Bi wrote:
>> Hi Jiufei,
>>
>> When target node down, mle is cleared from
>> dlm_do_local_recovery_cleanup()->dlm_clean_master_list()->dlm_clean_migration_mle()?
>> mle->woken is set to 1 in dlm_clean_migration_mle(), so the code to
>> detect target node down(if (dlm_is_node_dead(dlm, target))) will never
>> be run in dlm_migrate_lockres()?
>>
>>
>> 2621 while (1) {
>> 2622 ret = wait_event_interruptible_timeout(mle->wq,
>> 2623 (atomic_read(&mle->woken)
>> == 1),
>> 2624 msecs_to_jiffies(5000));
>> 2625
>> 2626 if (ret >= 0) {
>> 2627 if (atomic_read(&mle->woken) == 1 ||
>> 2628 res->owner == target)
>> 2629 break;
>> 2630
>> 2631 mlog(0, "%s:%.*s: timed out during
>> migration\n",
>> 2632 dlm->name, res->lockname.len,
>> res->lockname.name);
>> 2633 /* avoid hang during shutdown when
>> migrating lockres
>> 2634 * to a node which also goes down */
>> 2635 if (dlm_is_node_dead(dlm, target)) {
>> 2636 mlog(0, "%s:%.*s: expected migration "
>> 2637 "target %u is no longer up,
>> restarting\n",
>> 2638 dlm->name, res->lockname.len,
>> 2639 res->lockname.name, target);
>> 2640 ret = -EINVAL;
>> 2641 /* migration failed, detach and
>> clean up mle */
>> 2642 dlm_mle_detach_hb_events(dlm, mle);
>> 2643 dlm_put_mle(mle);
>> 2644 dlm_put_mle_inuse(mle);
>> 2645 spin_lock(&res->spinlock);
>> 2646 res->state &= ~DLM_LOCK_RES_MIGRATING;
>> 2647 wake = 1;
>> 2648 spin_unlock(&res->spinlock);
>> 2649 goto leave;
>> 2650 }
>> 2651 } else
>> 2652 mlog(0, "%s:%.*s: caught signal during
>> migration\n",
>> 2653 dlm->name, res->lockname.len,
>> res->lockname.name);
>> 2654 }
>>
>>
>> Thanks,
>> Junxiao.
>> On 12/28/2015 03:44 PM, xuejiufei wrote:
>>> We have found that migration source will trigger a BUG that the
>>> refcount of mle is already zero before put when the target is
>>> down during migration. The situation is as follows:
>>>
>>> dlm_migrate_lockres
>>> dlm_add_migration_mle
>>> dlm_mark_lockres_migrating
>>> dlm_get_mle_inuse
>>> <<<<<< Now the refcount of the mle is 2.
>>> dlm_send_one_lockres and wait for the target to become the
>>> new master.
>>> <<<<<< o2hb detect the target down and clean the migration
>>> mle. Now the refcount is 1.
>>>
>>> dlm_migrate_lockres woken, and put the mle twice when found
>>> the target goes down which trigger the BUG with the following
>>> message:
>>> "ERROR: bad mle: ".
>>>
>>> Signed-off-by: Jiufei Xue <xuejiufei@huawei.com>
>>> Reviewed-by: Joseph Qi <joseph.qi@huawei.com>
>>> ---
>>> fs/ocfs2/dlm/dlmmaster.c | 26 +++++++++++++++-----------
>>> 1 file changed, 15 insertions(+), 11 deletions(-)
>>>
>>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>>> index 936e11b..b713140 100644
>>> --- a/fs/ocfs2/dlm/dlmmaster.c
>>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>>> @@ -2519,6 +2519,11 @@ static int dlm_migrate_lockres(struct dlm_ctxt *dlm,
>>> spin_lock(&dlm->master_lock);
>>> ret = dlm_add_migration_mle(dlm, res, mle, &oldmle, name,
>>> namelen, target, dlm->node_num);
>>> + /* get an extra reference on the mle.
>>> + * otherwise the assert_master from the new
>>> + * master will destroy this.
>>> + */
>>> + dlm_get_mle_inuse(mle);
>>> spin_unlock(&dlm->master_lock);
>>> spin_unlock(&dlm->spinlock);
>>>
>>> @@ -2554,6 +2559,7 @@ fail:
>>> if (mle_added) {
>>> dlm_mle_detach_hb_events(dlm, mle);
>>> dlm_put_mle(mle);
>>> + dlm_put_mle_inuse(mle);
>>> } else if (mle) {
>>> kmem_cache_free(dlm_mle_cache, mle);
>>> mle = NULL;
>>> @@ -2571,17 +2577,6 @@ fail:
>>> * ensure that all assert_master work is flushed. */
>>> flush_workqueue(dlm->dlm_worker);
>>>
>>> - /* get an extra reference on the mle.
>>> - * otherwise the assert_master from the new
>>> - * master will destroy this.
>>> - * also, make sure that all callers of dlm_get_mle
>>> - * take both dlm->spinlock and dlm->master_lock */
>>> - spin_lock(&dlm->spinlock);
>>> - spin_lock(&dlm->master_lock);
>>> - dlm_get_mle_inuse(mle);
>>> - spin_unlock(&dlm->master_lock);
>>> - spin_unlock(&dlm->spinlock);
>>> -
>>> /* notify new node and send all lock state */
>>> /* call send_one_lockres with migration flag.
>>> * this serves as notice to the target node that a
>>> @@ -3312,6 +3307,15 @@ top:
>>> mle->new_master != dead_node)
>>> continue;
>>>
>>> + if (mle->new_master == dead_node && mle->inuse) {
>>> + mlog(ML_NOTICE, "%s: target %u died during "
>>> + "migration from %u, the MLE is "
>>> + "still keep used, ignore it!\n",
>>> + dlm->name, dead_node,
>>> + mle->master);
>>> + continue;
>>> + }
>>> +
>>> /* If we have reached this point, this mle needs to be
>>> * removed from the list and freed. */
>>> dlm_clean_migration_mle(dlm, mle);
>>>
>>
>>
>> .
>>
>
^ permalink raw reply related [flat|nested] 8+ messages in thread
* [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse
2015-12-31 3:05 ` Junxiao Bi
@ 2015-12-31 7:15 ` xuejiufei
2016-01-04 8:20 ` Junxiao Bi
0 siblings, 1 reply; 8+ messages in thread
From: xuejiufei @ 2015-12-31 7:15 UTC (permalink / raw)
To: ocfs2-devel
Hi Junxiao,
On 2015/12/31 11:05, Junxiao Bi wrote:
> On 12/30/2015 05:56 PM, xuejiufei wrote:
>> Hi Junxiao,
>> You are right. But it may happen that mle->woken is set to 1 in
>> dlm_clean_migration_mle() just after atomic_read() in
>> dlm_migrate_lockres(). Actually we trigger this BUG when dlm_send_one_lockres()
>> return error.
> Yes, that's possible because of that 5s timeout wakeup, I think this
> timeout is useless and can be removed?
>>
>> And I think dlm_migrate_lockres() should not set owner to target and return 0
>> when mle->woken is set to 1 in dlm_clean_migration_mle(). This is another problem?
> Yes, it should be fixed, or the lockres owner will be set to a down node
> wrongly. Can the following fix these two issue?
>
It can not fix the first issue when dlm_send_one_lockres() or
dlm_mark_lockres_migrating() return error, right?
>
> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
> index 84f2f8079466..d0380ea62340 100644
> --- a/fs/ocfs2/dlm/dlmmaster.c
> +++ b/fs/ocfs2/dlm/dlmmaster.c
> @@ -2618,39 +2618,22 @@ fail:
>
>
> /* wait for new node to assert master */
> - while (1) {
> - ret = wait_event_interruptible_timeout(mle->wq,
> - (atomic_read(&mle->woken) == 1),
> - msecs_to_jiffies(5000));
> -
> - if (ret >= 0) {
> - if (atomic_read(&mle->woken) == 1 ||
> - res->owner == target)
> - break;
> -
> - mlog(0, "%s:%.*s: timed out during migration\n",
> - dlm->name, res->lockname.len, res->lockname.name);
> - /* avoid hang during shutdown when migrating lockres
> - * to a node which also goes down */
> - if (dlm_is_node_dead(dlm, target)) {
> - mlog(0, "%s:%.*s: expected migration "
> - "target %u is no longer up, restarting\n",
> - dlm->name, res->lockname.len,
> - res->lockname.name, target);
> - ret = -EINVAL;
> - /* migration failed, detach and clean up mle */
> - dlm_mle_detach_hb_events(dlm, mle);
> - dlm_put_mle(mle);
> - dlm_put_mle_inuse(mle);
> - spin_lock(&res->spinlock);
> - res->state &= ~DLM_LOCK_RES_MIGRATING;
> - wake = 1;
> - spin_unlock(&res->spinlock);
> - goto leave;
> - }
> - } else
> - mlog(0, "%s:%.*s: caught signal during migration\n",
> - dlm->name, res->lockname.len, res->lockname.name);
> + wait_event(mle->wq, ((atomic_read(&mle->woken) == 1) ||
> + (atomic_read(&mle->woken) == 2)));
> +
> + /* migrate target down */
> + if (atomic_read(&mle->woken) == 2) {
> + mlog(0, "%s:%.*s: expected migration "
> + "target %u is no longer up, restarting\n",
> + dlm->name, res->lockname.len,
> + res->lockname.name, target);
> + ret = -EINVAL;
> + dlm_put_mle_inuse(mle);
> + spin_lock(&res->spinlock);
> + res->state &= ~DLM_LOCK_RES_MIGRATING;
> + wake = 1;
> + spin_unlock(&res->spinlock);
> + goto leave;
> }
>
> /* all done, set the owner, clear the flag */
> @@ -3227,7 +3210,7 @@ static void dlm_clean_migration_mle(struct
> dlm_ctxt *dlm,
>
> spin_lock(&mle->spinlock);
> __dlm_unlink_mle(dlm, mle);
> - atomic_set(&mle->woken, 1);
> + atomic_set(&mle->woken, 2);
> spin_unlock(&mle->spinlock);
>
> wake_up(&mle->wq);
>
> Thanks,
> Junxiao.
>
>
>>
>> Thanks
>> Jiufei.
>>
>> On 2015/12/30 10:52, Junxiao Bi wrote:
>>> Hi Jiufei,
>>>
>>> When target node down, mle is cleared from
>>> dlm_do_local_recovery_cleanup()->dlm_clean_master_list()->dlm_clean_migration_mle()?
>>> mle->woken is set to 1 in dlm_clean_migration_mle(), so the code to
>>> detect target node down(if (dlm_is_node_dead(dlm, target))) will never
>>> be run in dlm_migrate_lockres()?
>>>
>>>
>>> 2621 while (1) {
>>> 2622 ret = wait_event_interruptible_timeout(mle->wq,
>>> 2623 (atomic_read(&mle->woken)
>>> == 1),
>>> 2624 msecs_to_jiffies(5000));
>>> 2625
>>> 2626 if (ret >= 0) {
>>> 2627 if (atomic_read(&mle->woken) == 1 ||
>>> 2628 res->owner == target)
>>> 2629 break;
>>> 2630
>>> 2631 mlog(0, "%s:%.*s: timed out during
>>> migration\n",
>>> 2632 dlm->name, res->lockname.len,
>>> res->lockname.name);
>>> 2633 /* avoid hang during shutdown when
>>> migrating lockres
>>> 2634 * to a node which also goes down */
>>> 2635 if (dlm_is_node_dead(dlm, target)) {
>>> 2636 mlog(0, "%s:%.*s: expected migration "
>>> 2637 "target %u is no longer up,
>>> restarting\n",
>>> 2638 dlm->name, res->lockname.len,
>>> 2639 res->lockname.name, target);
>>> 2640 ret = -EINVAL;
>>> 2641 /* migration failed, detach and
>>> clean up mle */
>>> 2642 dlm_mle_detach_hb_events(dlm, mle);
>>> 2643 dlm_put_mle(mle);
>>> 2644 dlm_put_mle_inuse(mle);
>>> 2645 spin_lock(&res->spinlock);
>>> 2646 res->state &= ~DLM_LOCK_RES_MIGRATING;
>>> 2647 wake = 1;
>>> 2648 spin_unlock(&res->spinlock);
>>> 2649 goto leave;
>>> 2650 }
>>> 2651 } else
>>> 2652 mlog(0, "%s:%.*s: caught signal during
>>> migration\n",
>>> 2653 dlm->name, res->lockname.len,
>>> res->lockname.name);
>>> 2654 }
>>>
>>>
>>> Thanks,
>>> Junxiao.
>>> On 12/28/2015 03:44 PM, xuejiufei wrote:
>>>> We have found that migration source will trigger a BUG that the
>>>> refcount of mle is already zero before put when the target is
>>>> down during migration. The situation is as follows:
>>>>
>>>> dlm_migrate_lockres
>>>> dlm_add_migration_mle
>>>> dlm_mark_lockres_migrating
>>>> dlm_get_mle_inuse
>>>> <<<<<< Now the refcount of the mle is 2.
>>>> dlm_send_one_lockres and wait for the target to become the
>>>> new master.
>>>> <<<<<< o2hb detect the target down and clean the migration
>>>> mle. Now the refcount is 1.
>>>>
>>>> dlm_migrate_lockres woken, and put the mle twice when found
>>>> the target goes down which trigger the BUG with the following
>>>> message:
>>>> "ERROR: bad mle: ".
>>>>
>>>> Signed-off-by: Jiufei Xue <xuejiufei@huawei.com>
>>>> Reviewed-by: Joseph Qi <joseph.qi@huawei.com>
>>>> ---
>>>> fs/ocfs2/dlm/dlmmaster.c | 26 +++++++++++++++-----------
>>>> 1 file changed, 15 insertions(+), 11 deletions(-)
>>>>
>>>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>>>> index 936e11b..b713140 100644
>>>> --- a/fs/ocfs2/dlm/dlmmaster.c
>>>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>>>> @@ -2519,6 +2519,11 @@ static int dlm_migrate_lockres(struct dlm_ctxt *dlm,
>>>> spin_lock(&dlm->master_lock);
>>>> ret = dlm_add_migration_mle(dlm, res, mle, &oldmle, name,
>>>> namelen, target, dlm->node_num);
>>>> + /* get an extra reference on the mle.
>>>> + * otherwise the assert_master from the new
>>>> + * master will destroy this.
>>>> + */
>>>> + dlm_get_mle_inuse(mle);
>>>> spin_unlock(&dlm->master_lock);
>>>> spin_unlock(&dlm->spinlock);
>>>>
>>>> @@ -2554,6 +2559,7 @@ fail:
>>>> if (mle_added) {
>>>> dlm_mle_detach_hb_events(dlm, mle);
>>>> dlm_put_mle(mle);
>>>> + dlm_put_mle_inuse(mle);
>>>> } else if (mle) {
>>>> kmem_cache_free(dlm_mle_cache, mle);
>>>> mle = NULL;
>>>> @@ -2571,17 +2577,6 @@ fail:
>>>> * ensure that all assert_master work is flushed. */
>>>> flush_workqueue(dlm->dlm_worker);
>>>>
>>>> - /* get an extra reference on the mle.
>>>> - * otherwise the assert_master from the new
>>>> - * master will destroy this.
>>>> - * also, make sure that all callers of dlm_get_mle
>>>> - * take both dlm->spinlock and dlm->master_lock */
>>>> - spin_lock(&dlm->spinlock);
>>>> - spin_lock(&dlm->master_lock);
>>>> - dlm_get_mle_inuse(mle);
>>>> - spin_unlock(&dlm->master_lock);
>>>> - spin_unlock(&dlm->spinlock);
>>>> -
>>>> /* notify new node and send all lock state */
>>>> /* call send_one_lockres with migration flag.
>>>> * this serves as notice to the target node that a
>>>> @@ -3312,6 +3307,15 @@ top:
>>>> mle->new_master != dead_node)
>>>> continue;
>>>>
>>>> + if (mle->new_master == dead_node && mle->inuse) {
>>>> + mlog(ML_NOTICE, "%s: target %u died during "
>>>> + "migration from %u, the MLE is "
>>>> + "still keep used, ignore it!\n",
>>>> + dlm->name, dead_node,
>>>> + mle->master);
>>>> + continue;
>>>> + }
>>>> +
>>>> /* If we have reached this point, this mle needs to be
>>>> * removed from the list and freed. */
>>>> dlm_clean_migration_mle(dlm, mle);
>>>>
>>>
>>>
>>> .
>>>
>>
>
>
> .
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse
2015-12-31 7:15 ` xuejiufei
@ 2016-01-04 8:20 ` Junxiao Bi
2016-01-06 0:57 ` xuejiufei
0 siblings, 1 reply; 8+ messages in thread
From: Junxiao Bi @ 2016-01-04 8:20 UTC (permalink / raw)
To: ocfs2-devel
Hi Jiufei,
On 12/31/2015 03:15 PM, xuejiufei wrote:
> Hi Junxiao,
> On 2015/12/31 11:05, Junxiao Bi wrote:
>> On 12/30/2015 05:56 PM, xuejiufei wrote:
>>> Hi Junxiao,
>>> You are right. But it may happen that mle->woken is set to 1 in
>>> dlm_clean_migration_mle() just after atomic_read() in
>>> dlm_migrate_lockres(). Actually we trigger this BUG when dlm_send_one_lockres()
>>> return error.
>> Yes, that's possible because of that 5s timeout wakeup, I think this
>> timeout is useless and can be removed?
>>>
>>> And I think dlm_migrate_lockres() should not set owner to target and return 0
>>> when mle->woken is set to 1 in dlm_clean_migration_mle(). This is another problem?
>> Yes, it should be fixed, or the lockres owner will be set to a down node
>> wrongly. Can the following fix these two issue?
>>
> It can not fix the first issue when dlm_send_one_lockres() or
> dlm_mark_lockres_migrating() return error, right?
Right. This is for second issue. Any way please consider these two
issue, I think they can be fixed in one patch since both happened when
target node down during migrate. You can make your own one or merge this
patch to yours if you like.
Thanks,
Junxiao.
>
>>
>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>> index 84f2f8079466..d0380ea62340 100644
>> --- a/fs/ocfs2/dlm/dlmmaster.c
>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>> @@ -2618,39 +2618,22 @@ fail:
>>
>>
>> /* wait for new node to assert master */
>> - while (1) {
>> - ret = wait_event_interruptible_timeout(mle->wq,
>> - (atomic_read(&mle->woken) == 1),
>> - msecs_to_jiffies(5000));
>> -
>> - if (ret >= 0) {
>> - if (atomic_read(&mle->woken) == 1 ||
>> - res->owner == target)
>> - break;
>> -
>> - mlog(0, "%s:%.*s: timed out during migration\n",
>> - dlm->name, res->lockname.len, res->lockname.name);
>> - /* avoid hang during shutdown when migrating lockres
>> - * to a node which also goes down */
>> - if (dlm_is_node_dead(dlm, target)) {
>> - mlog(0, "%s:%.*s: expected migration "
>> - "target %u is no longer up, restarting\n",
>> - dlm->name, res->lockname.len,
>> - res->lockname.name, target);
>> - ret = -EINVAL;
>> - /* migration failed, detach and clean up mle */
>> - dlm_mle_detach_hb_events(dlm, mle);
>> - dlm_put_mle(mle);
>> - dlm_put_mle_inuse(mle);
>> - spin_lock(&res->spinlock);
>> - res->state &= ~DLM_LOCK_RES_MIGRATING;
>> - wake = 1;
>> - spin_unlock(&res->spinlock);
>> - goto leave;
>> - }
>> - } else
>> - mlog(0, "%s:%.*s: caught signal during migration\n",
>> - dlm->name, res->lockname.len, res->lockname.name);
>> + wait_event(mle->wq, ((atomic_read(&mle->woken) == 1) ||
>> + (atomic_read(&mle->woken) == 2)));
>> +
>> + /* migrate target down */
>> + if (atomic_read(&mle->woken) == 2) {
>> + mlog(0, "%s:%.*s: expected migration "
>> + "target %u is no longer up, restarting\n",
>> + dlm->name, res->lockname.len,
>> + res->lockname.name, target);
>> + ret = -EINVAL;
>> + dlm_put_mle_inuse(mle);
>> + spin_lock(&res->spinlock);
>> + res->state &= ~DLM_LOCK_RES_MIGRATING;
>> + wake = 1;
>> + spin_unlock(&res->spinlock);
>> + goto leave;
>> }
>>
>> /* all done, set the owner, clear the flag */
>> @@ -3227,7 +3210,7 @@ static void dlm_clean_migration_mle(struct
>> dlm_ctxt *dlm,
>>
>> spin_lock(&mle->spinlock);
>> __dlm_unlink_mle(dlm, mle);
>> - atomic_set(&mle->woken, 1);
>> + atomic_set(&mle->woken, 2);
>> spin_unlock(&mle->spinlock);
>>
>> wake_up(&mle->wq);
>>
>> Thanks,
>> Junxiao.
>>
>>
>>>
>>> Thanks
>>> Jiufei.
>>>
>>> On 2015/12/30 10:52, Junxiao Bi wrote:
>>>> Hi Jiufei,
>>>>
>>>> When target node down, mle is cleared from
>>>> dlm_do_local_recovery_cleanup()->dlm_clean_master_list()->dlm_clean_migration_mle()?
>>>> mle->woken is set to 1 in dlm_clean_migration_mle(), so the code to
>>>> detect target node down(if (dlm_is_node_dead(dlm, target))) will never
>>>> be run in dlm_migrate_lockres()?
>>>>
>>>>
>>>> 2621 while (1) {
>>>> 2622 ret = wait_event_interruptible_timeout(mle->wq,
>>>> 2623 (atomic_read(&mle->woken)
>>>> == 1),
>>>> 2624 msecs_to_jiffies(5000));
>>>> 2625
>>>> 2626 if (ret >= 0) {
>>>> 2627 if (atomic_read(&mle->woken) == 1 ||
>>>> 2628 res->owner == target)
>>>> 2629 break;
>>>> 2630
>>>> 2631 mlog(0, "%s:%.*s: timed out during
>>>> migration\n",
>>>> 2632 dlm->name, res->lockname.len,
>>>> res->lockname.name);
>>>> 2633 /* avoid hang during shutdown when
>>>> migrating lockres
>>>> 2634 * to a node which also goes down */
>>>> 2635 if (dlm_is_node_dead(dlm, target)) {
>>>> 2636 mlog(0, "%s:%.*s: expected migration "
>>>> 2637 "target %u is no longer up,
>>>> restarting\n",
>>>> 2638 dlm->name, res->lockname.len,
>>>> 2639 res->lockname.name, target);
>>>> 2640 ret = -EINVAL;
>>>> 2641 /* migration failed, detach and
>>>> clean up mle */
>>>> 2642 dlm_mle_detach_hb_events(dlm, mle);
>>>> 2643 dlm_put_mle(mle);
>>>> 2644 dlm_put_mle_inuse(mle);
>>>> 2645 spin_lock(&res->spinlock);
>>>> 2646 res->state &= ~DLM_LOCK_RES_MIGRATING;
>>>> 2647 wake = 1;
>>>> 2648 spin_unlock(&res->spinlock);
>>>> 2649 goto leave;
>>>> 2650 }
>>>> 2651 } else
>>>> 2652 mlog(0, "%s:%.*s: caught signal during
>>>> migration\n",
>>>> 2653 dlm->name, res->lockname.len,
>>>> res->lockname.name);
>>>> 2654 }
>>>>
>>>>
>>>> Thanks,
>>>> Junxiao.
>>>> On 12/28/2015 03:44 PM, xuejiufei wrote:
>>>>> We have found that migration source will trigger a BUG that the
>>>>> refcount of mle is already zero before put when the target is
>>>>> down during migration. The situation is as follows:
>>>>>
>>>>> dlm_migrate_lockres
>>>>> dlm_add_migration_mle
>>>>> dlm_mark_lockres_migrating
>>>>> dlm_get_mle_inuse
>>>>> <<<<<< Now the refcount of the mle is 2.
>>>>> dlm_send_one_lockres and wait for the target to become the
>>>>> new master.
>>>>> <<<<<< o2hb detect the target down and clean the migration
>>>>> mle. Now the refcount is 1.
>>>>>
>>>>> dlm_migrate_lockres woken, and put the mle twice when found
>>>>> the target goes down which trigger the BUG with the following
>>>>> message:
>>>>> "ERROR: bad mle: ".
>>>>>
>>>>> Signed-off-by: Jiufei Xue <xuejiufei@huawei.com>
>>>>> Reviewed-by: Joseph Qi <joseph.qi@huawei.com>
>>>>> ---
>>>>> fs/ocfs2/dlm/dlmmaster.c | 26 +++++++++++++++-----------
>>>>> 1 file changed, 15 insertions(+), 11 deletions(-)
>>>>>
>>>>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>>>>> index 936e11b..b713140 100644
>>>>> --- a/fs/ocfs2/dlm/dlmmaster.c
>>>>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>>>>> @@ -2519,6 +2519,11 @@ static int dlm_migrate_lockres(struct dlm_ctxt *dlm,
>>>>> spin_lock(&dlm->master_lock);
>>>>> ret = dlm_add_migration_mle(dlm, res, mle, &oldmle, name,
>>>>> namelen, target, dlm->node_num);
>>>>> + /* get an extra reference on the mle.
>>>>> + * otherwise the assert_master from the new
>>>>> + * master will destroy this.
>>>>> + */
>>>>> + dlm_get_mle_inuse(mle);
>>>>> spin_unlock(&dlm->master_lock);
>>>>> spin_unlock(&dlm->spinlock);
>>>>>
>>>>> @@ -2554,6 +2559,7 @@ fail:
>>>>> if (mle_added) {
>>>>> dlm_mle_detach_hb_events(dlm, mle);
>>>>> dlm_put_mle(mle);
>>>>> + dlm_put_mle_inuse(mle);
>>>>> } else if (mle) {
>>>>> kmem_cache_free(dlm_mle_cache, mle);
>>>>> mle = NULL;
>>>>> @@ -2571,17 +2577,6 @@ fail:
>>>>> * ensure that all assert_master work is flushed. */
>>>>> flush_workqueue(dlm->dlm_worker);
>>>>>
>>>>> - /* get an extra reference on the mle.
>>>>> - * otherwise the assert_master from the new
>>>>> - * master will destroy this.
>>>>> - * also, make sure that all callers of dlm_get_mle
>>>>> - * take both dlm->spinlock and dlm->master_lock */
>>>>> - spin_lock(&dlm->spinlock);
>>>>> - spin_lock(&dlm->master_lock);
>>>>> - dlm_get_mle_inuse(mle);
>>>>> - spin_unlock(&dlm->master_lock);
>>>>> - spin_unlock(&dlm->spinlock);
>>>>> -
>>>>> /* notify new node and send all lock state */
>>>>> /* call send_one_lockres with migration flag.
>>>>> * this serves as notice to the target node that a
>>>>> @@ -3312,6 +3307,15 @@ top:
>>>>> mle->new_master != dead_node)
>>>>> continue;
>>>>>
>>>>> + if (mle->new_master == dead_node && mle->inuse) {
>>>>> + mlog(ML_NOTICE, "%s: target %u died during "
>>>>> + "migration from %u, the MLE is "
>>>>> + "still keep used, ignore it!\n",
>>>>> + dlm->name, dead_node,
>>>>> + mle->master);
>>>>> + continue;
>>>>> + }
>>>>> +
>>>>> /* If we have reached this point, this mle needs to be
>>>>> * removed from the list and freed. */
>>>>> dlm_clean_migration_mle(dlm, mle);
>>>>>
>>>>
>>>>
>>>> .
>>>>
>>>
>>
>>
>> .
>>
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse
2016-01-04 8:20 ` Junxiao Bi
@ 2016-01-06 0:57 ` xuejiufei
2016-01-07 1:46 ` Junxiao Bi
0 siblings, 1 reply; 8+ messages in thread
From: xuejiufei @ 2016-01-06 0:57 UTC (permalink / raw)
To: ocfs2-devel
Hi Junxiao,
Function dlm_clean_migration_mle() will not be called after the first
patch applied. So the issue that the owner of lockres will be set to
a down node is not exist.
Thanks,
Jiufei
On 2016/1/4 16:20, Junxiao Bi wrote:
> Hi Jiufei,
>
> On 12/31/2015 03:15 PM, xuejiufei wrote:
>> Hi Junxiao,
>> On 2015/12/31 11:05, Junxiao Bi wrote:
>>> On 12/30/2015 05:56 PM, xuejiufei wrote:
>>>> Hi Junxiao,
>>>> You are right. But it may happen that mle->woken is set to 1 in
>>>> dlm_clean_migration_mle() just after atomic_read() in
>>>> dlm_migrate_lockres(). Actually we trigger this BUG when dlm_send_one_lockres()
>>>> return error.
>>> Yes, that's possible because of that 5s timeout wakeup, I think this
>>> timeout is useless and can be removed?
>>>>
>>>> And I think dlm_migrate_lockres() should not set owner to target and return 0
>>>> when mle->woken is set to 1 in dlm_clean_migration_mle(). This is another problem?
>>> Yes, it should be fixed, or the lockres owner will be set to a down node
>>> wrongly. Can the following fix these two issue?
>>>
>> It can not fix the first issue when dlm_send_one_lockres() or
>> dlm_mark_lockres_migrating() return error, right?
> Right. This is for second issue. Any way please consider these two
> issue, I think they can be fixed in one patch since both happened when
> target node down during migrate. You can make your own one or merge this
> patch to yours if you like.
>
> Thanks,
> Junxiao.
>>
>>>
>>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>>> index 84f2f8079466..d0380ea62340 100644
>>> --- a/fs/ocfs2/dlm/dlmmaster.c
>>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>>> @@ -2618,39 +2618,22 @@ fail:
>>>
>>>
>>> /* wait for new node to assert master */
>>> - while (1) {
>>> - ret = wait_event_interruptible_timeout(mle->wq,
>>> - (atomic_read(&mle->woken) == 1),
>>> - msecs_to_jiffies(5000));
>>> -
>>> - if (ret >= 0) {
>>> - if (atomic_read(&mle->woken) == 1 ||
>>> - res->owner == target)
>>> - break;
>>> -
>>> - mlog(0, "%s:%.*s: timed out during migration\n",
>>> - dlm->name, res->lockname.len, res->lockname.name);
>>> - /* avoid hang during shutdown when migrating lockres
>>> - * to a node which also goes down */
>>> - if (dlm_is_node_dead(dlm, target)) {
>>> - mlog(0, "%s:%.*s: expected migration "
>>> - "target %u is no longer up, restarting\n",
>>> - dlm->name, res->lockname.len,
>>> - res->lockname.name, target);
>>> - ret = -EINVAL;
>>> - /* migration failed, detach and clean up mle */
>>> - dlm_mle_detach_hb_events(dlm, mle);
>>> - dlm_put_mle(mle);
>>> - dlm_put_mle_inuse(mle);
>>> - spin_lock(&res->spinlock);
>>> - res->state &= ~DLM_LOCK_RES_MIGRATING;
>>> - wake = 1;
>>> - spin_unlock(&res->spinlock);
>>> - goto leave;
>>> - }
>>> - } else
>>> - mlog(0, "%s:%.*s: caught signal during migration\n",
>>> - dlm->name, res->lockname.len, res->lockname.name);
>>> + wait_event(mle->wq, ((atomic_read(&mle->woken) == 1) ||
>>> + (atomic_read(&mle->woken) == 2)));
>>> +
>>> + /* migrate target down */
>>> + if (atomic_read(&mle->woken) == 2) {
>>> + mlog(0, "%s:%.*s: expected migration "
>>> + "target %u is no longer up, restarting\n",
>>> + dlm->name, res->lockname.len,
>>> + res->lockname.name, target);
>>> + ret = -EINVAL;
>>> + dlm_put_mle_inuse(mle);
>>> + spin_lock(&res->spinlock);
>>> + res->state &= ~DLM_LOCK_RES_MIGRATING;
>>> + wake = 1;
>>> + spin_unlock(&res->spinlock);
>>> + goto leave;
>>> }
>>>
>>> /* all done, set the owner, clear the flag */
>>> @@ -3227,7 +3210,7 @@ static void dlm_clean_migration_mle(struct
>>> dlm_ctxt *dlm,
>>>
>>> spin_lock(&mle->spinlock);
>>> __dlm_unlink_mle(dlm, mle);
>>> - atomic_set(&mle->woken, 1);
>>> + atomic_set(&mle->woken, 2);
>>> spin_unlock(&mle->spinlock);
>>>
>>> wake_up(&mle->wq);
>>>
>>> Thanks,
>>> Junxiao.
>>>
>>>
>>>>
>>>> Thanks
>>>> Jiufei.
>>>>
>>>> On 2015/12/30 10:52, Junxiao Bi wrote:
>>>>> Hi Jiufei,
>>>>>
>>>>> When target node down, mle is cleared from
>>>>> dlm_do_local_recovery_cleanup()->dlm_clean_master_list()->dlm_clean_migration_mle()?
>>>>> mle->woken is set to 1 in dlm_clean_migration_mle(), so the code to
>>>>> detect target node down(if (dlm_is_node_dead(dlm, target))) will never
>>>>> be run in dlm_migrate_lockres()?
>>>>>
>>>>>
>>>>> 2621 while (1) {
>>>>> 2622 ret = wait_event_interruptible_timeout(mle->wq,
>>>>> 2623 (atomic_read(&mle->woken)
>>>>> == 1),
>>>>> 2624 msecs_to_jiffies(5000));
>>>>> 2625
>>>>> 2626 if (ret >= 0) {
>>>>> 2627 if (atomic_read(&mle->woken) == 1 ||
>>>>> 2628 res->owner == target)
>>>>> 2629 break;
>>>>> 2630
>>>>> 2631 mlog(0, "%s:%.*s: timed out during
>>>>> migration\n",
>>>>> 2632 dlm->name, res->lockname.len,
>>>>> res->lockname.name);
>>>>> 2633 /* avoid hang during shutdown when
>>>>> migrating lockres
>>>>> 2634 * to a node which also goes down */
>>>>> 2635 if (dlm_is_node_dead(dlm, target)) {
>>>>> 2636 mlog(0, "%s:%.*s: expected migration "
>>>>> 2637 "target %u is no longer up,
>>>>> restarting\n",
>>>>> 2638 dlm->name, res->lockname.len,
>>>>> 2639 res->lockname.name, target);
>>>>> 2640 ret = -EINVAL;
>>>>> 2641 /* migration failed, detach and
>>>>> clean up mle */
>>>>> 2642 dlm_mle_detach_hb_events(dlm, mle);
>>>>> 2643 dlm_put_mle(mle);
>>>>> 2644 dlm_put_mle_inuse(mle);
>>>>> 2645 spin_lock(&res->spinlock);
>>>>> 2646 res->state &= ~DLM_LOCK_RES_MIGRATING;
>>>>> 2647 wake = 1;
>>>>> 2648 spin_unlock(&res->spinlock);
>>>>> 2649 goto leave;
>>>>> 2650 }
>>>>> 2651 } else
>>>>> 2652 mlog(0, "%s:%.*s: caught signal during
>>>>> migration\n",
>>>>> 2653 dlm->name, res->lockname.len,
>>>>> res->lockname.name);
>>>>> 2654 }
>>>>>
>>>>>
>>>>> Thanks,
>>>>> Junxiao.
>>>>> On 12/28/2015 03:44 PM, xuejiufei wrote:
>>>>>> We have found that migration source will trigger a BUG that the
>>>>>> refcount of mle is already zero before put when the target is
>>>>>> down during migration. The situation is as follows:
>>>>>>
>>>>>> dlm_migrate_lockres
>>>>>> dlm_add_migration_mle
>>>>>> dlm_mark_lockres_migrating
>>>>>> dlm_get_mle_inuse
>>>>>> <<<<<< Now the refcount of the mle is 2.
>>>>>> dlm_send_one_lockres and wait for the target to become the
>>>>>> new master.
>>>>>> <<<<<< o2hb detect the target down and clean the migration
>>>>>> mle. Now the refcount is 1.
>>>>>>
>>>>>> dlm_migrate_lockres woken, and put the mle twice when found
>>>>>> the target goes down which trigger the BUG with the following
>>>>>> message:
>>>>>> "ERROR: bad mle: ".
>>>>>>
>>>>>> Signed-off-by: Jiufei Xue <xuejiufei@huawei.com>
>>>>>> Reviewed-by: Joseph Qi <joseph.qi@huawei.com>
>>>>>> ---
>>>>>> fs/ocfs2/dlm/dlmmaster.c | 26 +++++++++++++++-----------
>>>>>> 1 file changed, 15 insertions(+), 11 deletions(-)
>>>>>>
>>>>>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>>>>>> index 936e11b..b713140 100644
>>>>>> --- a/fs/ocfs2/dlm/dlmmaster.c
>>>>>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>>>>>> @@ -2519,6 +2519,11 @@ static int dlm_migrate_lockres(struct dlm_ctxt *dlm,
>>>>>> spin_lock(&dlm->master_lock);
>>>>>> ret = dlm_add_migration_mle(dlm, res, mle, &oldmle, name,
>>>>>> namelen, target, dlm->node_num);
>>>>>> + /* get an extra reference on the mle.
>>>>>> + * otherwise the assert_master from the new
>>>>>> + * master will destroy this.
>>>>>> + */
>>>>>> + dlm_get_mle_inuse(mle);
>>>>>> spin_unlock(&dlm->master_lock);
>>>>>> spin_unlock(&dlm->spinlock);
>>>>>>
>>>>>> @@ -2554,6 +2559,7 @@ fail:
>>>>>> if (mle_added) {
>>>>>> dlm_mle_detach_hb_events(dlm, mle);
>>>>>> dlm_put_mle(mle);
>>>>>> + dlm_put_mle_inuse(mle);
>>>>>> } else if (mle) {
>>>>>> kmem_cache_free(dlm_mle_cache, mle);
>>>>>> mle = NULL;
>>>>>> @@ -2571,17 +2577,6 @@ fail:
>>>>>> * ensure that all assert_master work is flushed. */
>>>>>> flush_workqueue(dlm->dlm_worker);
>>>>>>
>>>>>> - /* get an extra reference on the mle.
>>>>>> - * otherwise the assert_master from the new
>>>>>> - * master will destroy this.
>>>>>> - * also, make sure that all callers of dlm_get_mle
>>>>>> - * take both dlm->spinlock and dlm->master_lock */
>>>>>> - spin_lock(&dlm->spinlock);
>>>>>> - spin_lock(&dlm->master_lock);
>>>>>> - dlm_get_mle_inuse(mle);
>>>>>> - spin_unlock(&dlm->master_lock);
>>>>>> - spin_unlock(&dlm->spinlock);
>>>>>> -
>>>>>> /* notify new node and send all lock state */
>>>>>> /* call send_one_lockres with migration flag.
>>>>>> * this serves as notice to the target node that a
>>>>>> @@ -3312,6 +3307,15 @@ top:
>>>>>> mle->new_master != dead_node)
>>>>>> continue;
>>>>>>
>>>>>> + if (mle->new_master == dead_node && mle->inuse) {
>>>>>> + mlog(ML_NOTICE, "%s: target %u died during "
>>>>>> + "migration from %u, the MLE is "
>>>>>> + "still keep used, ignore it!\n",
>>>>>> + dlm->name, dead_node,
>>>>>> + mle->master);
>>>>>> + continue;
>>>>>> + }
>>>>>> +
>>>>>> /* If we have reached this point, this mle needs to be
>>>>>> * removed from the list and freed. */
>>>>>> dlm_clean_migration_mle(dlm, mle);
>>>>>>
>>>>>
>>>>>
>>>>> .
>>>>>
>>>>
>>>
>>>
>>> .
>>>
>>
>
>
> .
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse
2016-01-06 0:57 ` xuejiufei
@ 2016-01-07 1:46 ` Junxiao Bi
0 siblings, 0 replies; 8+ messages in thread
From: Junxiao Bi @ 2016-01-07 1:46 UTC (permalink / raw)
To: ocfs2-devel
On 01/06/2016 08:57 AM, xuejiufei wrote:
> Hi Junxiao,
> Function dlm_clean_migration_mle() will not be called after the first
> patch applied. So the issue that the owner of lockres will be set to
> a down node is not exist.
Right. Thank you for the explanation.
Thanks,
Junxiao.
>
> Thanks,
> Jiufei
>
> On 2016/1/4 16:20, Junxiao Bi wrote:
>> Hi Jiufei,
>>
>> On 12/31/2015 03:15 PM, xuejiufei wrote:
>>> Hi Junxiao,
>>> On 2015/12/31 11:05, Junxiao Bi wrote:
>>>> On 12/30/2015 05:56 PM, xuejiufei wrote:
>>>>> Hi Junxiao,
>>>>> You are right. But it may happen that mle->woken is set to 1 in
>>>>> dlm_clean_migration_mle() just after atomic_read() in
>>>>> dlm_migrate_lockres(). Actually we trigger this BUG when dlm_send_one_lockres()
>>>>> return error.
>>>> Yes, that's possible because of that 5s timeout wakeup, I think this
>>>> timeout is useless and can be removed?
>>>>>
>>>>> And I think dlm_migrate_lockres() should not set owner to target and return 0
>>>>> when mle->woken is set to 1 in dlm_clean_migration_mle(). This is another problem?
>>>> Yes, it should be fixed, or the lockres owner will be set to a down node
>>>> wrongly. Can the following fix these two issue?
>>>>
>>> It can not fix the first issue when dlm_send_one_lockres() or
>>> dlm_mark_lockres_migrating() return error, right?
>> Right. This is for second issue. Any way please consider these two
>> issue, I think they can be fixed in one patch since both happened when
>> target node down during migrate. You can make your own one or merge this
>> patch to yours if you like.
>>
>> Thanks,
>> Junxiao.
>>>
>>>>
>>>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>>>> index 84f2f8079466..d0380ea62340 100644
>>>> --- a/fs/ocfs2/dlm/dlmmaster.c
>>>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>>>> @@ -2618,39 +2618,22 @@ fail:
>>>>
>>>>
>>>> /* wait for new node to assert master */
>>>> - while (1) {
>>>> - ret = wait_event_interruptible_timeout(mle->wq,
>>>> - (atomic_read(&mle->woken) == 1),
>>>> - msecs_to_jiffies(5000));
>>>> -
>>>> - if (ret >= 0) {
>>>> - if (atomic_read(&mle->woken) == 1 ||
>>>> - res->owner == target)
>>>> - break;
>>>> -
>>>> - mlog(0, "%s:%.*s: timed out during migration\n",
>>>> - dlm->name, res->lockname.len, res->lockname.name);
>>>> - /* avoid hang during shutdown when migrating lockres
>>>> - * to a node which also goes down */
>>>> - if (dlm_is_node_dead(dlm, target)) {
>>>> - mlog(0, "%s:%.*s: expected migration "
>>>> - "target %u is no longer up, restarting\n",
>>>> - dlm->name, res->lockname.len,
>>>> - res->lockname.name, target);
>>>> - ret = -EINVAL;
>>>> - /* migration failed, detach and clean up mle */
>>>> - dlm_mle_detach_hb_events(dlm, mle);
>>>> - dlm_put_mle(mle);
>>>> - dlm_put_mle_inuse(mle);
>>>> - spin_lock(&res->spinlock);
>>>> - res->state &= ~DLM_LOCK_RES_MIGRATING;
>>>> - wake = 1;
>>>> - spin_unlock(&res->spinlock);
>>>> - goto leave;
>>>> - }
>>>> - } else
>>>> - mlog(0, "%s:%.*s: caught signal during migration\n",
>>>> - dlm->name, res->lockname.len, res->lockname.name);
>>>> + wait_event(mle->wq, ((atomic_read(&mle->woken) == 1) ||
>>>> + (atomic_read(&mle->woken) == 2)));
>>>> +
>>>> + /* migrate target down */
>>>> + if (atomic_read(&mle->woken) == 2) {
>>>> + mlog(0, "%s:%.*s: expected migration "
>>>> + "target %u is no longer up, restarting\n",
>>>> + dlm->name, res->lockname.len,
>>>> + res->lockname.name, target);
>>>> + ret = -EINVAL;
>>>> + dlm_put_mle_inuse(mle);
>>>> + spin_lock(&res->spinlock);
>>>> + res->state &= ~DLM_LOCK_RES_MIGRATING;
>>>> + wake = 1;
>>>> + spin_unlock(&res->spinlock);
>>>> + goto leave;
>>>> }
>>>>
>>>> /* all done, set the owner, clear the flag */
>>>> @@ -3227,7 +3210,7 @@ static void dlm_clean_migration_mle(struct
>>>> dlm_ctxt *dlm,
>>>>
>>>> spin_lock(&mle->spinlock);
>>>> __dlm_unlink_mle(dlm, mle);
>>>> - atomic_set(&mle->woken, 1);
>>>> + atomic_set(&mle->woken, 2);
>>>> spin_unlock(&mle->spinlock);
>>>>
>>>> wake_up(&mle->wq);
>>>>
>>>> Thanks,
>>>> Junxiao.
>>>>
>>>>
>>>>>
>>>>> Thanks
>>>>> Jiufei.
>>>>>
>>>>> On 2015/12/30 10:52, Junxiao Bi wrote:
>>>>>> Hi Jiufei,
>>>>>>
>>>>>> When target node down, mle is cleared from
>>>>>> dlm_do_local_recovery_cleanup()->dlm_clean_master_list()->dlm_clean_migration_mle()?
>>>>>> mle->woken is set to 1 in dlm_clean_migration_mle(), so the code to
>>>>>> detect target node down(if (dlm_is_node_dead(dlm, target))) will never
>>>>>> be run in dlm_migrate_lockres()?
>>>>>>
>>>>>>
>>>>>> 2621 while (1) {
>>>>>> 2622 ret = wait_event_interruptible_timeout(mle->wq,
>>>>>> 2623 (atomic_read(&mle->woken)
>>>>>> == 1),
>>>>>> 2624 msecs_to_jiffies(5000));
>>>>>> 2625
>>>>>> 2626 if (ret >= 0) {
>>>>>> 2627 if (atomic_read(&mle->woken) == 1 ||
>>>>>> 2628 res->owner == target)
>>>>>> 2629 break;
>>>>>> 2630
>>>>>> 2631 mlog(0, "%s:%.*s: timed out during
>>>>>> migration\n",
>>>>>> 2632 dlm->name, res->lockname.len,
>>>>>> res->lockname.name);
>>>>>> 2633 /* avoid hang during shutdown when
>>>>>> migrating lockres
>>>>>> 2634 * to a node which also goes down */
>>>>>> 2635 if (dlm_is_node_dead(dlm, target)) {
>>>>>> 2636 mlog(0, "%s:%.*s: expected migration "
>>>>>> 2637 "target %u is no longer up,
>>>>>> restarting\n",
>>>>>> 2638 dlm->name, res->lockname.len,
>>>>>> 2639 res->lockname.name, target);
>>>>>> 2640 ret = -EINVAL;
>>>>>> 2641 /* migration failed, detach and
>>>>>> clean up mle */
>>>>>> 2642 dlm_mle_detach_hb_events(dlm, mle);
>>>>>> 2643 dlm_put_mle(mle);
>>>>>> 2644 dlm_put_mle_inuse(mle);
>>>>>> 2645 spin_lock(&res->spinlock);
>>>>>> 2646 res->state &= ~DLM_LOCK_RES_MIGRATING;
>>>>>> 2647 wake = 1;
>>>>>> 2648 spin_unlock(&res->spinlock);
>>>>>> 2649 goto leave;
>>>>>> 2650 }
>>>>>> 2651 } else
>>>>>> 2652 mlog(0, "%s:%.*s: caught signal during
>>>>>> migration\n",
>>>>>> 2653 dlm->name, res->lockname.len,
>>>>>> res->lockname.name);
>>>>>> 2654 }
>>>>>>
>>>>>>
>>>>>> Thanks,
>>>>>> Junxiao.
>>>>>> On 12/28/2015 03:44 PM, xuejiufei wrote:
>>>>>>> We have found that migration source will trigger a BUG that the
>>>>>>> refcount of mle is already zero before put when the target is
>>>>>>> down during migration. The situation is as follows:
>>>>>>>
>>>>>>> dlm_migrate_lockres
>>>>>>> dlm_add_migration_mle
>>>>>>> dlm_mark_lockres_migrating
>>>>>>> dlm_get_mle_inuse
>>>>>>> <<<<<< Now the refcount of the mle is 2.
>>>>>>> dlm_send_one_lockres and wait for the target to become the
>>>>>>> new master.
>>>>>>> <<<<<< o2hb detect the target down and clean the migration
>>>>>>> mle. Now the refcount is 1.
>>>>>>>
>>>>>>> dlm_migrate_lockres woken, and put the mle twice when found
>>>>>>> the target goes down which trigger the BUG with the following
>>>>>>> message:
>>>>>>> "ERROR: bad mle: ".
>>>>>>>
>>>>>>> Signed-off-by: Jiufei Xue <xuejiufei@huawei.com>
>>>>>>> Reviewed-by: Joseph Qi <joseph.qi@huawei.com>
>>>>>>> ---
>>>>>>> fs/ocfs2/dlm/dlmmaster.c | 26 +++++++++++++++-----------
>>>>>>> 1 file changed, 15 insertions(+), 11 deletions(-)
>>>>>>>
>>>>>>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>>>>>>> index 936e11b..b713140 100644
>>>>>>> --- a/fs/ocfs2/dlm/dlmmaster.c
>>>>>>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>>>>>>> @@ -2519,6 +2519,11 @@ static int dlm_migrate_lockres(struct dlm_ctxt *dlm,
>>>>>>> spin_lock(&dlm->master_lock);
>>>>>>> ret = dlm_add_migration_mle(dlm, res, mle, &oldmle, name,
>>>>>>> namelen, target, dlm->node_num);
>>>>>>> + /* get an extra reference on the mle.
>>>>>>> + * otherwise the assert_master from the new
>>>>>>> + * master will destroy this.
>>>>>>> + */
>>>>>>> + dlm_get_mle_inuse(mle);
>>>>>>> spin_unlock(&dlm->master_lock);
>>>>>>> spin_unlock(&dlm->spinlock);
>>>>>>>
>>>>>>> @@ -2554,6 +2559,7 @@ fail:
>>>>>>> if (mle_added) {
>>>>>>> dlm_mle_detach_hb_events(dlm, mle);
>>>>>>> dlm_put_mle(mle);
>>>>>>> + dlm_put_mle_inuse(mle);
>>>>>>> } else if (mle) {
>>>>>>> kmem_cache_free(dlm_mle_cache, mle);
>>>>>>> mle = NULL;
>>>>>>> @@ -2571,17 +2577,6 @@ fail:
>>>>>>> * ensure that all assert_master work is flushed. */
>>>>>>> flush_workqueue(dlm->dlm_worker);
>>>>>>>
>>>>>>> - /* get an extra reference on the mle.
>>>>>>> - * otherwise the assert_master from the new
>>>>>>> - * master will destroy this.
>>>>>>> - * also, make sure that all callers of dlm_get_mle
>>>>>>> - * take both dlm->spinlock and dlm->master_lock */
>>>>>>> - spin_lock(&dlm->spinlock);
>>>>>>> - spin_lock(&dlm->master_lock);
>>>>>>> - dlm_get_mle_inuse(mle);
>>>>>>> - spin_unlock(&dlm->master_lock);
>>>>>>> - spin_unlock(&dlm->spinlock);
>>>>>>> -
>>>>>>> /* notify new node and send all lock state */
>>>>>>> /* call send_one_lockres with migration flag.
>>>>>>> * this serves as notice to the target node that a
>>>>>>> @@ -3312,6 +3307,15 @@ top:
>>>>>>> mle->new_master != dead_node)
>>>>>>> continue;
>>>>>>>
>>>>>>> + if (mle->new_master == dead_node && mle->inuse) {
>>>>>>> + mlog(ML_NOTICE, "%s: target %u died during "
>>>>>>> + "migration from %u, the MLE is "
>>>>>>> + "still keep used, ignore it!\n",
>>>>>>> + dlm->name, dead_node,
>>>>>>> + mle->master);
>>>>>>> + continue;
>>>>>>> + }
>>>>>>> +
>>>>>>> /* If we have reached this point, this mle needs to be
>>>>>>> * removed from the list and freed. */
>>>>>>> dlm_clean_migration_mle(dlm, mle);
>>>>>>>
>>>>>>
>>>>>>
>>>>>> .
>>>>>>
>>>>>
>>>>
>>>>
>>>> .
>>>>
>>>
>>
>>
>> .
>>
>
^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2016-01-07 1:46 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
[not found] <567E3C07.2000305@huawei.com>
2015-12-28 7:44 ` [Ocfs2-devel] [PATCH] ocfs2/dlm: ignore cleaning the migration mle that is inuse xuejiufei
2015-12-30 2:52 ` Junxiao Bi
2015-12-30 9:56 ` xuejiufei
2015-12-31 3:05 ` Junxiao Bi
2015-12-31 7:15 ` xuejiufei
2016-01-04 8:20 ` Junxiao Bi
2016-01-06 0:57 ` xuejiufei
2016-01-07 1:46 ` Junxiao Bi
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.