Loading fs/ocfs2/dlm/dlmmaster.c +6 −5 Original line number Diff line number Diff line Loading @@ -1639,13 +1639,13 @@ static int dlm_do_assert_master(struct dlm_ctxt *dlm, const char *lockname, tmpret = o2net_send_message(DLM_ASSERT_MASTER_MSG, dlm->key, &assert, sizeof(assert), to, &r); if (tmpret < 0) { mlog(ML_ERROR, "assert_master returned %d!\n", tmpret); mlog(0, "assert_master returned %d!\n", tmpret); if (!dlm_is_host_down(tmpret)) { mlog(ML_ERROR, "unhandled error!\n"); mlog(ML_ERROR, "unhandled error=%d!\n", tmpret); BUG(); } /* a node died. finish out the rest of the nodes. */ mlog(ML_ERROR, "link to %d went down!\n", to); mlog(0, "link to %d went down!\n", to); /* any nonzero status return will do */ ret = tmpret; } else if (r < 0) { Loading Loading @@ -2029,6 +2029,7 @@ static void dlm_assert_master_worker(struct dlm_work_item *item, void *data) nodemap, flags); if (ret < 0) { /* no need to restart, we are done */ if (!dlm_is_host_down(ret)) mlog_errno(ret); } Loading Loading @@ -2808,7 +2809,7 @@ void dlm_clean_master_list(struct dlm_ctxt *dlm, u8 dead_node) * may result in the mle being unlinked and * freed, but there may still be a process * waiting in the dlmlock path which is fine. */ mlog(ML_ERROR, "node %u was expected master\n", mlog(0, "node %u was expected master\n", dead_node); atomic_set(&mle->woken, 1); spin_unlock(&mle->spinlock); Loading fs/ocfs2/dlm/dlmrecovery.c +1 −1 Original line number Diff line number Diff line Loading @@ -410,7 +410,7 @@ static int dlm_in_recovery(struct dlm_ctxt *dlm) void dlm_wait_for_recovery(struct dlm_ctxt *dlm) { if (dlm_in_recovery(dlm)) { mlog(ML_NOTICE, "%s: reco thread %d in recovery: " mlog(0, "%s: reco thread %d in recovery: " "state=%d, master=%u, dead=%u\n", dlm->name, dlm->dlm_reco_thread_task->pid, dlm->reco.state, dlm->reco.new_master, Loading Loading
fs/ocfs2/dlm/dlmmaster.c +6 −5 Original line number Diff line number Diff line Loading @@ -1639,13 +1639,13 @@ static int dlm_do_assert_master(struct dlm_ctxt *dlm, const char *lockname, tmpret = o2net_send_message(DLM_ASSERT_MASTER_MSG, dlm->key, &assert, sizeof(assert), to, &r); if (tmpret < 0) { mlog(ML_ERROR, "assert_master returned %d!\n", tmpret); mlog(0, "assert_master returned %d!\n", tmpret); if (!dlm_is_host_down(tmpret)) { mlog(ML_ERROR, "unhandled error!\n"); mlog(ML_ERROR, "unhandled error=%d!\n", tmpret); BUG(); } /* a node died. finish out the rest of the nodes. */ mlog(ML_ERROR, "link to %d went down!\n", to); mlog(0, "link to %d went down!\n", to); /* any nonzero status return will do */ ret = tmpret; } else if (r < 0) { Loading Loading @@ -2029,6 +2029,7 @@ static void dlm_assert_master_worker(struct dlm_work_item *item, void *data) nodemap, flags); if (ret < 0) { /* no need to restart, we are done */ if (!dlm_is_host_down(ret)) mlog_errno(ret); } Loading Loading @@ -2808,7 +2809,7 @@ void dlm_clean_master_list(struct dlm_ctxt *dlm, u8 dead_node) * may result in the mle being unlinked and * freed, but there may still be a process * waiting in the dlmlock path which is fine. */ mlog(ML_ERROR, "node %u was expected master\n", mlog(0, "node %u was expected master\n", dead_node); atomic_set(&mle->woken, 1); spin_unlock(&mle->spinlock); Loading
fs/ocfs2/dlm/dlmrecovery.c +1 −1 Original line number Diff line number Diff line Loading @@ -410,7 +410,7 @@ static int dlm_in_recovery(struct dlm_ctxt *dlm) void dlm_wait_for_recovery(struct dlm_ctxt *dlm) { if (dlm_in_recovery(dlm)) { mlog(ML_NOTICE, "%s: reco thread %d in recovery: " mlog(0, "%s: reco thread %d in recovery: " "state=%d, master=%u, dead=%u\n", dlm->name, dlm->dlm_reco_thread_task->pid, dlm->reco.state, dlm->reco.new_master, Loading