mirror of
https://github.com/torvalds/linux.git
synced 2024-11-13 23:51:39 +00:00
8304d6f24c
Change how callbacks are recorded for locks. Previously, information about multiple callbacks was combined into a couple of variables that indicated what the end result should be. In some situations, we could not tell from this combined state what the exact sequence of callbacks were, and would end up either delivering the callbacks in the wrong order, or suppress redundant callbacks incorrectly. This new approach records all the data for each callback, leaving no uncertainty about what needs to be delivered. Signed-off-by: David Teigland <teigland@redhat.com>
347 lines
8.3 KiB
C
347 lines
8.3 KiB
C
/******************************************************************************
|
|
*******************************************************************************
|
|
**
|
|
** Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved.
|
|
** Copyright (C) 2004-2010 Red Hat, Inc. All rights reserved.
|
|
**
|
|
** This copyrighted material is made available to anyone wishing to use,
|
|
** modify, copy, or redistribute it subject to the terms and conditions
|
|
** of the GNU General Public License v.2.
|
|
**
|
|
*******************************************************************************
|
|
******************************************************************************/
|
|
|
|
#include "dlm_internal.h"
|
|
#include "lock.h"
|
|
#include "user.h"
|
|
#include "ast.h"
|
|
|
|
#define WAKE_ASTS 0
|
|
|
|
static uint64_t ast_seq_count;
|
|
static struct list_head ast_queue;
|
|
static spinlock_t ast_queue_lock;
|
|
static struct task_struct * astd_task;
|
|
static unsigned long astd_wakeflags;
|
|
static struct mutex astd_running;
|
|
|
|
|
|
static void dlm_dump_lkb_callbacks(struct dlm_lkb *lkb)
|
|
{
|
|
int i;
|
|
|
|
log_print("last_bast %x %llu flags %x mode %d sb %d %x",
|
|
lkb->lkb_id,
|
|
(unsigned long long)lkb->lkb_last_bast.seq,
|
|
lkb->lkb_last_bast.flags,
|
|
lkb->lkb_last_bast.mode,
|
|
lkb->lkb_last_bast.sb_status,
|
|
lkb->lkb_last_bast.sb_flags);
|
|
|
|
log_print("last_cast %x %llu flags %x mode %d sb %d %x",
|
|
lkb->lkb_id,
|
|
(unsigned long long)lkb->lkb_last_cast.seq,
|
|
lkb->lkb_last_cast.flags,
|
|
lkb->lkb_last_cast.mode,
|
|
lkb->lkb_last_cast.sb_status,
|
|
lkb->lkb_last_cast.sb_flags);
|
|
|
|
for (i = 0; i < DLM_CALLBACKS_SIZE; i++) {
|
|
log_print("cb %x %llu flags %x mode %d sb %d %x",
|
|
lkb->lkb_id,
|
|
(unsigned long long)lkb->lkb_callbacks[i].seq,
|
|
lkb->lkb_callbacks[i].flags,
|
|
lkb->lkb_callbacks[i].mode,
|
|
lkb->lkb_callbacks[i].sb_status,
|
|
lkb->lkb_callbacks[i].sb_flags);
|
|
}
|
|
}
|
|
|
|
void dlm_del_ast(struct dlm_lkb *lkb)
|
|
{
|
|
spin_lock(&ast_queue_lock);
|
|
if (!list_empty(&lkb->lkb_astqueue))
|
|
list_del_init(&lkb->lkb_astqueue);
|
|
spin_unlock(&ast_queue_lock);
|
|
}
|
|
|
|
int dlm_add_lkb_callback(struct dlm_lkb *lkb, uint32_t flags, int mode,
|
|
int status, uint32_t sbflags, uint64_t seq)
|
|
{
|
|
struct dlm_ls *ls = lkb->lkb_resource->res_ls;
|
|
uint64_t prev_seq;
|
|
int prev_mode;
|
|
int i;
|
|
|
|
for (i = 0; i < DLM_CALLBACKS_SIZE; i++) {
|
|
if (lkb->lkb_callbacks[i].seq)
|
|
continue;
|
|
|
|
/*
|
|
* Suppress some redundant basts here, do more on removal.
|
|
* Don't even add a bast if the callback just before it
|
|
* is a bast for the same mode or a more restrictive mode.
|
|
* (the addional > PR check is needed for PR/CW inversion)
|
|
*/
|
|
|
|
if ((i > 0) && (flags & DLM_CB_BAST) &&
|
|
(lkb->lkb_callbacks[i-1].flags & DLM_CB_BAST)) {
|
|
|
|
prev_seq = lkb->lkb_callbacks[i-1].seq;
|
|
prev_mode = lkb->lkb_callbacks[i-1].mode;
|
|
|
|
if ((prev_mode == mode) ||
|
|
(prev_mode > mode && prev_mode > DLM_LOCK_PR)) {
|
|
|
|
log_debug(ls, "skip %x add bast %llu mode %d "
|
|
"for bast %llu mode %d",
|
|
lkb->lkb_id,
|
|
(unsigned long long)seq,
|
|
mode,
|
|
(unsigned long long)prev_seq,
|
|
prev_mode);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
lkb->lkb_callbacks[i].seq = seq;
|
|
lkb->lkb_callbacks[i].flags = flags;
|
|
lkb->lkb_callbacks[i].mode = mode;
|
|
lkb->lkb_callbacks[i].sb_status = status;
|
|
lkb->lkb_callbacks[i].sb_flags = (sbflags & 0x000000FF);
|
|
break;
|
|
}
|
|
|
|
if (i == DLM_CALLBACKS_SIZE) {
|
|
log_error(ls, "no callbacks %x %llu flags %x mode %d sb %d %x",
|
|
lkb->lkb_id, (unsigned long long)seq,
|
|
flags, mode, status, sbflags);
|
|
dlm_dump_lkb_callbacks(lkb);
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int dlm_rem_lkb_callback(struct dlm_ls *ls, struct dlm_lkb *lkb,
|
|
struct dlm_callback *cb, int *resid)
|
|
{
|
|
int i;
|
|
|
|
*resid = 0;
|
|
|
|
if (!lkb->lkb_callbacks[0].seq)
|
|
return -ENOENT;
|
|
|
|
/* oldest undelivered cb is callbacks[0] */
|
|
|
|
memcpy(cb, &lkb->lkb_callbacks[0], sizeof(struct dlm_callback));
|
|
memset(&lkb->lkb_callbacks[0], 0, sizeof(struct dlm_callback));
|
|
|
|
/* shift others down */
|
|
|
|
for (i = 1; i < DLM_CALLBACKS_SIZE; i++) {
|
|
if (!lkb->lkb_callbacks[i].seq)
|
|
break;
|
|
memcpy(&lkb->lkb_callbacks[i-1], &lkb->lkb_callbacks[i],
|
|
sizeof(struct dlm_callback));
|
|
memset(&lkb->lkb_callbacks[i], 0, sizeof(struct dlm_callback));
|
|
(*resid)++;
|
|
}
|
|
|
|
/* if cb is a bast, it should be skipped if the blocking mode is
|
|
compatible with the last granted mode */
|
|
|
|
if ((cb->flags & DLM_CB_BAST) && lkb->lkb_last_cast.seq) {
|
|
if (dlm_modes_compat(cb->mode, lkb->lkb_last_cast.mode)) {
|
|
cb->flags |= DLM_CB_SKIP;
|
|
|
|
log_debug(ls, "skip %x bast %llu mode %d "
|
|
"for cast %llu mode %d",
|
|
lkb->lkb_id,
|
|
(unsigned long long)cb->seq,
|
|
cb->mode,
|
|
(unsigned long long)lkb->lkb_last_cast.seq,
|
|
lkb->lkb_last_cast.mode);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
if (cb->flags & DLM_CB_CAST) {
|
|
memcpy(&lkb->lkb_last_cast, cb, sizeof(struct dlm_callback));
|
|
lkb->lkb_last_cast_time = ktime_get();
|
|
}
|
|
|
|
if (cb->flags & DLM_CB_BAST) {
|
|
memcpy(&lkb->lkb_last_bast, cb, sizeof(struct dlm_callback));
|
|
lkb->lkb_last_bast_time = ktime_get();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void dlm_add_ast(struct dlm_lkb *lkb, uint32_t flags, int mode, int status,
|
|
uint32_t sbflags)
|
|
{
|
|
uint64_t seq;
|
|
int rv;
|
|
|
|
spin_lock(&ast_queue_lock);
|
|
|
|
seq = ++ast_seq_count;
|
|
|
|
if (lkb->lkb_flags & DLM_IFL_USER) {
|
|
spin_unlock(&ast_queue_lock);
|
|
dlm_user_add_ast(lkb, flags, mode, status, sbflags, seq);
|
|
return;
|
|
}
|
|
|
|
rv = dlm_add_lkb_callback(lkb, flags, mode, status, sbflags, seq);
|
|
if (rv < 0) {
|
|
spin_unlock(&ast_queue_lock);
|
|
return;
|
|
}
|
|
|
|
if (list_empty(&lkb->lkb_astqueue)) {
|
|
kref_get(&lkb->lkb_ref);
|
|
list_add_tail(&lkb->lkb_astqueue, &ast_queue);
|
|
}
|
|
spin_unlock(&ast_queue_lock);
|
|
|
|
set_bit(WAKE_ASTS, &astd_wakeflags);
|
|
wake_up_process(astd_task);
|
|
}
|
|
|
|
static void process_asts(void)
|
|
{
|
|
struct dlm_ls *ls = NULL;
|
|
struct dlm_rsb *r = NULL;
|
|
struct dlm_lkb *lkb;
|
|
void (*castfn) (void *astparam);
|
|
void (*bastfn) (void *astparam, int mode);
|
|
struct dlm_callback callbacks[DLM_CALLBACKS_SIZE];
|
|
int i, rv, resid;
|
|
|
|
repeat:
|
|
spin_lock(&ast_queue_lock);
|
|
list_for_each_entry(lkb, &ast_queue, lkb_astqueue) {
|
|
r = lkb->lkb_resource;
|
|
ls = r->res_ls;
|
|
|
|
if (dlm_locking_stopped(ls))
|
|
continue;
|
|
|
|
/* we remove from astqueue list and remove everything in
|
|
lkb_callbacks before releasing the spinlock so empty
|
|
lkb_astqueue is always consistent with empty lkb_callbacks */
|
|
|
|
list_del_init(&lkb->lkb_astqueue);
|
|
|
|
castfn = lkb->lkb_astfn;
|
|
bastfn = lkb->lkb_bastfn;
|
|
|
|
memset(&callbacks, 0, sizeof(callbacks));
|
|
|
|
for (i = 0; i < DLM_CALLBACKS_SIZE; i++) {
|
|
rv = dlm_rem_lkb_callback(ls, lkb, &callbacks[i], &resid);
|
|
if (rv < 0)
|
|
break;
|
|
}
|
|
spin_unlock(&ast_queue_lock);
|
|
|
|
if (resid) {
|
|
/* shouldn't happen, for loop should have removed all */
|
|
log_error(ls, "callback resid %d lkb %x",
|
|
resid, lkb->lkb_id);
|
|
}
|
|
|
|
for (i = 0; i < DLM_CALLBACKS_SIZE; i++) {
|
|
if (!callbacks[i].seq)
|
|
break;
|
|
if (callbacks[i].flags & DLM_CB_SKIP) {
|
|
continue;
|
|
} else if (callbacks[i].flags & DLM_CB_BAST) {
|
|
bastfn(lkb->lkb_astparam, callbacks[i].mode);
|
|
} else if (callbacks[i].flags & DLM_CB_CAST) {
|
|
lkb->lkb_lksb->sb_status = callbacks[i].sb_status;
|
|
lkb->lkb_lksb->sb_flags = callbacks[i].sb_flags;
|
|
castfn(lkb->lkb_astparam);
|
|
}
|
|
}
|
|
|
|
/* removes ref for ast_queue, may cause lkb to be freed */
|
|
dlm_put_lkb(lkb);
|
|
|
|
cond_resched();
|
|
goto repeat;
|
|
}
|
|
spin_unlock(&ast_queue_lock);
|
|
}
|
|
|
|
static inline int no_asts(void)
|
|
{
|
|
int ret;
|
|
|
|
spin_lock(&ast_queue_lock);
|
|
ret = list_empty(&ast_queue);
|
|
spin_unlock(&ast_queue_lock);
|
|
return ret;
|
|
}
|
|
|
|
static int dlm_astd(void *data)
|
|
{
|
|
while (!kthread_should_stop()) {
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
if (!test_bit(WAKE_ASTS, &astd_wakeflags))
|
|
schedule();
|
|
set_current_state(TASK_RUNNING);
|
|
|
|
mutex_lock(&astd_running);
|
|
if (test_and_clear_bit(WAKE_ASTS, &astd_wakeflags))
|
|
process_asts();
|
|
mutex_unlock(&astd_running);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
void dlm_astd_wake(void)
|
|
{
|
|
if (!no_asts()) {
|
|
set_bit(WAKE_ASTS, &astd_wakeflags);
|
|
wake_up_process(astd_task);
|
|
}
|
|
}
|
|
|
|
int dlm_astd_start(void)
|
|
{
|
|
struct task_struct *p;
|
|
int error = 0;
|
|
|
|
INIT_LIST_HEAD(&ast_queue);
|
|
spin_lock_init(&ast_queue_lock);
|
|
mutex_init(&astd_running);
|
|
|
|
p = kthread_run(dlm_astd, NULL, "dlm_astd");
|
|
if (IS_ERR(p))
|
|
error = PTR_ERR(p);
|
|
else
|
|
astd_task = p;
|
|
return error;
|
|
}
|
|
|
|
void dlm_astd_stop(void)
|
|
{
|
|
kthread_stop(astd_task);
|
|
}
|
|
|
|
void dlm_astd_suspend(void)
|
|
{
|
|
mutex_lock(&astd_running);
|
|
}
|
|
|
|
void dlm_astd_resume(void)
|
|
{
|
|
mutex_unlock(&astd_running);
|
|
}
|
|
|