forked from luck/tmp_suning_uos_patched
4a7727725d
This patch fixes a couple problems dealing with spectators who remain with gfs2 mounts after the last non-spectator node fails. Before this patch, spectator mounts would try to acquire the dlm's mounted lock EX as part of its normal recovery sequence. The mounted lock is only used to determine whether the node is the first mounter, the first node to mount the file system, for the purposes of file system recovery and journal replay. It's not necessary for spectators: they should never do journal recovery. If they acquire the lock it will prevent another "real" first-mounter from acquiring the lock in EX mode, which means it also cannot do journal recovery because it doesn't think it's the first node to mount the file system. This patch checks if the mounter is a spectator, and if so, avoids grabbing the mounted lock. This allows a secondary mounter who is really the first non-spectator mounter, to do journal recovery: since the spectator doesn't acquire the lock, it can grab it in EX mode, and therefore consider itself to be the first mounter both as a "real" first mount, and as a first-real-after-spectator. Note that the control lock still needs to be taken in PR mode in order to fetch the lvb value so it has the current status of all journal's recovery. This is used as it is today by a first mounter to replay the journals. For spectators, it's merely used to fetch the status bits. All recovery is bypassed and the node waits until recovery is completed by a non-spectator node. I also improved the cryptic message given by control_mount when a spectator is waiting for a non-spectator to perform recovery. It also fixes a problem in gfs2_recover_set whereby spectators were never queueing recovery work for their own journal. They cannot do recovery themselves, but they still need to queue the work so they can check the recovery bits and clear the DFL_BLOCK_LOCKS bit once the recovery happens on another node. When the work queue runs on a spectator, it bypasses most of the work so it won't print a bunch of annoying messages. All it will print is a bunch of messages that look like this until recovery completes on the non-spectator node: GFS2: fsid=mycluster:scratch.s: recover generation 3 jid 0 GFS2: fsid=mycluster:scratch.s: recover jid 0 result busy These continue every 1.5 seconds until the recovery is done by the non-spectator, at which time it says: GFS2: fsid=mycluster:scratch.s: recover generation 4 done Then it proceeds with its mount. If the file system is mounted in spectator node and the last remaining non-spectator is fenced, any IO to the file system is blocked by dlm and the spectator waits until recovery is performed by a non-spectator. If a spectator tries to mount the file system before any non-spectators, it blocks and repeatedly gives this kernel message: GFS2: fsid=mycluster:scratch: Recovery is required. Waiting for a non-spectator to mount. GFS2: fsid=mycluster:scratch: Recovery is required. Waiting for a non-spectator to mount. Signed-off-by: Bob Peterson <rpeterso@redhat.com> Signed-off-by: Andreas Gruenbacher <agruenba@redhat.com>
738 lines
18 KiB
C
738 lines
18 KiB
C
/*
|
|
* Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved.
|
|
* Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This copyrighted material is made available to anyone wishing to use,
|
|
* modify, copy, or redistribute it subject to the terms and conditions
|
|
* of the GNU General Public License version 2.
|
|
*/
|
|
|
|
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
|
|
|
|
#include <linux/sched.h>
|
|
#include <linux/cred.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/completion.h>
|
|
#include <linux/buffer_head.h>
|
|
#include <linux/module.h>
|
|
#include <linux/kobject.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/gfs2_ondisk.h>
|
|
#include <linux/genhd.h>
|
|
|
|
#include "gfs2.h"
|
|
#include "incore.h"
|
|
#include "sys.h"
|
|
#include "super.h"
|
|
#include "glock.h"
|
|
#include "quota.h"
|
|
#include "util.h"
|
|
#include "glops.h"
|
|
#include "recovery.h"
|
|
|
|
struct gfs2_attr {
|
|
struct attribute attr;
|
|
ssize_t (*show)(struct gfs2_sbd *, char *);
|
|
ssize_t (*store)(struct gfs2_sbd *, const char *, size_t);
|
|
};
|
|
|
|
static ssize_t gfs2_attr_show(struct kobject *kobj, struct attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct gfs2_sbd *sdp = container_of(kobj, struct gfs2_sbd, sd_kobj);
|
|
struct gfs2_attr *a = container_of(attr, struct gfs2_attr, attr);
|
|
return a->show ? a->show(sdp, buf) : 0;
|
|
}
|
|
|
|
static ssize_t gfs2_attr_store(struct kobject *kobj, struct attribute *attr,
|
|
const char *buf, size_t len)
|
|
{
|
|
struct gfs2_sbd *sdp = container_of(kobj, struct gfs2_sbd, sd_kobj);
|
|
struct gfs2_attr *a = container_of(attr, struct gfs2_attr, attr);
|
|
return a->store ? a->store(sdp, buf, len) : len;
|
|
}
|
|
|
|
static const struct sysfs_ops gfs2_attr_ops = {
|
|
.show = gfs2_attr_show,
|
|
.store = gfs2_attr_store,
|
|
};
|
|
|
|
|
|
static struct kset *gfs2_kset;
|
|
|
|
static ssize_t id_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
return snprintf(buf, PAGE_SIZE, "%u:%u\n",
|
|
MAJOR(sdp->sd_vfs->s_dev), MINOR(sdp->sd_vfs->s_dev));
|
|
}
|
|
|
|
static ssize_t fsname_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
return snprintf(buf, PAGE_SIZE, "%s\n", sdp->sd_fsname);
|
|
}
|
|
|
|
static ssize_t uuid_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
struct super_block *s = sdp->sd_vfs;
|
|
|
|
buf[0] = '\0';
|
|
if (uuid_is_null(&s->s_uuid))
|
|
return 0;
|
|
return snprintf(buf, PAGE_SIZE, "%pUB\n", &s->s_uuid);
|
|
}
|
|
|
|
static ssize_t freeze_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
struct super_block *sb = sdp->sd_vfs;
|
|
int frozen = (sb->s_writers.frozen == SB_UNFROZEN) ? 0 : 1;
|
|
|
|
return snprintf(buf, PAGE_SIZE, "%d\n", frozen);
|
|
}
|
|
|
|
static ssize_t freeze_store(struct gfs2_sbd *sdp, const char *buf, size_t len)
|
|
{
|
|
int error, n;
|
|
|
|
error = kstrtoint(buf, 0, &n);
|
|
if (error)
|
|
return error;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
switch (n) {
|
|
case 0:
|
|
error = thaw_super(sdp->sd_vfs);
|
|
break;
|
|
case 1:
|
|
error = freeze_super(sdp->sd_vfs);
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (error) {
|
|
fs_warn(sdp, "freeze %d error %d\n", n, error);
|
|
return error;
|
|
}
|
|
|
|
return len;
|
|
}
|
|
|
|
static ssize_t withdraw_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
unsigned int b = test_bit(SDF_SHUTDOWN, &sdp->sd_flags);
|
|
return snprintf(buf, PAGE_SIZE, "%u\n", b);
|
|
}
|
|
|
|
static ssize_t withdraw_store(struct gfs2_sbd *sdp, const char *buf, size_t len)
|
|
{
|
|
int error, val;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
error = kstrtoint(buf, 0, &val);
|
|
if (error)
|
|
return error;
|
|
|
|
if (val != 1)
|
|
return -EINVAL;
|
|
|
|
gfs2_lm_withdraw(sdp, "withdrawing from cluster at user's request\n");
|
|
|
|
return len;
|
|
}
|
|
|
|
static ssize_t statfs_sync_store(struct gfs2_sbd *sdp, const char *buf,
|
|
size_t len)
|
|
{
|
|
int error, val;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
error = kstrtoint(buf, 0, &val);
|
|
if (error)
|
|
return error;
|
|
|
|
if (val != 1)
|
|
return -EINVAL;
|
|
|
|
gfs2_statfs_sync(sdp->sd_vfs, 0);
|
|
return len;
|
|
}
|
|
|
|
static ssize_t quota_sync_store(struct gfs2_sbd *sdp, const char *buf,
|
|
size_t len)
|
|
{
|
|
int error, val;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
error = kstrtoint(buf, 0, &val);
|
|
if (error)
|
|
return error;
|
|
|
|
if (val != 1)
|
|
return -EINVAL;
|
|
|
|
gfs2_quota_sync(sdp->sd_vfs, 0);
|
|
return len;
|
|
}
|
|
|
|
static ssize_t quota_refresh_user_store(struct gfs2_sbd *sdp, const char *buf,
|
|
size_t len)
|
|
{
|
|
struct kqid qid;
|
|
int error;
|
|
u32 id;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
error = kstrtou32(buf, 0, &id);
|
|
if (error)
|
|
return error;
|
|
|
|
qid = make_kqid(current_user_ns(), USRQUOTA, id);
|
|
if (!qid_valid(qid))
|
|
return -EINVAL;
|
|
|
|
error = gfs2_quota_refresh(sdp, qid);
|
|
return error ? error : len;
|
|
}
|
|
|
|
static ssize_t quota_refresh_group_store(struct gfs2_sbd *sdp, const char *buf,
|
|
size_t len)
|
|
{
|
|
struct kqid qid;
|
|
int error;
|
|
u32 id;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
error = kstrtou32(buf, 0, &id);
|
|
if (error)
|
|
return error;
|
|
|
|
qid = make_kqid(current_user_ns(), GRPQUOTA, id);
|
|
if (!qid_valid(qid))
|
|
return -EINVAL;
|
|
|
|
error = gfs2_quota_refresh(sdp, qid);
|
|
return error ? error : len;
|
|
}
|
|
|
|
static ssize_t demote_rq_store(struct gfs2_sbd *sdp, const char *buf, size_t len)
|
|
{
|
|
struct gfs2_glock *gl;
|
|
const struct gfs2_glock_operations *glops;
|
|
unsigned int glmode;
|
|
unsigned int gltype;
|
|
unsigned long long glnum;
|
|
char mode[16];
|
|
int rv;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
rv = sscanf(buf, "%u:%llu %15s", &gltype, &glnum,
|
|
mode);
|
|
if (rv != 3)
|
|
return -EINVAL;
|
|
|
|
if (strcmp(mode, "EX") == 0)
|
|
glmode = LM_ST_UNLOCKED;
|
|
else if ((strcmp(mode, "CW") == 0) || (strcmp(mode, "DF") == 0))
|
|
glmode = LM_ST_DEFERRED;
|
|
else if ((strcmp(mode, "PR") == 0) || (strcmp(mode, "SH") == 0))
|
|
glmode = LM_ST_SHARED;
|
|
else
|
|
return -EINVAL;
|
|
|
|
if (gltype > LM_TYPE_JOURNAL)
|
|
return -EINVAL;
|
|
if (gltype == LM_TYPE_NONDISK && glnum == GFS2_FREEZE_LOCK)
|
|
glops = &gfs2_freeze_glops;
|
|
else
|
|
glops = gfs2_glops_list[gltype];
|
|
if (glops == NULL)
|
|
return -EINVAL;
|
|
if (!test_and_set_bit(SDF_DEMOTE, &sdp->sd_flags))
|
|
fs_info(sdp, "demote interface used\n");
|
|
rv = gfs2_glock_get(sdp, glnum, glops, 0, &gl);
|
|
if (rv)
|
|
return rv;
|
|
gfs2_glock_cb(gl, glmode);
|
|
gfs2_glock_put(gl);
|
|
return len;
|
|
}
|
|
|
|
|
|
#define GFS2_ATTR(name, mode, show, store) \
|
|
static struct gfs2_attr gfs2_attr_##name = __ATTR(name, mode, show, store)
|
|
|
|
GFS2_ATTR(id, 0444, id_show, NULL);
|
|
GFS2_ATTR(fsname, 0444, fsname_show, NULL);
|
|
GFS2_ATTR(uuid, 0444, uuid_show, NULL);
|
|
GFS2_ATTR(freeze, 0644, freeze_show, freeze_store);
|
|
GFS2_ATTR(withdraw, 0644, withdraw_show, withdraw_store);
|
|
GFS2_ATTR(statfs_sync, 0200, NULL, statfs_sync_store);
|
|
GFS2_ATTR(quota_sync, 0200, NULL, quota_sync_store);
|
|
GFS2_ATTR(quota_refresh_user, 0200, NULL, quota_refresh_user_store);
|
|
GFS2_ATTR(quota_refresh_group, 0200, NULL, quota_refresh_group_store);
|
|
GFS2_ATTR(demote_rq, 0200, NULL, demote_rq_store);
|
|
|
|
static struct attribute *gfs2_attrs[] = {
|
|
&gfs2_attr_id.attr,
|
|
&gfs2_attr_fsname.attr,
|
|
&gfs2_attr_uuid.attr,
|
|
&gfs2_attr_freeze.attr,
|
|
&gfs2_attr_withdraw.attr,
|
|
&gfs2_attr_statfs_sync.attr,
|
|
&gfs2_attr_quota_sync.attr,
|
|
&gfs2_attr_quota_refresh_user.attr,
|
|
&gfs2_attr_quota_refresh_group.attr,
|
|
&gfs2_attr_demote_rq.attr,
|
|
NULL,
|
|
};
|
|
|
|
static void gfs2_sbd_release(struct kobject *kobj)
|
|
{
|
|
struct gfs2_sbd *sdp = container_of(kobj, struct gfs2_sbd, sd_kobj);
|
|
|
|
kfree(sdp);
|
|
}
|
|
|
|
static struct kobj_type gfs2_ktype = {
|
|
.release = gfs2_sbd_release,
|
|
.default_attrs = gfs2_attrs,
|
|
.sysfs_ops = &gfs2_attr_ops,
|
|
};
|
|
|
|
|
|
/*
|
|
* lock_module. Originally from lock_dlm
|
|
*/
|
|
|
|
static ssize_t proto_name_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
const struct lm_lockops *ops = sdp->sd_lockstruct.ls_ops;
|
|
return sprintf(buf, "%s\n", ops->lm_proto_name);
|
|
}
|
|
|
|
static ssize_t block_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
struct lm_lockstruct *ls = &sdp->sd_lockstruct;
|
|
ssize_t ret;
|
|
int val = 0;
|
|
|
|
if (test_bit(DFL_BLOCK_LOCKS, &ls->ls_recover_flags))
|
|
val = 1;
|
|
ret = sprintf(buf, "%d\n", val);
|
|
return ret;
|
|
}
|
|
|
|
static ssize_t block_store(struct gfs2_sbd *sdp, const char *buf, size_t len)
|
|
{
|
|
struct lm_lockstruct *ls = &sdp->sd_lockstruct;
|
|
int ret, val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (val == 1)
|
|
set_bit(DFL_BLOCK_LOCKS, &ls->ls_recover_flags);
|
|
else if (val == 0) {
|
|
clear_bit(DFL_BLOCK_LOCKS, &ls->ls_recover_flags);
|
|
smp_mb__after_atomic();
|
|
gfs2_glock_thaw(sdp);
|
|
} else {
|
|
return -EINVAL;
|
|
}
|
|
return len;
|
|
}
|
|
|
|
static ssize_t wdack_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
int val = completion_done(&sdp->sd_wdack) ? 1 : 0;
|
|
|
|
return sprintf(buf, "%d\n", val);
|
|
}
|
|
|
|
static ssize_t wdack_store(struct gfs2_sbd *sdp, const char *buf, size_t len)
|
|
{
|
|
int ret, val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if ((val == 1) &&
|
|
!strcmp(sdp->sd_lockstruct.ls_ops->lm_proto_name, "lock_dlm"))
|
|
complete(&sdp->sd_wdack);
|
|
else
|
|
return -EINVAL;
|
|
return len;
|
|
}
|
|
|
|
static ssize_t lkfirst_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
struct lm_lockstruct *ls = &sdp->sd_lockstruct;
|
|
return sprintf(buf, "%d\n", ls->ls_first);
|
|
}
|
|
|
|
static ssize_t lkfirst_store(struct gfs2_sbd *sdp, const char *buf, size_t len)
|
|
{
|
|
unsigned first;
|
|
int rv;
|
|
|
|
rv = sscanf(buf, "%u", &first);
|
|
if (rv != 1 || first > 1)
|
|
return -EINVAL;
|
|
rv = wait_for_completion_killable(&sdp->sd_locking_init);
|
|
if (rv)
|
|
return rv;
|
|
spin_lock(&sdp->sd_jindex_spin);
|
|
rv = -EBUSY;
|
|
if (test_bit(SDF_NOJOURNALID, &sdp->sd_flags) == 0)
|
|
goto out;
|
|
rv = -EINVAL;
|
|
if (sdp->sd_args.ar_spectator)
|
|
goto out;
|
|
if (sdp->sd_lockstruct.ls_ops->lm_mount == NULL)
|
|
goto out;
|
|
sdp->sd_lockstruct.ls_first = first;
|
|
rv = 0;
|
|
out:
|
|
spin_unlock(&sdp->sd_jindex_spin);
|
|
return rv ? rv : len;
|
|
}
|
|
|
|
static ssize_t first_done_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
struct lm_lockstruct *ls = &sdp->sd_lockstruct;
|
|
return sprintf(buf, "%d\n", !!test_bit(DFL_FIRST_MOUNT_DONE, &ls->ls_recover_flags));
|
|
}
|
|
|
|
int gfs2_recover_set(struct gfs2_sbd *sdp, unsigned jid)
|
|
{
|
|
struct gfs2_jdesc *jd;
|
|
int rv;
|
|
|
|
/* Wait for our primary journal to be initialized */
|
|
wait_for_completion(&sdp->sd_journal_ready);
|
|
|
|
spin_lock(&sdp->sd_jindex_spin);
|
|
rv = -EBUSY;
|
|
/**
|
|
* If we're a spectator, we use journal0, but it's not really ours.
|
|
* So we need to wait for its recovery too. If we skip it we'd never
|
|
* queue work to the recovery workqueue, and so its completion would
|
|
* never clear the DFL_BLOCK_LOCKS flag, so all our locks would
|
|
* permanently stop working.
|
|
*/
|
|
if (sdp->sd_jdesc->jd_jid == jid && !sdp->sd_args.ar_spectator)
|
|
goto out;
|
|
rv = -ENOENT;
|
|
list_for_each_entry(jd, &sdp->sd_jindex_list, jd_list) {
|
|
if (jd->jd_jid != jid && !sdp->sd_args.ar_spectator)
|
|
continue;
|
|
rv = gfs2_recover_journal(jd, false);
|
|
break;
|
|
}
|
|
out:
|
|
spin_unlock(&sdp->sd_jindex_spin);
|
|
return rv;
|
|
}
|
|
|
|
static ssize_t recover_store(struct gfs2_sbd *sdp, const char *buf, size_t len)
|
|
{
|
|
unsigned jid;
|
|
int rv;
|
|
|
|
rv = sscanf(buf, "%u", &jid);
|
|
if (rv != 1)
|
|
return -EINVAL;
|
|
|
|
if (test_bit(SDF_NORECOVERY, &sdp->sd_flags)) {
|
|
rv = -ESHUTDOWN;
|
|
goto out;
|
|
}
|
|
|
|
rv = gfs2_recover_set(sdp, jid);
|
|
out:
|
|
return rv ? rv : len;
|
|
}
|
|
|
|
static ssize_t recover_done_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
struct lm_lockstruct *ls = &sdp->sd_lockstruct;
|
|
return sprintf(buf, "%d\n", ls->ls_recover_jid_done);
|
|
}
|
|
|
|
static ssize_t recover_status_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
struct lm_lockstruct *ls = &sdp->sd_lockstruct;
|
|
return sprintf(buf, "%d\n", ls->ls_recover_jid_status);
|
|
}
|
|
|
|
static ssize_t jid_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
return sprintf(buf, "%d\n", sdp->sd_lockstruct.ls_jid);
|
|
}
|
|
|
|
static ssize_t jid_store(struct gfs2_sbd *sdp, const char *buf, size_t len)
|
|
{
|
|
int jid;
|
|
int rv;
|
|
|
|
rv = sscanf(buf, "%d", &jid);
|
|
if (rv != 1)
|
|
return -EINVAL;
|
|
rv = wait_for_completion_killable(&sdp->sd_locking_init);
|
|
if (rv)
|
|
return rv;
|
|
spin_lock(&sdp->sd_jindex_spin);
|
|
rv = -EINVAL;
|
|
if (sdp->sd_lockstruct.ls_ops->lm_mount == NULL)
|
|
goto out;
|
|
rv = -EBUSY;
|
|
if (test_bit(SDF_NOJOURNALID, &sdp->sd_flags) == 0)
|
|
goto out;
|
|
rv = 0;
|
|
if (sdp->sd_args.ar_spectator && jid > 0)
|
|
rv = jid = -EINVAL;
|
|
sdp->sd_lockstruct.ls_jid = jid;
|
|
clear_bit(SDF_NOJOURNALID, &sdp->sd_flags);
|
|
smp_mb__after_atomic();
|
|
wake_up_bit(&sdp->sd_flags, SDF_NOJOURNALID);
|
|
out:
|
|
spin_unlock(&sdp->sd_jindex_spin);
|
|
return rv ? rv : len;
|
|
}
|
|
|
|
#define GDLM_ATTR(_name,_mode,_show,_store) \
|
|
static struct gfs2_attr gdlm_attr_##_name = __ATTR(_name,_mode,_show,_store)
|
|
|
|
GDLM_ATTR(proto_name, 0444, proto_name_show, NULL);
|
|
GDLM_ATTR(block, 0644, block_show, block_store);
|
|
GDLM_ATTR(withdraw, 0644, wdack_show, wdack_store);
|
|
GDLM_ATTR(jid, 0644, jid_show, jid_store);
|
|
GDLM_ATTR(first, 0644, lkfirst_show, lkfirst_store);
|
|
GDLM_ATTR(first_done, 0444, first_done_show, NULL);
|
|
GDLM_ATTR(recover, 0600, NULL, recover_store);
|
|
GDLM_ATTR(recover_done, 0444, recover_done_show, NULL);
|
|
GDLM_ATTR(recover_status, 0444, recover_status_show, NULL);
|
|
|
|
static struct attribute *lock_module_attrs[] = {
|
|
&gdlm_attr_proto_name.attr,
|
|
&gdlm_attr_block.attr,
|
|
&gdlm_attr_withdraw.attr,
|
|
&gdlm_attr_jid.attr,
|
|
&gdlm_attr_first.attr,
|
|
&gdlm_attr_first_done.attr,
|
|
&gdlm_attr_recover.attr,
|
|
&gdlm_attr_recover_done.attr,
|
|
&gdlm_attr_recover_status.attr,
|
|
NULL,
|
|
};
|
|
|
|
/*
|
|
* get and set struct gfs2_tune fields
|
|
*/
|
|
|
|
static ssize_t quota_scale_show(struct gfs2_sbd *sdp, char *buf)
|
|
{
|
|
return snprintf(buf, PAGE_SIZE, "%u %u\n",
|
|
sdp->sd_tune.gt_quota_scale_num,
|
|
sdp->sd_tune.gt_quota_scale_den);
|
|
}
|
|
|
|
static ssize_t quota_scale_store(struct gfs2_sbd *sdp, const char *buf,
|
|
size_t len)
|
|
{
|
|
struct gfs2_tune *gt = &sdp->sd_tune;
|
|
unsigned int x, y;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
if (sscanf(buf, "%u %u", &x, &y) != 2 || !y)
|
|
return -EINVAL;
|
|
|
|
spin_lock(>->gt_spin);
|
|
gt->gt_quota_scale_num = x;
|
|
gt->gt_quota_scale_den = y;
|
|
spin_unlock(>->gt_spin);
|
|
return len;
|
|
}
|
|
|
|
static ssize_t tune_set(struct gfs2_sbd *sdp, unsigned int *field,
|
|
int check_zero, const char *buf, size_t len)
|
|
{
|
|
struct gfs2_tune *gt = &sdp->sd_tune;
|
|
unsigned int x;
|
|
int error;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
error = kstrtouint(buf, 0, &x);
|
|
if (error)
|
|
return error;
|
|
|
|
if (check_zero && !x)
|
|
return -EINVAL;
|
|
|
|
spin_lock(>->gt_spin);
|
|
*field = x;
|
|
spin_unlock(>->gt_spin);
|
|
return len;
|
|
}
|
|
|
|
#define TUNE_ATTR_3(name, show, store) \
|
|
static struct gfs2_attr tune_attr_##name = __ATTR(name, 0644, show, store)
|
|
|
|
#define TUNE_ATTR_2(name, store) \
|
|
static ssize_t name##_show(struct gfs2_sbd *sdp, char *buf) \
|
|
{ \
|
|
return snprintf(buf, PAGE_SIZE, "%u\n", sdp->sd_tune.gt_##name); \
|
|
} \
|
|
TUNE_ATTR_3(name, name##_show, store)
|
|
|
|
#define TUNE_ATTR(name, check_zero) \
|
|
static ssize_t name##_store(struct gfs2_sbd *sdp, const char *buf, size_t len)\
|
|
{ \
|
|
return tune_set(sdp, &sdp->sd_tune.gt_##name, check_zero, buf, len); \
|
|
} \
|
|
TUNE_ATTR_2(name, name##_store)
|
|
|
|
TUNE_ATTR(quota_warn_period, 0);
|
|
TUNE_ATTR(quota_quantum, 0);
|
|
TUNE_ATTR(max_readahead, 0);
|
|
TUNE_ATTR(complain_secs, 0);
|
|
TUNE_ATTR(statfs_slow, 0);
|
|
TUNE_ATTR(new_files_jdata, 0);
|
|
TUNE_ATTR(statfs_quantum, 1);
|
|
TUNE_ATTR_3(quota_scale, quota_scale_show, quota_scale_store);
|
|
|
|
static struct attribute *tune_attrs[] = {
|
|
&tune_attr_quota_warn_period.attr,
|
|
&tune_attr_quota_quantum.attr,
|
|
&tune_attr_max_readahead.attr,
|
|
&tune_attr_complain_secs.attr,
|
|
&tune_attr_statfs_slow.attr,
|
|
&tune_attr_statfs_quantum.attr,
|
|
&tune_attr_quota_scale.attr,
|
|
&tune_attr_new_files_jdata.attr,
|
|
NULL,
|
|
};
|
|
|
|
static const struct attribute_group tune_group = {
|
|
.name = "tune",
|
|
.attrs = tune_attrs,
|
|
};
|
|
|
|
static const struct attribute_group lock_module_group = {
|
|
.name = "lock_module",
|
|
.attrs = lock_module_attrs,
|
|
};
|
|
|
|
int gfs2_sys_fs_add(struct gfs2_sbd *sdp)
|
|
{
|
|
struct super_block *sb = sdp->sd_vfs;
|
|
int error;
|
|
char ro[20];
|
|
char spectator[20];
|
|
char *envp[] = { ro, spectator, NULL };
|
|
int sysfs_frees_sdp = 0;
|
|
|
|
sprintf(ro, "RDONLY=%d", sb_rdonly(sb));
|
|
sprintf(spectator, "SPECTATOR=%d", sdp->sd_args.ar_spectator ? 1 : 0);
|
|
|
|
sdp->sd_kobj.kset = gfs2_kset;
|
|
error = kobject_init_and_add(&sdp->sd_kobj, &gfs2_ktype, NULL,
|
|
"%s", sdp->sd_table_name);
|
|
if (error)
|
|
goto fail_reg;
|
|
|
|
sysfs_frees_sdp = 1; /* Freeing sdp is now done by sysfs calling
|
|
function gfs2_sbd_release. */
|
|
error = sysfs_create_group(&sdp->sd_kobj, &tune_group);
|
|
if (error)
|
|
goto fail_reg;
|
|
|
|
error = sysfs_create_group(&sdp->sd_kobj, &lock_module_group);
|
|
if (error)
|
|
goto fail_tune;
|
|
|
|
error = sysfs_create_link(&sdp->sd_kobj,
|
|
&disk_to_dev(sb->s_bdev->bd_disk)->kobj,
|
|
"device");
|
|
if (error)
|
|
goto fail_lock_module;
|
|
|
|
kobject_uevent_env(&sdp->sd_kobj, KOBJ_ADD, envp);
|
|
return 0;
|
|
|
|
fail_lock_module:
|
|
sysfs_remove_group(&sdp->sd_kobj, &lock_module_group);
|
|
fail_tune:
|
|
sysfs_remove_group(&sdp->sd_kobj, &tune_group);
|
|
fail_reg:
|
|
free_percpu(sdp->sd_lkstats);
|
|
fs_err(sdp, "error %d adding sysfs files\n", error);
|
|
if (sysfs_frees_sdp)
|
|
kobject_put(&sdp->sd_kobj);
|
|
else
|
|
kfree(sdp);
|
|
sb->s_fs_info = NULL;
|
|
return error;
|
|
}
|
|
|
|
void gfs2_sys_fs_del(struct gfs2_sbd *sdp)
|
|
{
|
|
sysfs_remove_link(&sdp->sd_kobj, "device");
|
|
sysfs_remove_group(&sdp->sd_kobj, &tune_group);
|
|
sysfs_remove_group(&sdp->sd_kobj, &lock_module_group);
|
|
kobject_put(&sdp->sd_kobj);
|
|
}
|
|
|
|
static int gfs2_uevent(struct kset *kset, struct kobject *kobj,
|
|
struct kobj_uevent_env *env)
|
|
{
|
|
struct gfs2_sbd *sdp = container_of(kobj, struct gfs2_sbd, sd_kobj);
|
|
struct super_block *s = sdp->sd_vfs;
|
|
|
|
add_uevent_var(env, "LOCKTABLE=%s", sdp->sd_table_name);
|
|
add_uevent_var(env, "LOCKPROTO=%s", sdp->sd_proto_name);
|
|
if (!test_bit(SDF_NOJOURNALID, &sdp->sd_flags))
|
|
add_uevent_var(env, "JOURNALID=%d", sdp->sd_lockstruct.ls_jid);
|
|
if (!uuid_is_null(&s->s_uuid))
|
|
add_uevent_var(env, "UUID=%pUB", &s->s_uuid);
|
|
return 0;
|
|
}
|
|
|
|
static const struct kset_uevent_ops gfs2_uevent_ops = {
|
|
.uevent = gfs2_uevent,
|
|
};
|
|
|
|
int gfs2_sys_init(void)
|
|
{
|
|
gfs2_kset = kset_create_and_add("gfs2", &gfs2_uevent_ops, fs_kobj);
|
|
if (!gfs2_kset)
|
|
return -ENOMEM;
|
|
return 0;
|
|
}
|
|
|
|
void gfs2_sys_uninit(void)
|
|
{
|
|
kset_unregister(gfs2_kset);
|
|
}
|
|
|