#include <linux/buffer_head.h>
#include <linux/sort.h>
#include <linux/fs.h>
+#include <linux/bio.h>
#include <linux/gfs2_ondisk.h>
+#include <linux/lm_interface.h>
#include "gfs2.h"
-#include "lm_interface.h"
#include "incore.h"
#include "bmap.h"
#include "glock.h"
struct gfs2_sbd *sdp = qd->qd_gl->gl_sbd;
struct gfs2_inode *ip = GFS2_I(sdp->sd_qc_inode);
unsigned int block, offset;
- u64 dblock;
- int new = 0;
struct buffer_head *bh;
int error;
- int boundary;
+ struct buffer_head bh_map = { .b_state = 0, .b_blocknr = 0 };
mutex_lock(&sdp->sd_quota_mutex);
block = qd->qd_slot / sdp->sd_qc_per_block;
offset = qd->qd_slot % sdp->sd_qc_per_block;;
- error = gfs2_block_map(&ip->i_inode, block, &new, &dblock, &boundary);
+ bh_map.b_size = 1 << ip->i_inode.i_blkbits;
+ error = gfs2_block_map(&ip->i_inode, block, 0, &bh_map);
if (error)
goto fail;
- error = gfs2_meta_read(ip->i_gl, dblock, DIO_START | DIO_WAIT, &bh);
+ error = gfs2_meta_read(ip->i_gl, bh_map.b_blocknr, DIO_WAIT, &bh);
if (error)
goto fail;
error = -EIO;
if (sdp->sd_args.ar_quota == GFS2_QUOTA_OFF)
return 0;
- error = qdsb_get(sdp, QUOTA_USER, ip->i_di.di_uid, CREATE, qd);
+ error = qdsb_get(sdp, QUOTA_USER, ip->i_inode.i_uid, CREATE, qd);
if (error)
goto out;
al->al_qd_num++;
qd++;
- error = qdsb_get(sdp, QUOTA_GROUP, ip->i_di.di_gid, CREATE, qd);
+ error = qdsb_get(sdp, QUOTA_GROUP, ip->i_inode.i_gid, CREATE, qd);
if (error)
goto out;
al->al_qd_num++;
qd++;
- if (uid != NO_QUOTA_CHANGE && uid != ip->i_di.di_uid) {
+ if (uid != NO_QUOTA_CHANGE && uid != ip->i_inode.i_uid) {
error = qdsb_get(sdp, QUOTA_USER, uid, CREATE, qd);
if (error)
goto out;
qd++;
}
- if (gid != NO_QUOTA_CHANGE && gid != ip->i_di.di_gid) {
+ if (gid != NO_QUOTA_CHANGE && gid != ip->i_inode.i_gid) {
error = qdsb_get(sdp, QUOTA_GROUP, gid, CREATE, qd);
if (error)
goto out;
static int sort_qd(const void *a, const void *b)
{
- struct gfs2_quota_data *qd_a = *(struct gfs2_quota_data **)a;
- struct gfs2_quota_data *qd_b = *(struct gfs2_quota_data **)b;
- int ret = 0;
+ const struct gfs2_quota_data *qd_a = *(const struct gfs2_quota_data **)a;
+ const struct gfs2_quota_data *qd_b = *(const struct gfs2_quota_data **)b;
if (!test_bit(QDF_USER, &qd_a->qd_flags) !=
!test_bit(QDF_USER, &qd_b->qd_flags)) {
if (test_bit(QDF_USER, &qd_a->qd_flags))
- ret = -1;
+ return -1;
else
- ret = 1;
- } else {
- if (qd_a->qd_id < qd_b->qd_id)
- ret = -1;
- else if (qd_a->qd_id > qd_b->qd_id)
- ret = 1;
+ return 1;
}
+ if (qd_a->qd_id < qd_b->qd_id)
+ return -1;
+ if (qd_a->qd_id > qd_b->qd_id)
+ return 1;
- return ret;
+ return 0;
}
static void do_qc(struct gfs2_quota_data *qd, s64 change)
qc->qc_id = cpu_to_be32(qd->qd_id);
}
- x = qc->qc_change;
- x = be64_to_cpu(x) + change;
+ x = be64_to_cpu(qc->qc_change) + change;
qc->qc_change = cpu_to_be64(x);
spin_lock(&sdp->sd_quota_spin);
qd_hold(qd);
slot_hold(qd);
}
-
+
mutex_unlock(&sdp->sd_quota_mutex);
}
set_buffer_uptodate(bh);
if (!buffer_uptodate(bh)) {
- ll_rw_block(READ, 1, &bh);
+ ll_rw_block(READ_META, 1, &bh);
wait_on_buffer(bh);
if (!buffer_uptodate(bh))
goto unlock;
gfs2_trans_add_bh(ip->i_gl, bh, 0);
kaddr = kmap_atomic(page, KM_USER0);
- ptr = (__be64 *)(kaddr + offset);
+ ptr = kaddr + offset;
value = (s64)be64_to_cpu(*ptr) + change;
*ptr = cpu_to_be64(value);
flush_dcache_page(page);
kunmap_atomic(kaddr, KM_USER0);
err = 0;
qd->qd_qb.qb_magic = cpu_to_be32(GFS2_MAGIC);
-#if 0
- qd->qd_qb.qb_limit = cpu_to_be64(q.qu_limit);
- qd->qd_qb.qb_warn = cpu_to_be64(q.qu_warn);
-#endif
qd->qd_qb.qb_value = cpu_to_be64(value);
unlock:
unlock_page(page);
struct gfs2_sbd *sdp = qd->qd_gl->gl_sbd;
struct gfs2_inode *ip = GFS2_I(sdp->sd_quota_inode);
struct gfs2_holder i_gh;
- struct gfs2_quota q;
+ struct gfs2_quota_host q;
char buf[sizeof(struct gfs2_quota)];
struct file_ra_state ra_state;
int error;
gfs2_glock_dq_uninit(&i_gh);
-
+
gfs2_quota_in(&q, buf);
qlvb = (struct gfs2_quota_lvb *)qd->qd_gl->gl_lvb;
qlvb->qb_magic = cpu_to_be32(GFS2_MAGIC);
return error;
}
-#if 0
-int gfs2_quota_read(struct gfs2_sbd *sdp, int user, u32 id,
- struct gfs2_quota *q)
-{
- struct gfs2_quota_data *qd;
- struct gfs2_holder q_gh;
- int error;
-
- if (((user) ? (id != current->fsuid) : (!in_group_p(id))) &&
- !capable(CAP_SYS_ADMIN))
- return -EACCES;
-
- error = qd_get(sdp, user, id, CREATE, &qd);
- if (error)
- return error;
-
- error = do_glock(qd, NO_FORCE, &q_gh);
- if (error)
- goto out;
-
- memset(q, 0, sizeof(struct gfs2_quota));
- q->qu_limit = be64_to_cpu(qd->qd_qb.qb_limit);
- q->qu_warn = be64_to_cpu(qd->qd_qb.qb_warn);
- q->qu_value = be64_to_cpu(qd->qd_qb.qb_value);
-
- spin_lock(&sdp->sd_quota_spin);
- q->qu_value += qd->qd_change;
- spin_unlock(&sdp->sd_quota_spin);
-
- gfs2_glock_dq_uninit(&q_gh);
-
-out:
- qd_put(qd);
- return error;
-}
-#endif /* 0 */
-
int gfs2_quota_init(struct gfs2_sbd *sdp)
{
struct gfs2_inode *ip = GFS2_I(sdp->sd_qc_inode);
u32 extlen = 0;
int error;
- if (!ip->i_di.di_size ||
- ip->i_di.di_size > (64 << 20) ||
+ if (!ip->i_di.di_size || ip->i_di.di_size > (64 << 20) ||
ip->i_di.di_size & (sdp->sd_sb.sb_bsize - 1)) {
gfs2_consist_inode(ip);
- return -EIO;
+ return -EIO;
}
sdp->sd_quota_slots = blocks * sdp->sd_qc_per_block;
sdp->sd_quota_chunks = DIV_ROUND_UP(sdp->sd_quota_slots, 8 * PAGE_SIZE);
if (error)
goto fail;
}
- gfs2_meta_ra(ip->i_gl, dblock, extlen);
- error = gfs2_meta_read(ip->i_gl, dblock, DIO_START | DIO_WAIT,
- &bh);
- if (error)
- goto fail;
error = -EIO;
+ bh = gfs2_meta_ra(ip->i_gl, dblock, extlen);
+ if (!bh)
+ goto fail;
if (gfs2_metatype_check(sdp, bh, GFS2_METATYPE_QC)) {
brelse(bh);
goto fail;
}
- for (y = 0;
- y < sdp->sd_qc_per_block && slot < sdp->sd_quota_slots;
+ for (y = 0; y < sdp->sd_qc_per_block && slot < sdp->sd_quota_slots;
y++, slot++) {
- struct gfs2_quota_change qc;
+ struct gfs2_quota_change_host qc;
struct gfs2_quota_data *qd;
gfs2_quota_change_in(&qc, bh->b_data +