@@ -1539,6 +1539,18 @@ struct lu_tgt_desc {
ltd_connecting:1; /* target is connecting */
};
+static inline u64 tgt_statfs_bavail(struct lu_tgt_desc *tgt)
+{
+ struct obd_statfs *statfs = &tgt->ltd_statfs;
+
+ return statfs->os_bavail * statfs->os_bsize;
+}
+
+static inline u64 tgt_statfs_iavail(struct lu_tgt_desc *tgt)
+{
+ return tgt->ltd_statfs.os_ffree;
+}
+
/* number of pointers at 2nd level */
#define TGT_PTRS_PER_BLOCK (PAGE_SIZE / sizeof(void *))
/* number of pointers at 1st level - only need as many as max OST/MDT count */
@@ -1593,7 +1605,7 @@ struct lu_tgt_descs {
u64 lu_prandom_u64_max(u64 ep_ro);
int lu_qos_add_tgt(struct lu_qos *qos, struct lu_tgt_desc *ltd);
int lu_qos_del_tgt(struct lu_qos *qos, struct lu_tgt_desc *ltd);
-void lu_tgt_qos_weight_calc(struct lu_tgt_desc *tgt);
+void lu_tgt_qos_weight_calc(struct lu_tgt_desc *tgt, bool is_mdt);
int lu_tgt_descs_init(struct lu_tgt_descs *ltd, bool is_mdt);
void lu_tgt_descs_fini(struct lu_tgt_descs *ltd);
@@ -1512,7 +1512,7 @@ static struct lu_tgt_desc *lmv_locate_tgt_qos(struct lmv_obd *lmv,
}
tgt->ltd_qos.ltq_usable = 1;
- lu_tgt_qos_weight_calc(tgt);
+ lu_tgt_qos_weight_calc(tgt, true);
if (tgt->ltd_index == op_data->op_mds)
cur = tgt;
total_avail += tgt->ltd_qos.ltq_avail;
@@ -1613,7 +1613,7 @@ static struct lu_tgt_desc *lmv_locate_tgt_lf(struct lmv_obd *lmv)
}
tgt->ltd_qos.ltq_usable = 1;
- lu_tgt_qos_weight_calc(tgt);
+ lu_tgt_qos_weight_calc(tgt, true);
avail += tgt->ltd_qos.ltq_avail;
if (!min || min->ltd_qos.ltq_avail > tgt->ltd_qos.ltq_avail)
min = tgt;
@@ -198,33 +198,26 @@ int lu_qos_del_tgt(struct lu_qos *qos, struct lu_tgt_desc *ltd)
}
EXPORT_SYMBOL(lu_qos_del_tgt);
-static inline u64 tgt_statfs_bavail(struct lu_tgt_desc *tgt)
-{
- struct obd_statfs *statfs = &tgt->ltd_statfs;
-
- return statfs->os_bavail * statfs->os_bsize;
-}
-
-static inline u64 tgt_statfs_iavail(struct lu_tgt_desc *tgt)
-{
- return tgt->ltd_statfs.os_ffree;
-}
-
/**
* Calculate weight for a given tgt.
*
- * The final tgt weight is bavail >> 16 * iavail >> 8 minus the tgt and server
- * penalties. See ltd_qos_penalties_calc() for how penalties are calculated.
+ * The final tgt weight uses only free space for OSTs, but combines
+ * both free space and inodes for MDTs, minus tgt and server penalties.
+ * See ltd_qos_penalties_calc() for how penalties are calculated.
*
* @tgt target descriptor
+ * @is_mdt target table is for MDT selection (use inodes)
*/
-void lu_tgt_qos_weight_calc(struct lu_tgt_desc *tgt)
+void lu_tgt_qos_weight_calc(struct lu_tgt_desc *tgt, bool is_mdt)
{
struct lu_tgt_qos *ltq = &tgt->ltd_qos;
u64 penalty;
- ltq->ltq_avail = (tgt_statfs_bavail(tgt) >> 16) *
- (tgt_statfs_iavail(tgt) >> 8);
+ if (is_mdt)
+ ltq->ltq_avail = (tgt_statfs_bavail(tgt) >> 16) *
+ (tgt_statfs_iavail(tgt) >> 8);
+ else
+ ltq->ltq_avail = tgt_statfs_bavail(tgt) >> 8;
penalty = ltq->ltq_penalty + ltq->ltq_svr->lsq_penalty;
if (ltq->ltq_avail < penalty)
ltq->ltq_weight = 0;
@@ -512,11 +505,10 @@ int ltd_qos_penalties_calc(struct lu_tgt_descs *ltd)
/*
* per-tgt penalty is
- * prio * bavail * iavail / (num_tgt - 1) / 2
+ * prio * bavail * iavail / (num_tgt - 1) / prio_max / 2
*/
- tgt->ltd_qos.ltq_penalty_per_obj = prio_wide * ba * ia >> 8;
+ tgt->ltd_qos.ltq_penalty_per_obj = prio_wide * ba * ia >> 9;
do_div(tgt->ltd_qos.ltq_penalty_per_obj, num_active);
- tgt->ltd_qos.ltq_penalty_per_obj >>= 1;
age = (now - tgt->ltd_qos.ltq_used) >> 3;
if (test_bit(LQ_RESET, &qos->lq_flags) ||
@@ -563,14 +555,11 @@ int ltd_qos_penalties_calc(struct lu_tgt_descs *ltd)
svr->lsq_penalty >>= age / desc->ld_qos_maxage;
}
- clear_bit(LQ_DIRTY, &qos->lq_flags);
- clear_bit(LQ_RESET, &qos->lq_flags);
/*
* If each tgt has almost same free space, do rr allocation for better
* creation performance
*/
- clear_bit(LQ_SAME_SPACE, &qos->lq_flags);
if (((ba_max * (QOS_THRESHOLD_MAX - qos->lq_threshold_rr)) /
QOS_THRESHOLD_MAX) < ba_min &&
((ia_max * (QOS_THRESHOLD_MAX - qos->lq_threshold_rr)) /
@@ -578,7 +567,11 @@ int ltd_qos_penalties_calc(struct lu_tgt_descs *ltd)
set_bit(LQ_SAME_SPACE, &qos->lq_flags);
/* Reset weights for the next time we enter qos mode */
set_bit(LQ_RESET, &qos->lq_flags);
+ } else {
+ clear_bit(LQ_SAME_SPACE, &qos->lq_flags);
+ clear_bit(LQ_RESET, &qos->lq_flags);
}
+ clear_bit(LQ_DIRTY, &qos->lq_flags);
rc = 0;
out:
@@ -653,7 +646,7 @@ int ltd_qos_update(struct lu_tgt_descs *ltd, struct lu_tgt_desc *tgt,
else
ltq->ltq_penalty -= ltq->ltq_penalty_per_obj;
- lu_tgt_qos_weight_calc(tgt);
+ lu_tgt_qos_weight_calc(tgt, ltd->ltd_is_mdt);
/* Recalc the total weight of usable osts */
if (ltq->ltq_usable)