/* object cache. */
cfs_mem_cache_t *lmv_object_cache;
-atomic_t lmv_object_count = ATOMIC_INIT(0);
+cfs_atomic_t lmv_object_count = CFS_ATOMIC_INIT(0);
static void lmv_activate_target(struct lmv_obd *lmv,
struct lmv_tgt_desc *tgt,
CDEBUG(D_INFO, "Searching in lmv %p for uuid %s (activate=%d)\n",
lmv, uuid->uuid, activate);
- spin_lock(&lmv->lmv_lock);
+ cfs_spin_lock(&lmv->lmv_lock);
for (i = 0, tgt = lmv->tgts; i < lmv->desc.ld_tgt_count; i++, tgt++) {
if (tgt->ltd_exp == NULL)
continue;
EXIT;
out_lmv_lock:
- spin_unlock(&lmv->lmv_lock);
+ cfs_spin_unlock(&lmv->lmv_lock);
return rc;
}
LASSERT(data != NULL);
- spin_lock(&lmv->lmv_lock);
+ cfs_spin_lock(&lmv->lmv_lock);
for (i = 0, tgt = lmv->tgts; i < lmv->desc.ld_tgt_count; i++, tgt++) {
if (tgt->ltd_exp == NULL)
continue;
break;
}
}
- spin_unlock(&lmv->lmv_lock);
+ cfs_spin_unlock(&lmv->lmv_lock);
RETURN(0);
}
CDEBUG(D_CONFIG, "Connected to %s(%s) successfully (%d)\n",
mdc_obd->obd_name, mdc_obd->obd_uuid.uuid,
- atomic_read(&obd->obd_refcount));
+ cfs_atomic_read(&obd->obd_refcount));
#ifdef __KERNEL__
lmv_proc_dir = lprocfs_srch(obd->obd_proc_entry, "target_obds");
CERROR("lmv failed to setup llogging subsystems\n");
}
}
- spin_lock(&lmv->lmv_lock);
+ cfs_spin_lock(&lmv->lmv_lock);
tgt = lmv->tgts + lmv->desc.ld_tgt_count++;
tgt->ltd_uuid = *tgt_uuid;
- spin_unlock(&lmv->lmv_lock);
+ cfs_spin_unlock(&lmv->lmv_lock);
if (lmv->connected) {
rc = lmv_connect_mdc(obd, tgt);
if (rc) {
- spin_lock(&lmv->lmv_lock);
+ cfs_spin_lock(&lmv->lmv_lock);
lmv->desc.ld_tgt_count--;
memset(tgt, 0, sizeof(*tgt));
- spin_unlock(&lmv->lmv_lock);
+ cfs_spin_unlock(&lmv->lmv_lock);
} else {
int easize = sizeof(struct lmv_stripe_md) +
lmv->desc.ld_tgt_count *
RETURN(-EINVAL);
rc = obd_statfs(mdc_obd, &stat_buf,
- cfs_time_current_64() - HZ, 0);
+ cfs_time_current_64() - CFS_HZ, 0);
if (rc)
RETURN(rc);
- if (copy_to_user(data->ioc_pbuf1, &stat_buf, data->ioc_plen1))
+ if (cfs_copy_to_user(data->ioc_pbuf1, &stat_buf,
+ data->ioc_plen1))
RETURN(-EFAULT);
- if (copy_to_user(data->ioc_pbuf2, obd2cli_tgt(mdc_obd),
- data->ioc_plen2))
+ if (cfs_copy_to_user(data->ioc_pbuf2, obd2cli_tgt(mdc_obd),
+ data->ioc_plen2))
RETURN(-EFAULT);
break;
}
* New seq alloc and FLD setup should be atomic. Otherwise we may find
* on server that seq in new allocated fid is not yet known.
*/
- down(&tgt->ltd_fid_sem);
+ cfs_down(&tgt->ltd_fid_sem);
if (!tgt->ltd_active)
GOTO(out, rc = -ENODEV);
EXIT;
out:
- up(&tgt->ltd_fid_sem);
+ cfs_up(&tgt->ltd_fid_sem);
return rc;
}
RETURN(-ENOMEM);
for (i = 0; i < LMV_MAX_TGT_COUNT; i++) {
- sema_init(&lmv->tgts[i].ltd_fid_sem, 1);
+ cfs_sema_init(&lmv->tgts[i].ltd_fid_sem, 1);
lmv->tgts[i].ltd_idx = i;
}
lmv->max_easize = 0;
lmv->lmv_placement = PLACEMENT_CHAR_POLICY;
- spin_lock_init(&lmv->lmv_lock);
- sema_init(&lmv->init_sem, 1);
+ cfs_spin_lock_init(&lmv->lmv_lock);
+ cfs_sema_init(&lmv->init_sem, 1);
rc = lmv_object_setup(obd);
if (rc) {
lprocfs_lmv_init_vars(&lvars);
- request_module("lquota");
+ cfs_request_module("lquota");
quota_interface = PORTAL_SYMBOL_GET(lmv_quota_interface);
init_obd_quota_ops(quota_interface, &lmv_obd_ops);
class_unregister_type(LUSTRE_LMV_NAME);
- LASSERTF(atomic_read(&lmv_object_count) == 0,
+ LASSERTF(cfs_atomic_read(&lmv_object_count) == 0,
"Can't free lmv objects cache, %d object(s) busy\n",
- atomic_read(&lmv_object_count));
+ cfs_atomic_read(&lmv_object_count));
cfs_mem_cache_destroy(lmv_object_cache);
}