static int ll_rd_max_cached_mb(char *page, char **start, off_t off,
int count, int *eof, void *data)
{
- struct super_block *sb = data;
- struct ll_sb_info *sbi = ll_s2sbi(sb);
- struct cl_client_lru *lru = &sbi->ll_lru;
+ struct super_block *sb = data;
+ struct ll_sb_info *sbi = ll_s2sbi(sb);
+ struct cl_client_cache *cache = &sbi->ll_cache;
int shift = 20 - CFS_PAGE_SHIFT;
int max_cached_mb;
int unused_mb;
*eof = 1;
- max_cached_mb = lru->ccl_page_max >> shift;
- unused_mb = cfs_atomic_read(&lru->ccl_page_left) >> shift;
+ max_cached_mb = cache->ccc_lru_max >> shift;
+ unused_mb = cfs_atomic_read(&cache->ccc_lru_left) >> shift;
return snprintf(page, count,
"users: %d\n"
"max_cached_mb: %d\n"
"used_mb: %d\n"
"unused_mb: %d\n"
"reclaim_count: %u\n",
- cfs_atomic_read(&lru->ccl_users),
+ cfs_atomic_read(&cache->ccc_users),
max_cached_mb,
max_cached_mb - unused_mb,
unused_mb,
- lru->ccl_reclaim_count);
+ cache->ccc_lru_shrinkers);
}
static int ll_wr_max_cached_mb(struct file *file, const char *buffer,
{
struct super_block *sb = data;
struct ll_sb_info *sbi = ll_s2sbi(sb);
- struct cl_client_lru *lru = &sbi->ll_lru;
+ struct cl_client_cache *cache = &sbi->ll_cache;
int mult, rc, pages_number;
int diff = 0;
int nrpages = 0;
RETURN(-ENODEV);
cfs_spin_lock(&sbi->ll_lock);
- diff = pages_number - lru->ccl_page_max;
+ diff = pages_number - cache->ccc_lru_max;
cfs_spin_unlock(&sbi->ll_lock);
/* easy - add more LRU slots. */
if (diff >= 0) {
- cfs_atomic_add(diff, &lru->ccl_page_left);
+ cfs_atomic_add(diff, &cache->ccc_lru_left);
GOTO(out, rc = 0);
}
do {
int ov, nv;
- ov = cfs_atomic_read(&lru->ccl_page_left);
+ ov = cfs_atomic_read(&cache->ccc_lru_left);
if (ov == 0)
break;
nv = ov > diff ? ov - diff : 0;
- rc = cfs_atomic_cmpxchg(&lru->ccl_page_left, ov, nv);
+ rc = cfs_atomic_cmpxchg(&cache->ccc_lru_left, ov, nv);
if (likely(ov == rc)) {
diff -= ov - nv;
nrpages += ov - nv;
/* difficult - have to ask OSCs to drop LRU slots. */
tmp = diff << 1;
rc = obd_set_info_async(NULL, sbi->ll_dt_exp,
- sizeof(KEY_LRU_SHRINK), KEY_LRU_SHRINK,
+ sizeof(KEY_CACHE_LRU_SHRINK),
+ KEY_CACHE_LRU_SHRINK,
sizeof(tmp), &tmp, NULL);
if (rc < 0)
break;
out:
if (rc >= 0) {
cfs_spin_lock(&sbi->ll_lock);
- lru->ccl_page_max = pages_number;
+ cache->ccc_lru_max = pages_number;
cfs_spin_unlock(&sbi->ll_lock);
rc = count;
} else {
- cfs_atomic_add(nrpages, &lru->ccl_page_left);
+ cfs_atomic_add(nrpages, &cache->ccc_lru_left);
}
return rc;
}