*/
/* returns the page unlocked, but with a reference */
-static int ll_dir_readpage(struct file *file, struct page *page0)
+static int ll_dir_filler(void *_hash, struct page *page0)
{
struct inode *inode = page0->mapping->host;
int hash64 = ll_i2sbi(inode)->ll_flags & LL_SBI_64BIT_HASH;
struct ptlrpc_request *request;
struct mdt_body *body;
struct md_op_data *op_data;
- __u64 hash;
+ __u64 hash = *((__u64 *)_hash);
struct page **page_pool;
struct page *page;
#ifndef HAVE_ADD_TO_PAGE_CACHE_LRU
int rc;
ENTRY;
- if (file) {
- struct ll_file_data *fd = LUSTRE_FPRIVATE(file);
-
- hash = fd->fd_dir.lfd_next;
- } else {
- hash = ll_i2info(inode)->lli_sa_pos;
- }
CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p) hash "LPU64"\n",
inode->i_ino, inode->i_generation, inode, hash);
return rc;
}
-#ifndef MS_HAS_NEW_AOPS
-struct address_space_operations ll_dir_aops = {
- .readpage = ll_dir_readpage,
-};
-#else
-struct address_space_operations_ext ll_dir_aops = {
- .orig_aops.readpage = ll_dir_readpage,
-};
-#endif
-
static void ll_check_page(struct inode *dir, struct page *page)
{
/* XXX: check page format later */
* hence, can avoid restart.
*
* In fact, page cannot be locked here at all, because
- * ll_dir_readpage() does synchronous io.
+ * ll_dir_filler() does synchronous io.
*/
wait_on_page(page);
if (PageUptodate(page)) {
return page;
}
-struct page *ll_get_dir_page(struct file *filp, struct inode *dir, __u64 hash,
+struct page *ll_get_dir_page(struct inode *dir, __u64 hash,
struct ll_dir_chain *chain)
{
ldlm_policy_data_t policy = {.l_inodebits = {MDS_INODELOCK_UPDATE} };
}
page = read_cache_page(mapping, hash_x_index(hash, hash64),
- (filler_t*)mapping->a_ops->readpage, filp);
+ ll_dir_filler, &lhash);
if (IS_ERR(page)) {
CERROR("read cache page: "DFID" at "LPU64": rc %ld\n",
PFID(ll_inode2fid(dir)), hash, PTR_ERR(page));
goto out_unlock;
}
-int ll_readdir(struct file *filp, void *cookie, filldir_t filldir)
+int ll_dir_read(struct inode *inode, __u64 *_pos, void *cookie,
+ filldir_t filldir)
{
- struct inode *inode = filp->f_dentry->d_inode;
struct ll_inode_info *info = ll_i2info(inode);
struct ll_sb_info *sbi = ll_i2sbi(inode);
- struct ll_file_data *fd = LUSTRE_FPRIVATE(filp);
- __u64 pos = fd->fd_dir.lfd_pos;
+ __u64 pos = *_pos;
int api32 = ll_need_32bit_api(sbi);
int hash64 = sbi->ll_flags & LL_SBI_64BIT_HASH;
struct page *page;
struct ll_dir_chain chain;
- int done;
- int rc;
+ int done = 0;
+ int rc = 0;
ENTRY;
- CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p) pos %lu/%llu 32bit_api %d\n",
- inode->i_ino, inode->i_generation, inode,
- (unsigned long)pos, i_size_read(inode), api32);
-
- if (pos == MDS_DIR_END_OFF)
- /*
- * end-of-file.
- */
- GOTO(out, rc = 0);
-
- rc = 0;
- done = 0;
ll_dir_chain_init(&chain);
- fd->fd_dir.lfd_next = pos;
- page = ll_get_dir_page(filp, inode, pos, &chain);
+ page = ll_get_dir_page(inode, pos, &chain);
while (rc == 0 && !done) {
struct lu_dirpage *dp;
ll_release_page(page,
le32_to_cpu(dp->ldp_flags) &
LDF_COLLIDE);
- fd->fd_dir.lfd_next = pos;
- page = ll_get_dir_page(filp, inode, pos,
+ next = pos;
+ page = ll_get_dir_page(inode, pos,
&chain);
} else {
/*
}
}
- fd->fd_dir.lfd_pos = pos;
+ *_pos = pos;
+ ll_dir_chain_fini(&chain);
+ RETURN(rc);
+}
+
+static int ll_readdir(struct file *filp, void *cookie, filldir_t filldir)
+{
+ struct inode *inode = filp->f_dentry->d_inode;
+ struct ll_file_data *lfd = LUSTRE_FPRIVATE(filp);
+ struct ll_sb_info *sbi = ll_i2sbi(inode);
+ __u64 pos = lfd->lfd_pos;
+ int hash64 = sbi->ll_flags & LL_SBI_64BIT_HASH;
+ int api32 = ll_need_32bit_api(sbi);
+ int rc;
+ ENTRY;
+
+ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p) pos %lu/%llu "
+ " 32bit_api %d\n", inode->i_ino, inode->i_generation,
+ inode, (unsigned long)pos, i_size_read(inode), api32);
+
+ if (pos == MDS_DIR_END_OFF)
+ /*
+ * end-of-file.
+ */
+ GOTO(out, rc = 0);
+
+ rc = ll_dir_read(inode, &pos, cookie, filldir);
+ lfd->lfd_pos = pos;
if (pos == MDS_DIR_END_OFF) {
if (api32)
filp->f_pos = LL_DIR_END_OFF_32BIT;
filp->f_version = inode->i_version;
touch_atime(filp->f_vfsmnt, filp->f_dentry);
- ll_dir_chain_fini(&chain);
-
out:
if (!rc)
ll_stats_ops_tally(sbi, LPROC_LL_READDIR, 1);
if (offset != file->f_pos) {
if ((api32 && offset == LL_DIR_END_OFF_32BIT) ||
(!api32 && offset == LL_DIR_END_OFF))
- fd->fd_dir.lfd_pos = MDS_DIR_END_OFF;
+ fd->lfd_pos = MDS_DIR_END_OFF;
else if (api32 && sbi->ll_flags & LL_SBI_64BIT_HASH)
- fd->fd_dir.lfd_pos = offset << 32;
+ fd->lfd_pos = offset << 32;
else
- fd->fd_dir.lfd_pos = offset;
+ fd->lfd_pos = offset;
file->f_pos = offset;
file->f_version = 0;
}
* cleanup the dir readahead. */
void *d_opendir_key;
struct ll_statahead_info *d_sai;
- __u64 d_sa_pos;
struct posix_acl *d_def_acl;
/* protect statahead stuff. */
cfs_spinlock_t d_sa_lock;
#define lli_readdir_mutex u.d.d_readdir_mutex
#define lli_opendir_key u.d.d_opendir_key
#define lli_sai u.d.d_sai
-#define lli_sa_pos u.d.d_sa_pos
#define lli_def_acl u.d.d_def_acl
#define lli_sa_lock u.d.d_sa_lock
#define lli_opendir_pid u.d.d_opendir_pid
* clustred nfs */
struct rmtacl_ctl_table ll_rct;
struct eacl_table ll_et;
- struct vfsmount *ll_mnt;
};
#define LL_DEFAULT_MAX_RW_CHUNK (32 * 1024 * 1024)
unsigned long ras_consecutive_stride_requests;
};
-struct ll_file_dir {
- __u64 lfd_pos;
- __u64 lfd_next;
-};
-
extern cfs_mem_cache_t *ll_file_data_slab;
struct lustre_handle;
struct ll_file_data {
struct ll_readahead_state fd_ras;
int fd_omode;
struct ccc_grouplock fd_grouplock;
- struct ll_file_dir fd_dir;
+ __u64 lfd_pos;
__u32 fd_flags;
struct file *fd_file;
/* Indicate whether need to report failure when close.
void ll_release_page(struct page *page, int remove);
extern struct file_operations ll_dir_operations;
extern struct inode_operations ll_dir_inode_operations;
-struct page *ll_get_dir_page(struct file *filp, struct inode *dir, __u64 hash,
+struct page *ll_get_dir_page(struct inode *dir, __u64 hash,
struct ll_dir_chain *chain);
-int ll_readdir(struct file *filp, void *cookie, filldir_t filldir);
+int ll_dir_read(struct inode *inode, __u64 *_pos, void *cookie,
+ filldir_t filldir);
int ll_get_mdt_idx(struct inode *inode);
char *ll_get_fsname(struct inode *inode);
#ifndef MS_HAS_NEW_AOPS
extern struct address_space_operations ll_aops;
-extern struct address_space_operations ll_dir_aops;
#else
extern struct address_space_operations_ext ll_aops;
-extern struct address_space_operations_ext ll_dir_aops;
#endif
#ifndef log2
CDEBUG(D_SUPER, "rootfid "DFID"\n", PFID(&sbi->ll_root_fid));
sb->s_op = &lustre_super_operations;
-#if THREAD_SIZE >= 8192 /*b=17630*/ && !defined(HAVE_FSTYPE_MOUNT) /*LU-812*/
+#if THREAD_SIZE >= 8192 /*b=17630*/
sb->s_export_op = &lustre_export_operations;
#endif
uuid = obd_get_uuid(sbi->ll_md_exp);
if (uuid != NULL)
sb->s_dev = get_uuid2int(uuid->uuid, strlen(uuid->uuid));
- sbi->ll_mnt = mnt;
if (data != NULL)
OBD_FREE_PTR(data);
cfs_mutex_init(&lli->lli_readdir_mutex);
lli->lli_opendir_key = NULL;
lli->lli_sai = NULL;
- lli->lli_sa_pos = 0;
lli->lli_def_acl = NULL;
cfs_spin_lock_init(&lli->lli_sa_lock);
lli->lli_opendir_pid = 0;
} else if (S_ISDIR(inode->i_mode)) {
inode->i_op = &ll_dir_inode_operations;
inode->i_fop = &ll_dir_operations;
- inode->i_mapping->a_ops = (struct address_space_operations *)&ll_dir_aops;
EXIT;
} else if (S_ISLNK(inode->i_mode)) {
inode->i_op = &ll_fast_symlink_inode_operations;
cfs_spin_unlock(&plli->lli_sa_lock);
cfs_waitq_signal(&thread->t_ctl_waitq);
- plli->lli_sa_pos = 0;
ll_dir_chain_init(&chain);
- page = ll_get_dir_page(NULL, dir, pos, &chain);
+ page = ll_get_dir_page(dir, pos, &chain);
while (1) {
struct lu_dirpage *dp;
*/
ll_release_page(page, le32_to_cpu(dp->ldp_flags) &
LDF_COLLIDE);
- plli->lli_sa_pos = pos;
sai->sai_in_readpage = 1;
- page = ll_get_dir_page(NULL, dir, pos, &chain);
+ page = ll_get_dir_page(dir, pos, &chain);
sai->sai_in_readpage = 0;
} else {
LASSERT(le32_to_cpu(dp->ldp_flags) & LDF_COLLIDE);
static int is_first_dirent(struct inode *dir, struct dentry *dentry)
{
- struct ll_inode_info *lli = ll_i2info(dir);
struct ll_dir_chain chain;
struct qstr *target = &dentry->d_name;
struct page *page;
int rc = LS_NONE_FIRST_DE;
ENTRY;
- lli->lli_sa_pos = 0;
ll_dir_chain_init(&chain);
- page = ll_get_dir_page(NULL, dir, pos, &chain);
+ page = ll_get_dir_page(dir, pos, &chain);
while (1) {
struct lu_dirpage *dp;
*/
ll_release_page(page, le32_to_cpu(dp->ldp_flags) &
LDF_COLLIDE);
- lli->lli_sa_pos = pos;
- page = ll_get_dir_page(NULL, dir, pos, &chain);
+ page = ll_get_dir_page(dir, pos, &chain);
} else {
/*
* go into overflow page.