X-Git-Url: https://git.whamcloud.com/?a=blobdiff_plain;f=lustre%2Finclude%2Flinux%2Flustre_compat25.h;h=db19c012f663e723fa69c03c3c237203fad4c4d4;hb=a8a7d2f30f7cb151741690ef62cb5ea3a7892468;hp=40620ac5eed8edc39969158e454208d77ae87363;hpb=efc75006b820c9b8e186c7650fc24edd5d0509af;p=fs%2Flustre-release.git diff --git a/lustre/include/linux/lustre_compat25.h b/lustre/include/linux/lustre_compat25.h index 40620ac..db19c01 100644 --- a/lustre/include/linux/lustre_compat25.h +++ b/lustre/include/linux/lustre_compat25.h @@ -31,6 +31,72 @@ #include +/* + * groups_info related staff + */ +#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,4) + +#define NGROUPS_SMALL NGROUPS +#define NGROUPS_PER_BLOCK ((int)(EXEC_PAGESIZE / sizeof(gid_t))) +struct group_info { + int ngroups; + atomic_t usage; + gid_t small_block[NGROUPS_SMALL]; + int nblocks; + gid_t *blocks[0]; +}; +#define current_ngroups current->ngroups + +struct group_info *groups_alloc(int gidsetsize); +void groups_free(struct group_info *ginfo); +int groups_search(struct group_info *ginfo, gid_t grp); + +#define get_group_info(group_info) \ + do { \ + atomic_inc(&(group_info)->usage); \ + } while (0) + +#define put_group_info(group_info) \ + do { \ + if (atomic_dec_and_test(&(group_info)->usage)) \ + groups_free(group_info); \ + } while (0) + +#define groups_sort(gi) do {} while (0) + +#define GROUP_AT(gi, i) ((gi)->small_block[(i)]) + +static inline int cleanup_group_info(void) +{ + /* Get rid of unneeded supplementary groups */ + current->ngroups = 0; + memset(current->groups, 0, sizeof(current->groups)); + return 0; +} + +#else /* >= 2.6.4 */ + +#define current_ngroups current->group_info->ngroups + +void groups_sort(struct group_info *ginfo); +int groups_search(struct group_info *ginfo, gid_t grp); + +static inline int cleanup_group_info(void) +{ + struct group_info *ginfo; + + ginfo = groups_alloc(0); + if (!ginfo) + return -ENOMEM; + + set_current_groups(ginfo); + put_group_info(ginfo); + + return 0; +} +#endif /* end of groups_info stuff */ + + #if LINUX_VERSION_CODE >= KERNEL_VERSION(2,5,0) /* @@ -38,8 +104,10 @@ * initialization routines must be called after device * driver initialization */ +#ifndef MODULE #undef module_init #define module_init(a) late_initcall(a) +#endif /* XXX our code should be using the 2.6 calls, not the other way around */ #define TryLockPage(page) TestSetPageLocked(page) @@ -50,10 +118,16 @@ #define LTIME_S(time) (time.tv_sec) #define ll_path_lookup path_lookup -#define ll_permission permission +#define ll_permission(inode,mask,nd) permission(inode,mask,nd) #define ll_pgcache_lock(mapping) spin_lock(&mapping->page_lock) #define ll_pgcache_unlock(mapping) spin_unlock(&mapping->page_lock) +#define ll_call_writepage(inode, page) \ + (inode)->i_mapping->a_ops->writepage(page, NULL) +#define ll_invalidate_inode_pages(inode) \ + invalidate_inode_pages((inode)->i_mapping) +#define ll_truncate_complete_page(page) \ + truncate_complete_page(page->mapping, page) #define ll_vfs_create(a,b,c,d) vfs_create(a,b,c,d) @@ -69,31 +143,35 @@ static inline void lustre_daemonize_helper(void) { LASSERT(current->signal != NULL); - current->session = 1; + current->signal->session = 1; if (current->group_leader) - current->group_leader->__pgrp = 1; + current->group_leader->signal->pgrp = 1; else CERROR("we aren't group leader\n"); - current->tty = NULL; + current->signal->tty = NULL; } -#define rb_node_s rb_node -#define rb_root_s rb_root -typedef struct rb_root_s rb_root_t; -typedef struct rb_node_s rb_node_t; - -#define smp_num_cpus NR_CPUS +#define __set_page_ll_data(page, llap) \ + do { \ + page_cache_get(page); \ + SetPagePrivate(page); \ + page->private = (unsigned long)llap; \ + } while (0) +#define __clear_page_ll_data(page) \ + do { \ + ClearPagePrivate(page); \ + page_cache_release(page); \ + page->private = 0; \ + } while(0) -#ifndef conditional_schedule -#define conditional_schedule() cond_resched() -#endif +#define kiobuf bio #include #else /* 2.4.. */ #define ll_vfs_create(a,b,c,d) vfs_create(a,b,c) -#define ll_permission(a,b,c) permission(a,b) +#define ll_permission(inode,mask,nd) permission(inode,mask) #define ILOOKUP(sb, ino, test, data) ilookup4(sb, ino, test, data); #define DCACHE_DISCONNECTED DCACHE_NFSD_DISCONNECTED #define ll_dev_t int @@ -126,19 +204,23 @@ static inline void clear_page_dirty(struct page *page) #define cpu_online(cpu) (cpu_online_map & (1<i_mapping->a_ops->writepage(page) +#define ll_invalidate_inode_pages(inode) invalidate_inode_pages(inode) +#define ll_truncate_complete_page(page) truncate_complete_page(page) static inline void __d_drop(struct dentry *dentry) { @@ -153,16 +235,107 @@ static inline void lustre_daemonize_helper(void) current->tty = NULL; } -#ifndef conditional_schedule -#define conditional_schedule() if (unlikely(need_resched())) schedule() +#ifndef HAVE_COND_RESCHED +static inline void cond_resched(void) +{ + if (unlikely(need_resched())) { + set_current_state(TASK_RUNNING); + schedule(); + } +} #endif +static inline int mapping_mapped(struct address_space *mapping) +{ + return mapping->i_mmap_shared ? 1 : 0; +} + /* to find proc_dir_entry from inode. 2.6 has native one -bzzz */ -#if LINUX_VERSION_CODE < KERNEL_VERSION(2,4,23) +#ifndef HAVE_PDE #define PDE(ii) ((ii)->u.generic_ip) #endif +#define __set_page_ll_data(page, llap) page->private = (unsigned long)llap +#define __clear_page_ll_data(page) page->private = 0 +#define PageWriteback(page) 0 +#define end_page_writeback(page) + +#ifdef ZAP_PAGE_RANGE_VMA +#define ll_zap_page_range(vma, addr, len) zap_page_range(vma, addr, len) +#else +#define ll_zap_page_range(vma, addr, len) zap_page_range(vma->vm_mm, addr, len) +#endif + #endif /* end of 2.4 compat macros */ +#ifdef HAVE_PAGE_LIST +static inline int mapping_has_pages(struct address_space *mapping) +{ + int rc = 1; + + ll_pgcache_lock(mapping); + if (list_empty(&mapping->dirty_pages) && + list_empty(&mapping->clean_pages) && + list_empty(&mapping->locked_pages)) { + rc = 0; + } + ll_pgcache_unlock(mapping); + + return rc; +} + +static inline int clear_page_dirty_for_io(struct page *page) +{ + struct address_space *mapping = page->mapping; + + if (page->mapping && PageDirty(page)) { + ClearPageDirty(page); + ll_pgcache_lock(mapping); + list_del(&page->list); + list_add(&page->list, &mapping->locked_pages); + ll_pgcache_unlock(mapping); + return 1; + } + return 0; +} +#else +static inline int mapping_has_pages(struct address_space *mapping) +{ + return mapping->nrpages > 0; +} +#endif + +#if (LINUX_VERSION_CODE < KERNEL_VERSION(2,6,7)) +#define ll_set_dflags(dentry, flags) do { dentry->d_vfs_flags |= flags; } while(0) +#define ll_vfs_symlink(dir, dentry, path, mode) vfs_symlink(dir, dentry, path) +#else +#define ll_set_dflags(dentry, flags) do { \ + spin_lock(&dentry->d_lock); \ + dentry->d_flags |= flags; \ + spin_unlock(&dentry->d_lock); \ + } while(0) +#define ll_vfs_symlink(dir, dentry, path, mode) vfs_symlink(dir, dentry, path, mode) +#endif + +#ifdef HAVE_I_ALLOC_SEM +#define UP_WRITE_I_ALLOC_SEM(i) do { up_write(&(i)->i_alloc_sem); } while (0) +#define DOWN_WRITE_I_ALLOC_SEM(i) do { down_write(&(i)->i_alloc_sem); } while(0) +#define LASSERT_MDS_ORPHAN_WRITE_LOCKED(i) LASSERT(down_read_trylock(&(i)->i_alloc_sem) == 0) + +#define UP_READ_I_ALLOC_SEM(i) do { up_read(&(i)->i_alloc_sem); } while (0) +#define DOWN_READ_I_ALLOC_SEM(i) do { down_read(&(i)->i_alloc_sem); } while (0) +#define LASSERT_MDS_ORPHAN_READ_LOCKED(i) LASSERT(down_write_trylock(&(i)->i_alloc_sem) == 0) +#define MDS_PACK_MD_LOCK 1 +#else +#define UP_READ_I_ALLOC_SEM(i) do { up(&(i)->i_sem); } while (0) +#define DOWN_READ_I_ALLOC_SEM(i) do { down(&(i)->i_sem); } while (0) +#define LASSERT_MDS_ORPHAN_READ_LOCKED(i) LASSERT(down_trylock(&(i)->i_sem) != 0) + +#define UP_WRITE_I_ALLOC_SEM(i) do { up(&(i)->i_sem); } while (0) +#define DOWN_WRITE_I_ALLOC_SEM(i) do { down(&(i)->i_sem); } while (0) +#define LASSERT_MDS_ORPHAN_WRITE_LOCKED(i) LASSERT(down_trylock(&(i)->i_sem) != 0) +#define MDS_PACK_MD_LOCK 0 +#endif + #endif /* __KERNEL__ */ #endif /* _COMPAT25_H */