X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=blobdiff_plain;f=lustre%2Finclude%2Fliblustre.h;h=67f0066a63931cbcf4bb8e49fadc6646b8b0daff;hp=da6cc8a41565eb24bc65f2261728c3e2aa1dd436;hb=689ea74b583aff80c50210e90548379f8f22c201;hpb=2dc9c16e770415d56839e1996015fec5fab93f29 diff --git a/lustre/include/liblustre.h b/lustre/include/liblustre.h index da6cc8a..67f0066 100644 --- a/lustre/include/liblustre.h +++ b/lustre/include/liblustre.h @@ -24,40 +24,73 @@ #ifndef LIBLUSTRE_H__ #define LIBLUSTRE_H__ -#include -#include -#ifndef __CYGWIN__ -#include -#include +#ifdef __KERNEL__ +#error Kernel files should not #include #else -#include -#include "ioctl.h" +/* + * The userspace implementations of linux/spinlock.h vary; we just + * include our own for all of them + */ +#define __LINUX_SPINLOCK_H +#endif + +#include +#ifdef HAVE_STDINT_H +# include +#endif +#ifdef HAVE_ASM_PAGE_H +# include +#endif +#ifdef HAVE_SYS_USER_H +# include +#endif +#ifdef HAVE_SYS_IOCTL_H +# include #endif + #include #include #include #include #include #include -#include +#ifdef HAVE_SYS_VFS_H +# include +#endif +#include +#include +#include -#include -#include -#include +#include +#include +#include +#include + +#ifndef _IOWR +# include "ioctl.h" +#endif /* definitions for liblustre */ #ifdef __CYGWIN__ -#define PAGE_SHIFT 12 -#define PAGE_SIZE (1UL << PAGE_SHIFT) -#define PAGE_MASK (~(PAGE_SIZE-1)) +#define CFS_PAGE_SHIFT 12 +#define CFS_PAGE_SIZE (1UL << CFS_PAGE_SHIFT) +#define CFS_PAGE_MASK (~((__u64)CFS_PAGE_SIZE-1)) #define loff_t long long #define ERESTART 2001 typedef unsigned short umode_t; #endif +#ifndef CURRENT_SECONDS +# define CURRENT_SECONDS time(0) +#endif + +#ifndef ARRAY_SIZE +#define ARRAY_SIZE(a) ((sizeof (a))/(sizeof ((a)[0]))) +#endif + /* This is because lprocfs_status.h gets included here indirectly. It would * be much better to just avoid lprocfs being included into liblustre entirely * but that requires more header surgery than I can handle right now. @@ -65,18 +98,28 @@ typedef unsigned short umode_t; #ifndef smp_processor_id #define smp_processor_id() 0 #endif +#ifndef num_online_cpus +#define num_online_cpus() 1 +#endif +#ifndef num_possible_cpus +#define num_possible_cpus() 1 +#endif /* always adopt 2.5 definitions */ #define KERNEL_VERSION(a,b,c) ((a)*100+(b)*10+c) -#define LINUX_VERSION_CODE (2*200+5*10+0) +#define LINUX_VERSION_CODE KERNEL_VERSION(2,6,5) + +#ifndef page_private +#define page_private(page) ((page)->private) +#define set_page_private(page, v) ((page)->private = (v)) +#endif + static inline void inter_module_put(void *a) { return; } -extern ptl_handle_ni_t tcpnal_ni; - void *inter_module_get(char *arg); /* cheats for now */ @@ -112,14 +155,10 @@ static inline void *kmalloc(int size, int prot) #define GFP_HIGHUSER 1 #define GFP_ATOMIC 1 #define GFP_NOFS 1 -#define IS_ERR(a) (((a) && abs((long)(a)) < 500) ? 1 : 0) +#define IS_ERR(a) ((unsigned long)(a) > (unsigned long)-1000L) #define PTR_ERR(a) ((long)(a)) #define ERR_PTR(a) ((void*)((long)(a))) -typedef struct { - void *cwd; -}mm_segment_t; - typedef int (read_proc_t)(char *page, char **start, off_t off, int count, int *eof, void *data); @@ -127,13 +166,6 @@ struct file; /* forward ref */ typedef int (write_proc_t)(struct file *file, const char *buffer, unsigned long count, void *data); -# define le16_to_cpu(x) __le16_to_cpu(x) -# define cpu_to_le16(x) __cpu_to_le16(x) -# define le32_to_cpu(x) __le32_to_cpu(x) -# define cpu_to_le32(x) __cpu_to_le32(x) -# define le64_to_cpu(x) __le64_to_cpu(x) -# define cpu_to_le64(x) __cpu_to_le64(x) - #define NIPQUAD(addr) \ ((unsigned char *)&addr)[0], \ ((unsigned char *)&addr)[1], \ @@ -153,50 +185,24 @@ typedef int (write_proc_t)(struct file *file, const char *buffer, #endif /* __LITTLE_ENDIAN */ /* bits ops */ -static __inline__ int set_bit(int nr,long * addr) -{ - int mask, retval; - - addr += nr >> 5; - mask = 1 << (nr & 0x1f); - retval = (mask & *addr) != 0; - *addr |= mask; - return retval; -} - -static __inline__ int clear_bit(int nr, long * addr) -{ - int mask, retval; - - addr += nr >> 5; - mask = 1 << (nr & 0x1f); - retval = (mask & *addr) != 0; - *addr &= ~mask; - return retval; -} -static __inline__ int test_bit(int nr, long * addr) -{ - int mask; - - addr += nr >> 5; - mask = 1 << (nr & 0x1f); - return ((mask & *addr) != 0); -} +/* a long can be more than 32 bits, so use BITS_PER_LONG + * to allow the compiler to adjust the bit shifting accordingly + */ static __inline__ int ext2_set_bit(int nr, void *addr) { - return set_bit(nr, (long*)addr); + return set_bit(nr, addr); } static __inline__ int ext2_clear_bit(int nr, void *addr) { - return clear_bit(nr, (long*)addr); + return clear_bit(nr, addr); } static __inline__ int ext2_test_bit(int nr, void *addr) { - return test_bit(nr, (long*)addr); + return test_bit(nr, addr); } /* modules */ @@ -257,6 +263,8 @@ extern int ldlm_init(void); extern int osc_init(void); extern int lov_init(void); extern int mdc_init(void); +extern int lmv_init(void); +extern int mgc_init(void); extern int echo_client_init(void); @@ -265,11 +273,14 @@ extern int echo_client_init(void); #define EXPORT_SYMBOL(S) +struct rcu_head { }; + typedef struct { } spinlock_t; typedef __u64 kdev_t; #define SPIN_LOCK_UNLOCKED (spinlock_t) { } #define LASSERT_SPIN_LOCKED(lock) do {} while(0) +#define LASSERT_SEM_LOCKED(sem) do {} while(0) static inline void spin_lock(spinlock_t *l) {return;} static inline void spin_unlock(spinlock_t *l) {return;} @@ -283,21 +294,32 @@ static inline void spin_unlock_bh(spinlock_t *l) {} static inline void spin_lock_irqsave(spinlock_t *a, unsigned long b) {} static inline void spin_unlock_irqrestore(spinlock_t *a, unsigned long b) {} +typedef spinlock_t rwlock_t; +#define RW_LOCK_UNLOCKED SPIN_LOCK_UNLOCKED +#define read_lock(l) spin_lock(l) +#define read_unlock(l) spin_unlock(l) +#define write_lock(l) spin_lock(l) +#define write_unlock(l) spin_unlock(l) + + #define min(x,y) ((x)<(y) ? (x) : (y)) #define max(x,y) ((x)>(y) ? (x) : (y)) #ifndef min_t #define min_t(type,x,y) \ - ({ type __x = (x); type __y = (y); __x < __y ? __x: __y; }) + ({ type __x = (x); type __y = (y); __x < __y ? __x: __y; }) #endif #ifndef max_t #define max_t(type,x,y) \ - ({ type __x = (x); type __y = (y); __x > __y ? __x: __y; }) + ({ type __x = (x); type __y = (y); __x > __y ? __x: __y; }) #endif -/* registering symbols */ +#define simple_strtol strtol +/* registering symbols */ +#ifndef ERESTARTSYS #define ERESTARTSYS ERESTART +#endif #define HZ 1 /* random */ @@ -306,8 +328,8 @@ void get_random_bytes(void *ptr, int size); /* memory */ -/* FIXME */ -#define num_physpages (16 * 1024) +/* memory size: used for some client tunables */ +#define num_physpages (256 * 1024) /* 1GB */ static inline int copy_from_user(void *a,void *b, int c) { @@ -349,23 +371,8 @@ static inline int kmem_cache_destroy(kmem_cache_t *a) free(a); return 0; } -#define kmem_cache_alloc(cache, prio) malloc(cache->size) -#define kmem_cache_free(cache, obj) free(obj) - -#define PAGE_CACHE_SIZE PAGE_SIZE -#define PAGE_CACHE_SHIFT 12 -#define PAGE_CACHE_MASK PAGE_MASK - -struct page { - void *addr; - unsigned long index; - struct list_head list; - unsigned long private; - /* internally used by liblustre file i/o */ - int _offset; - int _count; -}; +/* struct page decl moved out from here into portals/include/libcfs/user-prim.h */ /* 2.4 defines */ #define PAGE_LIST_ENTRY list @@ -374,16 +381,16 @@ struct page { #define kmap(page) (page)->addr #define kunmap(a) do {} while (0) -static inline struct page *alloc_pages(int mask, unsigned long order) +static inline cfs_page_t *alloc_pages(int mask, unsigned long order) { - struct page *pg = malloc(sizeof(*pg)); + cfs_page_t *pg = malloc(sizeof(*pg)); if (!pg) return NULL; #if 0 //#ifdef MAP_ANONYMOUS pg->addr = mmap(0, PAGE_SIZE << order, PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, 0, 0); #else - pg->addr = malloc(PAGE_SIZE << order); + pg->addr = malloc(CFS_PAGE_SIZE << order); #endif if (!pg->addr) { @@ -392,10 +399,12 @@ static inline struct page *alloc_pages(int mask, unsigned long order) } return pg; } +#define cfs_alloc_pages(mask, order) alloc_pages((mask), (order)) -#define alloc_page(mask) alloc_pages((mask), 0) +#define alloc_page(mask) alloc_pages((mask), 0) +#define cfs_alloc_page(mask) alloc_page(mask) -static inline void __free_pages(struct page *pg, int what) +static inline void __free_pages(cfs_page_t *pg, int what) { #if 0 //#ifdef MAP_ANONYMOUS munmap(pg->addr, PAGE_SIZE); @@ -404,13 +413,15 @@ static inline void __free_pages(struct page *pg, int what) #endif free(pg); } +#define __cfs_free_pages(pg, order) __free_pages((pg), (order)) #define __free_page(page) __free_pages((page), 0) #define free_page(page) __free_page(page) +#define __cfs_free_page(page) __cfs_free_pages((page), 0) -static inline struct page* __grab_cache_page(unsigned long index) +static inline cfs_page_t* __grab_cache_page(unsigned long index) { - struct page *pg = alloc_pages(0, 0); + cfs_page_t *pg = alloc_pages(0, 0); if (pg) pg->index = index; @@ -444,6 +455,9 @@ static inline struct page* __grab_cache_page(unsigned long index) #define ATTR_RAW 0x0800 /* file system, not vfs will massage attrs */ #define ATTR_FROM_OPEN 0x1000 /* called from open path, ie O_TRUNC */ #define ATTR_CTIME_SET 0x2000 +#define ATTR_BLOCKS 0x4000 +#define ATTR_KILL_SUID 0 +#define ATTR_KILL_SGID 0 struct iattr { unsigned int ia_valid; @@ -457,6 +471,8 @@ struct iattr { unsigned int ia_attr_flags; }; +#define ll_iattr iattr + #define IT_OPEN 0x0001 #define IT_CREAT 0x0002 #define IT_READDIR 0x0004 @@ -499,7 +515,6 @@ static inline void intent_init(struct lookup_intent *it, int op, int flags) it->it_flags = flags; } - struct dentry { int d_count; }; @@ -520,7 +535,9 @@ struct semaphore { /* use the macro's argument to avoid unused warnings */ #define down(a) do { (void)a; } while (0) +#define mutex_down(a) down(a) #define up(a) do { (void)a; } while (0) +#define mutex_up(a) up(a) #define down_read(a) do { (void)a; } while (0) #define up_read(a) do { (void)a; } while (0) #define down_write(a) do { (void)a; } while (0) @@ -533,7 +550,12 @@ static inline void init_MUTEX (struct semaphore *sem) { sema_init(sem, 1); } +static inline void init_MUTEX_LOCKED (struct semaphore *sem) +{ + sema_init(sem, 0); +} +#define init_mutex(s) init_MUTEX(s) typedef struct { struct list_head sleepers; @@ -548,15 +570,12 @@ struct signal { int signal; }; -struct fs_struct { - int umask; -}; - struct task_struct { - struct fs_struct *fs; int state; struct signal pending; char comm[32]; + int uid; + int gid; int pid; int fsuid; int fsgid; @@ -564,10 +583,13 @@ struct task_struct { int ngroups; gid_t *groups; __u32 cap_effective; - - struct fs_struct __fs; }; +typedef struct task_struct cfs_task_t; +#define cfs_current() current +#define cfs_curproc_pid() (current->pid) +#define cfs_curproc_comm() (current->comm) + extern struct task_struct *current; int in_group_p(gid_t gid); static inline int capable(int cap) @@ -616,6 +638,7 @@ static inline int schedule_timeout(signed long t) } #define lock_kernel() do {} while (0) +#define unlock_kernel() do {} while (0) #define daemonize(l) do {} while (0) #define sigfillset(l) do {} while (0) #define recalc_sigpending(l) do {} while (0) @@ -637,8 +660,10 @@ static inline int schedule_timeout(signed long t) _ret = tv.tv_sec; \ _ret; \ }) -#define time_after(a, b) ((long)(b) - (long)(a) > 0) +#define get_jiffies_64() (__u64)jiffies +#define time_after(a, b) ((long)(b) - (long)(a) < 0) #define time_before(a, b) time_after(b,a) +#define time_after_eq(a,b) ((long)(a) - (long)(b) >= 0) struct timer_list { struct list_head tl_list; @@ -649,7 +674,7 @@ struct timer_list { static inline int timer_pending(struct timer_list *l) { - if (l->expires > jiffies) + if (time_after(l->expires, jiffies)) return 1; else return 0; @@ -673,13 +698,20 @@ static inline void del_timer(struct timer_list *l) typedef struct { volatile int counter; } atomic_t; +#define ATOMIC_INIT(i) { (i) } + #define atomic_read(a) ((a)->counter) #define atomic_set(a,b) do {(a)->counter = b; } while (0) #define atomic_dec_and_test(a) ((--((a)->counter)) == 0) +#define atomic_dec_and_lock(a,b) ((--((a)->counter)) == 0) #define atomic_inc(a) (((a)->counter)++) #define atomic_dec(a) do { (a)->counter--; } while (0) #define atomic_add(b,a) do {(a)->counter += b;} while (0) +#define atomic_add_return(n,a) ((a)->counter += n) +#define atomic_inc_return(a) atomic_add_return(1,a) #define atomic_sub(b,a) do {(a)->counter -= b;} while (0) +#define atomic_sub_return(n,a) ((a)->counter -= n) +#define atomic_dec_return(a) atomic_sub_return(1,a) #ifndef likely #define likely(exp) (exp) @@ -688,9 +720,43 @@ typedef struct { volatile int counter; } atomic_t; #define unlikely(exp) (exp) #endif +#define might_sleep() +#define might_sleep_if(c) +#define smp_mb() + +static inline +int test_and_set_bit(int nr, unsigned long *addr) +{ + int oldbit; + + while (nr >= sizeof(long)) { + nr -= sizeof(long); + addr++; + } + + oldbit = (*addr) & (1 << nr); + *addr |= (1 << nr); + return oldbit; +} + +static inline +int test_and_clear_bit(int nr, unsigned long *addr) +{ + int oldbit; + + while (nr >= sizeof(long)) { + nr -= sizeof(long); + addr++; + } + + oldbit = (*addr) & (1 << nr); + *addr &= ~(1 << nr); + return oldbit; +} + /* FIXME sys/capability will finally included linux/fs.h thus * cause numerous trouble on x86-64. as temporary solution for - * build broken at cary, we copy definition we need from capability.h + * build broken at Cray, we copy definition we need from capability.h * FIXME */ struct _cap_struct; @@ -706,33 +772,22 @@ typedef enum { CAP_SET=1 } cap_flag_value_t; -#define CAP_FOWNER 3 -#define CAP_FSETID 4 -#define CAP_SYS_ADMIN 21 +#define CAP_DAC_OVERRIDE 1 +#define CAP_DAC_READ_SEARCH 2 +#define CAP_FOWNER 3 +#define CAP_FSETID 4 +#define CAP_SYS_ADMIN 21 cap_t cap_get_proc(void); int cap_get_flag(cap_t, cap_value_t, cap_flag_t, cap_flag_value_t *); - - -/* log related */ -static inline int llog_init_commit_master(void) { return 0; } -static inline int llog_cleanup_commit_master(int force) { return 0; } -static inline void portals_run_lbug_upcall(char *file, const char *fn, +static inline void libcfs_run_lbug_upcall(char *file, const char *fn, const int l){} -#define LBUG() \ - do { \ - printf("!!!LBUG at %s:%d\n", __FILE__, __LINE__); \ - sleep(1000000); \ - } while (0) - - - /* completion */ struct completion { unsigned int done; - wait_queue_head_t wait; + cfs_waitq_t wait; }; #define COMPLETION_INITIALIZER(work) \ @@ -751,20 +806,215 @@ static inline void init_completion(struct completion *x) struct liblustre_wait_callback { struct list_head llwc_list; + const char *llwc_name; int (*llwc_fn)(void *arg); void *llwc_arg; }; -void *liblustre_register_wait_callback(int (*fn)(void *arg), void *arg); +void *liblustre_register_wait_callback(const char *name, + int (*fn)(void *arg), void *arg); void liblustre_deregister_wait_callback(void *notifier); int liblustre_wait_event(int timeout); -#include -#include -#include -#include -#include -#include +void *liblustre_register_idle_callback(const char *name, + int (*fn)(void *arg), void *arg); +void liblustre_deregister_idle_callback(void *notifier); +void liblustre_wait_idle(void); + +/* flock related */ +struct nfs_lock_info { + __u32 state; + __u32 flags; + void *host; +}; + +typedef struct file_lock { + struct file_lock *fl_next; /* singly linked list for this inode */ + struct list_head fl_link; /* doubly linked list of all locks */ + struct list_head fl_block; /* circular list of blocked processes */ + void *fl_owner; + unsigned int fl_pid; + cfs_waitq_t fl_wait; + struct file *fl_file; + unsigned char fl_flags; + unsigned char fl_type; + loff_t fl_start; + loff_t fl_end; + + void (*fl_notify)(struct file_lock *); /* unblock callback */ + void (*fl_insert)(struct file_lock *); /* lock insertion callback */ + void (*fl_remove)(struct file_lock *); /* lock removal callback */ + + void *fl_fasync; /* for lease break notifications */ + unsigned long fl_break_time; /* for nonblocking lease breaks */ + + union { + struct nfs_lock_info nfs_fl; + } fl_u; +} cfs_flock_t; + +#define cfs_flock_type(fl) ((fl)->fl_type) +#define cfs_flock_set_type(fl, type) do { (fl)->fl_type = (type); } while(0) +#define cfs_flock_pid(fl) ((fl)->fl_pid) +#define cfs_flock_set_pid(fl, pid) do { (fl)->fl_pid = (pid); } while(0) +#define cfs_flock_start(fl) ((fl)->fl_start) +#define cfs_flock_set_start(fl, start) do { (fl)->fl_start = (start); } while(0) +#define cfs_flock_end(fl) ((fl)->fl_end) +#define cfs_flock_set_end(fl, end) do { (fl)->fl_end = (end); } while(0) + +#ifndef OFFSET_MAX +#define INT_LIMIT(x) (~((x)1 << (sizeof(x)*8 - 1))) +#define OFFSET_MAX INT_LIMIT(loff_t) +#endif + +/* XXX: defined in kernel */ +#define FL_POSIX 1 +#define FL_SLEEP 128 + +/* quota */ +#define QUOTA_OK 0 +#define NO_QUOTA 1 + +/* ACL */ +struct posix_acl_entry { + short e_tag; + unsigned short e_perm; + unsigned int e_id; +}; + +struct posix_acl { + atomic_t a_refcount; + unsigned int a_count; + struct posix_acl_entry a_entries[0]; +}; + +typedef struct { + __u16 e_tag; + __u16 e_perm; + __u32 e_id; +} xattr_acl_entry; + +typedef struct { + __u32 a_version; + xattr_acl_entry a_entries[0]; +} xattr_acl_header; + +static inline size_t xattr_acl_size(int count) +{ + return sizeof(xattr_acl_header) + count * sizeof(xattr_acl_entry); +} + +static inline +struct posix_acl * posix_acl_from_xattr(const void *value, size_t size) +{ + return NULL; +} + +static inline +int posix_acl_valid(const struct posix_acl *acl) +{ + return 0; +} + +static inline +void posix_acl_release(struct posix_acl *acl) +{ +} + +#ifdef LIBLUSTRE_POSIX_ACL +# ifndef posix_acl_xattr_entry +# define posix_acl_xattr_entry xattr_acl_entry +# endif +# ifndef posix_acl_xattr_header +# define posix_acl_xattr_header xattr_acl_header +# endif +# ifndef posix_acl_xattr_size +# define posix_acl_xattr_size(entry) xattr_acl_size(entry) +# endif +# ifndef CONFIG_FS_POSIX_ACL +# define CONFIG_FS_POSIX_ACL 1 +# endif +#endif + +#ifndef ENOTSUPP +#define ENOTSUPP ENOTSUP +#endif + +typedef int mm_segment_t; +enum { + KERNEL_DS, + USER_DS +}; +static inline mm_segment_t get_fs(void) +{ + return USER_DS; +} +static inline void set_fs(mm_segment_t seg) +{ +} + +#include +#include +#include +#include +#include +#include + +/* Fast hashing routine for a long. + (C) 2002 William Lee Irwin III, IBM */ + +/* + * Knuth recommends primes in approximately golden ratio to the maximum + * integer representable by a machine word for multiplicative hashing. + * Chuck Lever verified the effectiveness of this technique: + * http://www.citi.umich.edu/techreports/reports/citi-tr-00-1.pdf + * + * These primes are chosen to be bit-sparse, that is operations on + * them can use shifts and additions instead of multiplications for + * machines where multiplications are slow. + */ +#if BITS_PER_LONG == 32 +/* 2^31 + 2^29 - 2^25 + 2^22 - 2^19 - 2^16 + 1 */ +#define GOLDEN_RATIO_PRIME 0x9e370001UL +#elif BITS_PER_LONG == 64 +/* 2^63 + 2^61 - 2^57 + 2^54 - 2^51 - 2^18 + 1 */ +#define GOLDEN_RATIO_PRIME 0x9e37fffffffc0001UL +#else +#error Define GOLDEN_RATIO_PRIME for your wordsize. +#endif + +static inline unsigned long hash_long(unsigned long val, unsigned int bits) +{ + unsigned long hash = val; + +#if BITS_PER_LONG == 64 + /* Sigh, gcc can't optimise this alone like it does for 32 bits. */ + unsigned long n = hash; + n <<= 18; + hash -= n; + n <<= 33; + hash -= n; + n <<= 3; + hash += n; + n <<= 3; + hash -= n; + n <<= 4; + hash += n; + n <<= 2; + hash += n; +#else + /* On some cpus multiply is faster, on others gcc will do shifts */ + hash *= GOLDEN_RATIO_PRIME; +#endif + + /* High bits are more random, so use them. */ + return hash >> (BITS_PER_LONG - bits); +} + +static inline unsigned long hash_ptr(void *ptr, unsigned int bits) +{ + return hash_long((unsigned long)ptr, bits); +} #endif