X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=blobdiff_plain;f=lnet%2Finclude%2Flnet%2Flib-lnet.h;h=f028224f9e34416ec19dfe263eea33773a81b6d3;hp=99ca4f14e8d554dcb444eea344cdfeedc91f9ca9;hb=c0598f15dd502deadc6e84f362331d1ed77b6d14;hpb=1dfe6a018e7104d74b5578a47cda7b40a01538c2 diff --git a/lnet/include/lnet/lib-lnet.h b/lnet/include/lnet/lib-lnet.h index 99ca4f1..f028224 100644 --- a/lnet/include/lnet/lib-lnet.h +++ b/lnet/include/lnet/lib-lnet.h @@ -23,7 +23,7 @@ * Copyright (c) 2003, 2010, Oracle and/or its affiliates. All rights reserved. * Use is subject to license terms. * - * Copyright (c) 2012, 2016, Intel Corporation. + * Copyright (c) 2012, 2017, Intel Corporation. */ /* * This file is part of Lustre, http://www.lustre.org/ @@ -37,9 +37,10 @@ #ifndef __LNET_LIB_LNET_H__ #define __LNET_LIB_LNET_H__ -#ifndef __KERNEL__ -# error This include is only for kernel use. -#endif +/* LNET has 0xeXXX */ +#define CFS_FAIL_PTLRPC_OST_BULK_CB2 0xe000 + +#include #include #include @@ -73,16 +74,42 @@ extern struct lnet the_lnet; /* THE network */ /** exclusive lock */ #define LNET_LOCK_EX CFS_PERCPT_LOCK_EX -static inline int lnet_is_route_alive(struct lnet_route *route) -{ - if (!route->lr_gateway->lpni_alive) - return 0; /* gateway is down */ - if ((route->lr_gateway->lpni_ping_feats & - LNET_PING_FEAT_NI_STATUS) == 0) - return 1; /* no NI status, assume it's alive */ - /* has NI status, check # down NIs */ - return route->lr_downis == 0; -} +/* default timeout */ +#define DEFAULT_PEER_TIMEOUT 180 +#define LNET_LND_DEFAULT_TIMEOUT 5 + +#ifdef HAVE_KERN_SOCK_GETNAME_2ARGS +#define lnet_kernel_getpeername(sock, addr, addrlen) \ + kernel_getpeername(sock, addr) +#define lnet_kernel_getsockname(sock, addr, addrlen) \ + kernel_getsockname(sock, addr) +#else +#define lnet_kernel_getpeername(sock, addr, addrlen) \ + kernel_getpeername(sock, addr, addrlen) +#define lnet_kernel_getsockname(sock, addr, addrlen) \ + kernel_getsockname(sock, addr, addrlen) +#endif + +/* + * kernel 5.3: commit ef11db3310e272d3d8dbe8739e0770820dd20e52 + * added in_dev_for_each_ifa_rtnl and in_dev_for_each_ifa_rcu + * and removed for_ifa and endfor_ifa. + * Use the _rntl variant as the current locking is rtnl. + */ +#ifdef in_dev_for_each_ifa_rtnl +#define DECLARE_CONST_IN_IFADDR(ifa) const struct in_ifaddr *ifa +#define endfor_ifa(in_dev) +#else +#define DECLARE_CONST_IN_IFADDR(ifa) +#define in_dev_for_each_ifa_rtnl(ifa, in_dev) for_ifa((in_dev)) +#define in_dev_for_each_ifa_rcu(ifa, in_dev) for_ifa((in_dev)) +#endif + +int choose_ipv4_src(__u32 *ret, + int interface, __u32 dst_ipaddr, struct net *ns); + +bool lnet_is_route_alive(struct lnet_route *route); +bool lnet_is_gateway_alive(struct lnet_peer *gw); static inline int lnet_is_wire_handle_none(struct lnet_handle_wire *wh) { @@ -174,8 +201,6 @@ lnet_net_lock_current(void) #define lnet_ptl_lock(ptl) spin_lock(&(ptl)->ptl_lock) #define lnet_ptl_unlock(ptl) spin_unlock(&(ptl)->ptl_lock) -#define lnet_eq_wait_lock() spin_lock(&the_lnet.ln_eq_wait_lock) -#define lnet_eq_wait_unlock() spin_unlock(&the_lnet.ln_eq_wait_lock) #define lnet_ni_lock(ni) spin_lock(&(ni)->ni_lock) #define lnet_ni_unlock(ni) spin_unlock(&(ni)->ni_lock) @@ -185,68 +210,16 @@ lnet_net_lock_current(void) extern struct kmem_cache *lnet_mes_cachep; /* MEs kmem_cache */ extern struct kmem_cache *lnet_small_mds_cachep; /* <= LNET_SMALL_MD_SIZE bytes * MDs kmem_cache */ - -static inline struct lnet_eq * -lnet_eq_alloc (void) -{ - struct lnet_eq *eq; - - LIBCFS_ALLOC(eq, sizeof(*eq)); - return (eq); -} - -static inline void -lnet_eq_free(struct lnet_eq *eq) -{ - LIBCFS_FREE(eq, sizeof(*eq)); -} - -static inline struct lnet_libmd * -lnet_md_alloc(struct lnet_md *umd) -{ - struct lnet_libmd *md; - unsigned int size; - unsigned int niov; - - if ((umd->options & LNET_MD_KIOV) != 0) { - niov = umd->length; - size = offsetof(struct lnet_libmd, md_iov.kiov[niov]); - } else { - niov = ((umd->options & LNET_MD_IOVEC) != 0) ? - umd->length : 1; - size = offsetof(struct lnet_libmd, md_iov.iov[niov]); - } - - if (size <= LNET_SMALL_MD_SIZE) { - md = kmem_cache_alloc(lnet_small_mds_cachep, - GFP_NOFS | __GFP_ZERO); - if (md) { - CDEBUG(D_MALLOC, "slab-alloced 'md' of size %u at " - "%p.\n", size, md); - } else { - CDEBUG(D_MALLOC, "failed to allocate 'md' of size %u\n", - size); - return NULL; - } - } else { - LIBCFS_ALLOC(md, size); - } - - if (md != NULL) { - /* Set here in case of early free */ - md->md_options = umd->options; - md->md_niov = niov; - INIT_LIST_HEAD(&md->md_list); - } - - return md; -} +extern struct kmem_cache *lnet_rspt_cachep; +extern struct kmem_cache *lnet_msg_cachep; static inline void lnet_md_free(struct lnet_libmd *md) { unsigned int size; + LASSERTF(md->md_rspt_ptr == NULL, "md %p rsp %p\n", md, md->md_rspt_ptr); + if ((md->md_options & LNET_MD_KIOV) != 0) size = offsetof(struct lnet_libmd, md_iov.kiov[md->md_niov]); else @@ -260,28 +233,6 @@ lnet_md_free(struct lnet_libmd *md) } } -static inline struct lnet_me * -lnet_me_alloc (void) -{ - struct lnet_me *me; - - me = kmem_cache_alloc(lnet_mes_cachep, GFP_NOFS | __GFP_ZERO); - - if (me) - CDEBUG(D_MALLOC, "slab-alloced 'me' at %p.\n", me); - else - CDEBUG(D_MALLOC, "failed to allocate 'me'\n"); - - return me; -} - -static inline void -lnet_me_free(struct lnet_me *me) -{ - CDEBUG(D_MALLOC, "slab-freed 'me' at %p.\n", me); - kmem_cache_free(lnet_mes_cachep, me); -} - struct lnet_libhandle *lnet_res_lh_lookup(struct lnet_res_container *rec, __u64 cookie); void lnet_res_lh_initialize(struct lnet_res_container *rec, @@ -295,30 +246,6 @@ lnet_res_lh_invalidate(struct lnet_libhandle *lh) } static inline void -lnet_eq2handle(struct lnet_handle_eq *handle, struct lnet_eq *eq) -{ - if (eq == NULL) { - LNetInvalidateEQHandle(handle); - return; - } - - handle->cookie = eq->eq_lh.lh_cookie; -} - -static inline struct lnet_eq * -lnet_handle2eq(struct lnet_handle_eq *handle) -{ - /* ALWAYS called with resource lock held */ - struct lnet_libhandle *lh; - - lh = lnet_res_lh_lookup(&the_lnet.ln_eq_container, handle->cookie); - if (lh == NULL) - return NULL; - - return lh_entry(lh, struct lnet_eq, eq_lh); -} - -static inline void lnet_md2handle(struct lnet_handle_md *handle, struct lnet_libmd *md) { handle->cookie = md->md_lh.lh_cookie; @@ -360,25 +287,33 @@ lnet_wire_handle2md(struct lnet_handle_wire *wh) } static inline void -lnet_me2handle(struct lnet_handle_me *handle, struct lnet_me *me) +lnet_peer_net_addref_locked(struct lnet_peer_net *lpn) { - handle->cookie = me->me_lh.lh_cookie; + atomic_inc(&lpn->lpn_refcount); } -static inline struct lnet_me * -lnet_handle2me(struct lnet_handle_me *handle) +extern void lnet_destroy_peer_net_locked(struct lnet_peer_net *lpn); + +static inline void +lnet_peer_net_decref_locked(struct lnet_peer_net *lpn) { - /* ALWAYS called with resource lock held */ - struct lnet_libhandle *lh; - int cpt; + if (atomic_dec_and_test(&lpn->lpn_refcount)) + lnet_destroy_peer_net_locked(lpn); +} - cpt = lnet_cpt_of_cookie(handle->cookie); - lh = lnet_res_lh_lookup(the_lnet.ln_me_containers[cpt], - handle->cookie); - if (lh == NULL) - return NULL; +static inline void +lnet_peer_addref_locked(struct lnet_peer *lp) +{ + atomic_inc(&lp->lp_refcount); +} - return lh_entry(lh, struct lnet_me, me_lh); +extern void lnet_destroy_peer_locked(struct lnet_peer *lp); + +static inline void +lnet_peer_decref_locked(struct lnet_peer *lp) +{ + if (atomic_dec_and_test(&lp->lp_refcount)) + lnet_destroy_peer_locked(lp); } static inline void @@ -399,9 +334,9 @@ lnet_peer_ni_decref_locked(struct lnet_peer_ni *lp) } static inline int -lnet_isrouter(struct lnet_peer_ni *lp) +lnet_isrouter(struct lnet_peer_ni *lpni) { - return lp->lpni_rtr_refcount != 0; + return lpni->lpni_peer_net->lpn_peer->lp_rtr_refcount != 0; } static inline void @@ -443,9 +378,8 @@ lnet_msg_alloc(void) { struct lnet_msg *msg; - LIBCFS_ALLOC(msg, sizeof(*msg)); + msg = kmem_cache_alloc(lnet_msg_cachep, GFP_NOFS | __GFP_ZERO); - /* no need to zero, LIBCFS_ALLOC does for us */ return (msg); } @@ -453,7 +387,33 @@ static inline void lnet_msg_free(struct lnet_msg *msg) { LASSERT(!msg->msg_onactivelist); - LIBCFS_FREE(msg, sizeof(*msg)); + kmem_cache_free(lnet_msg_cachep, msg); +} + +static inline struct lnet_rsp_tracker * +lnet_rspt_alloc(int cpt) +{ + struct lnet_rsp_tracker *rspt; + + rspt = kmem_cache_alloc(lnet_rspt_cachep, GFP_NOFS | __GFP_ZERO); + if (rspt) { + lnet_net_lock(cpt); + the_lnet.ln_counters[cpt]->lct_health.lch_rst_alloc++; + lnet_net_unlock(cpt); + } + CDEBUG(D_MALLOC, "rspt alloc %p\n", rspt); + return rspt; +} + +static inline void +lnet_rspt_free(struct lnet_rsp_tracker *rspt, int cpt) +{ + CDEBUG(D_MALLOC, "rspt free %p\n", rspt); + + kmem_cache_free(lnet_rspt_cachep, rspt); + lnet_net_lock(cpt); + the_lnet.ln_counters[cpt]->lct_health.lch_rst_alloc--; + lnet_net_unlock(cpt); } void lnet_ni_free(struct lnet_ni *ni); @@ -483,7 +443,7 @@ lnet_net2rnethash(__u32 net) ((1U << the_lnet.ln_remote_nets_hbits) - 1)]; } -extern struct lnet_lnd the_lolnd; +extern const struct lnet_lnd the_lolnd; extern int avoid_asym_router_failure; extern unsigned int lnet_nid_cpt_hash(lnet_nid_t nid, unsigned int number); @@ -493,55 +453,61 @@ extern struct lnet_ni *lnet_nid2ni_locked(lnet_nid_t nid, int cpt); extern struct lnet_ni *lnet_nid2ni_addref(lnet_nid_t nid); extern struct lnet_ni *lnet_net2ni_locked(__u32 net, int cpt); extern struct lnet_ni *lnet_net2ni_addref(__u32 net); -bool lnet_is_ni_healthy_locked(struct lnet_ni *ni); struct lnet_net *lnet_get_net_locked(__u32 net_id); int lnet_lib_init(void); void lnet_lib_exit(void); +extern unsigned lnet_transaction_timeout; +extern unsigned lnet_retry_count; extern unsigned int lnet_numa_range; +extern unsigned int lnet_health_sensitivity; +extern unsigned int lnet_recovery_interval; +extern unsigned int lnet_peer_discovery_disabled; +extern unsigned int lnet_drop_asym_route; +extern unsigned int router_sensitivity_percentage; +extern int alive_router_check_interval; +extern int live_router_check_interval; +extern int dead_router_check_interval; extern int portal_rotor; -int lnet_notify(struct lnet_ni *ni, lnet_nid_t peer, int alive, - cfs_time_t when); +void lnet_mt_event_handler(struct lnet_event *event); + +int lnet_notify(struct lnet_ni *ni, lnet_nid_t peer, bool alive, bool reset, + time64_t when); void lnet_notify_locked(struct lnet_peer_ni *lp, int notifylnd, int alive, - cfs_time_t when); + time64_t when); int lnet_add_route(__u32 net, __u32 hops, lnet_nid_t gateway_nid, - unsigned int priority); -int lnet_check_routes(void); + __u32 priority, __u32 sensitivity); int lnet_del_route(__u32 net, lnet_nid_t gw_nid); +void lnet_move_route(struct lnet_route *route, struct lnet_peer *lp, + struct list_head *rt_list); void lnet_destroy_routes(void); int lnet_get_route(int idx, __u32 *net, __u32 *hops, - lnet_nid_t *gateway, __u32 *alive, __u32 *priority); + lnet_nid_t *gateway, __u32 *alive, __u32 *priority, + __u32 *sensitivity); int lnet_get_rtr_pool_cfg(int idx, struct lnet_ioctl_pool_cfg *pool_cfg); struct lnet_ni *lnet_get_next_ni_locked(struct lnet_net *mynet, struct lnet_ni *prev); struct lnet_ni *lnet_get_ni_idx_locked(int idx); -struct libcfs_ioctl_handler { - struct list_head item; - int (*handle_ioctl)(unsigned int cmd, struct libcfs_ioctl_hdr *hdr); -}; - -#define DECLARE_IOCTL_HANDLER(ident, func) \ - static struct libcfs_ioctl_handler ident = { \ - .item = LIST_HEAD_INIT(ident.item), \ - .handle_ioctl = func \ - } - -extern int libcfs_register_ioctl(struct libcfs_ioctl_handler *hand); -extern int libcfs_deregister_ioctl(struct libcfs_ioctl_handler *hand); extern int libcfs_ioctl_getdata(struct libcfs_ioctl_hdr **hdr_pp, struct libcfs_ioctl_hdr __user *uparam); +extern int lnet_get_peer_list(__u32 *countp, __u32 *sizep, + struct lnet_process_id __user *ids); +extern void lnet_peer_ni_set_healthv(lnet_nid_t nid, int value, bool all); +extern void lnet_peer_ni_add_to_recoveryq_locked(struct lnet_peer_ni *lpni); -void lnet_proc_init(void); -void lnet_proc_fini(void); +void lnet_router_debugfs_init(void); +void lnet_router_debugfs_fini(void); int lnet_rtrpools_alloc(int im_a_router); void lnet_destroy_rtrbuf(struct lnet_rtrbuf *rb, int npages); int lnet_rtrpools_adjust(int tiny, int small, int large); int lnet_rtrpools_enable(void); void lnet_rtrpools_disable(void); void lnet_rtrpools_free(int keep_pools); +void lnet_rtr_transfer_to_peer(struct lnet_peer *src, + struct lnet_peer *target); struct lnet_remotenet *lnet_find_rnet_locked(__u32 net); int lnet_dyn_add_net(struct lnet_ioctl_config_data *conf); int lnet_dyn_del_net(__u32 net); @@ -552,25 +518,28 @@ struct lnet_net *lnet_get_net_locked(__u32 net_id); int lnet_islocalnid(lnet_nid_t nid); int lnet_islocalnet(__u32 net); +int lnet_islocalnet_locked(__u32 net); void lnet_msg_attach_md(struct lnet_msg *msg, struct lnet_libmd *md, unsigned int offset, unsigned int mlen); -void lnet_msg_detach_md(struct lnet_msg *msg, int status); void lnet_build_unlink_event(struct lnet_libmd *md, struct lnet_event *ev); void lnet_build_msg_event(struct lnet_msg *msg, enum lnet_event_kind ev_type); void lnet_msg_commit(struct lnet_msg *msg, int cpt); void lnet_msg_decommit(struct lnet_msg *msg, int cpt, int status); -void lnet_eq_enqueue_event(struct lnet_eq *eq, struct lnet_event *ev); void lnet_prep_send(struct lnet_msg *msg, int type, struct lnet_process_id target, unsigned int offset, unsigned int len); int lnet_send(lnet_nid_t nid, struct lnet_msg *msg, lnet_nid_t rtr_nid); +int lnet_send_ping(lnet_nid_t dest_nid, struct lnet_handle_md *mdh, int nnis, + void *user_ptr, lnet_handler_t handler, bool recovery); void lnet_return_tx_credits_locked(struct lnet_msg *msg); void lnet_return_rx_credits_locked(struct lnet_msg *msg); void lnet_schedule_blocked_locked(struct lnet_rtrbufpool *rbp); void lnet_drop_routed_msgs_locked(struct list_head *list, int cpt); +struct list_head **lnet_create_array_of_queues(void); + /* portals functions */ /* portals attributes */ static inline int @@ -640,11 +609,16 @@ struct lnet_msg *lnet_create_reply_msg(struct lnet_ni *ni, struct lnet_msg *get_msg); void lnet_set_reply_msg_len(struct lnet_ni *ni, struct lnet_msg *msg, unsigned int len); +void lnet_detach_rsp_tracker(struct lnet_libmd *md, int cpt); +void lnet_clean_zombie_rstqs(void); void lnet_finalize(struct lnet_msg *msg, int rc); +bool lnet_send_error_simulation(struct lnet_msg *msg, + enum lnet_msg_hstatus *hstatus); +void lnet_handle_remote_failure_locked(struct lnet_peer_ni *lpni); void lnet_drop_message(struct lnet_ni *ni, int cpt, void *private, - unsigned int nob); + unsigned int nob, __u32 msg_type); void lnet_drop_delayed_msg_list(struct list_head *head, char *reason); void lnet_recv_delayed_msg_list(struct list_head *head); @@ -653,6 +627,7 @@ void lnet_msg_container_cleanup(struct lnet_msg_container *container); void lnet_msg_containers_destroy(void); int lnet_msg_containers_create(void); +char *lnet_health_error2str(enum lnet_msg_hstatus hstatus); char *lnet_msgtyp2str(int type); void lnet_print_hdr(struct lnet_hdr *hdr); int lnet_fail_nid(lnet_nid_t nid, unsigned int threshold); @@ -663,7 +638,8 @@ int lnet_fault_ctl(int cmd, struct libcfs_ioctl_data *data); int lnet_fault_init(void); void lnet_fault_fini(void); -bool lnet_drop_rule_match(struct lnet_hdr *hdr); +bool lnet_drop_rule_match(struct lnet_hdr *hdr, lnet_nid_t local_nid, + enum lnet_msg_hstatus *hstatus); int lnet_delay_rule_add(struct lnet_fault_attr *attr); int lnet_delay_rule_del(lnet_nid_t src, lnet_nid_t dst, bool shutdown); @@ -675,6 +651,7 @@ bool lnet_delay_rule_match_locked(struct lnet_hdr *hdr, struct lnet_msg *msg); /** @} lnet_fault_simulation */ +void lnet_counters_get_common(struct lnet_counters_common *common); void lnet_counters_get(struct lnet_counters *counters); void lnet_counters_reset(void); @@ -683,10 +660,10 @@ int lnet_extract_iov(int dst_niov, struct kvec *dst, int src_niov, struct kvec *src, unsigned int offset, unsigned int len); -unsigned int lnet_kiov_nob (unsigned int niov, lnet_kiov_t *iov); -int lnet_extract_kiov(int dst_niov, lnet_kiov_t *dst, - int src_niov, lnet_kiov_t *src, - unsigned int offset, unsigned int len); +unsigned int lnet_kiov_nob(unsigned int niov, struct bio_vec *iov); +int lnet_extract_kiov(int dst_niov, struct bio_vec *dst, + int src_niov, struct bio_vec *src, + unsigned int offset, unsigned int len); void lnet_copy_iov2iov(unsigned int ndiov, struct kvec *diov, unsigned int doffset, @@ -694,15 +671,15 @@ void lnet_copy_iov2iov(unsigned int ndiov, struct kvec *diov, unsigned int soffset, unsigned int nob); void lnet_copy_kiov2iov(unsigned int niov, struct kvec *iov, unsigned int iovoffset, - unsigned int nkiov, lnet_kiov_t *kiov, + unsigned int nkiov, struct bio_vec *kiov, unsigned int kiovoffset, unsigned int nob); -void lnet_copy_iov2kiov(unsigned int nkiov, lnet_kiov_t *kiov, +void lnet_copy_iov2kiov(unsigned int nkiov, struct bio_vec *kiov, unsigned int kiovoffset, unsigned int niov, struct kvec *iov, unsigned int iovoffset, unsigned int nob); -void lnet_copy_kiov2kiov(unsigned int ndkiov, lnet_kiov_t *dkiov, +void lnet_copy_kiov2kiov(unsigned int ndkiov, struct bio_vec *dkiov, unsigned int doffset, - unsigned int nskiov, lnet_kiov_t *skiov, + unsigned int nskiov, struct bio_vec *skiov, unsigned int soffset, unsigned int nob); static inline void @@ -718,7 +695,7 @@ lnet_copy_iov2flat(int dlen, void *dest, unsigned int doffset, static inline void lnet_copy_kiov2flat(int dlen, void *dest, unsigned int doffset, - unsigned int nsiov, lnet_kiov_t *skiov, + unsigned int nsiov, struct bio_vec *skiov, unsigned int soffset, unsigned int nob) { struct kvec diov = { .iov_base = dest, .iov_len = dlen }; @@ -738,7 +715,7 @@ lnet_copy_flat2iov(unsigned int ndiov, struct kvec *diov, unsigned int doffset, } static inline void -lnet_copy_flat2kiov(unsigned int ndiov, lnet_kiov_t *dkiov, +lnet_copy_flat2kiov(unsigned int ndiov, struct bio_vec *dkiov, unsigned int doffset, int slen, void *src, unsigned int soffset, unsigned int nob) { @@ -750,15 +727,16 @@ lnet_copy_flat2kiov(unsigned int ndiov, lnet_kiov_t *dkiov, void lnet_me_unlink(struct lnet_me *me); void lnet_md_unlink(struct lnet_libmd *md); -void lnet_md_deconstruct(struct lnet_libmd *lmd, struct lnet_md *umd); +void lnet_md_deconstruct(struct lnet_libmd *lmd, struct lnet_event *ev); struct page *lnet_kvaddr_to_page(unsigned long vaddr); int lnet_cpt_of_md(struct lnet_libmd *md, unsigned int offset); -void lnet_register_lnd(struct lnet_lnd *lnd); -void lnet_unregister_lnd(struct lnet_lnd *lnd); +unsigned int lnet_get_lnd_timeout(void); +void lnet_register_lnd(const struct lnet_lnd *lnd); +void lnet_unregister_lnd(const struct lnet_lnd *lnd); -int lnet_connect(struct socket **sockp, lnet_nid_t peer_nid, - __u32 local_ip, __u32 peer_ip, int peer_port); +struct socket *lnet_connect(lnet_nid_t peer_nid, int interface, __u32 peer_ip, + int peer_port, struct net *ns); void lnet_connect_console_error(int rc, lnet_nid_t peer_nid, __u32 peer_ip, int port); int lnet_count_acceptor_nets(void); @@ -767,27 +745,40 @@ int lnet_acceptor_port(void); int lnet_acceptor_start(void); void lnet_acceptor_stop(void); -int lnet_ipif_query(char *name, int *up, __u32 *ip, __u32 *mask); -int lnet_ipif_enumerate(char ***names); -void lnet_ipif_free_enumeration(char **names, int n); +struct lnet_inetdev { + u32 li_cpt; + u32 li_flags; + u32 li_ipaddr; + u32 li_netmask; + char li_name[IFNAMSIZ]; +}; + +int lnet_inet_enumerate(struct lnet_inetdev **dev_list, struct net *ns); int lnet_sock_setbuf(struct socket *socket, int txbufsize, int rxbufsize); int lnet_sock_getbuf(struct socket *socket, int *txbufsize, int *rxbufsize); int lnet_sock_getaddr(struct socket *socket, bool remote, __u32 *ip, int *port); int lnet_sock_write(struct socket *sock, void *buffer, int nob, int timeout); int lnet_sock_read(struct socket *sock, void *buffer, int nob, int timeout); -int lnet_sock_listen(struct socket **sockp, __u32 ip, int port, int backlog); -int lnet_sock_accept(struct socket **newsockp, struct socket *sock); -int lnet_sock_connect(struct socket **sockp, int *fatal, - __u32 local_ip, int local_port, - __u32 peer_ip, int peer_port); +struct socket *lnet_sock_listen(int port, int backlog, + struct net *ns); +struct socket *lnet_sock_connect(int interface, int local_port, + struct sockaddr *peeraddr, + struct net *ns); int lnet_peers_start_down(void); int lnet_peer_buffer_credits(struct lnet_net *net); +void lnet_consolidate_routes_locked(struct lnet_peer *orig_lp, + struct lnet_peer *new_lp); +void lnet_router_discovery_complete(struct lnet_peer *lp); +void lnet_router_discovery_ping_reply(struct lnet_peer *lp); -int lnet_router_checker_start(void); -void lnet_router_checker_stop(void); -void lnet_router_ni_update_locked(struct lnet_peer_ni *gw, __u32 net); +int lnet_monitor_thr_start(void); +void lnet_monitor_thr_stop(void); + +bool lnet_router_checker_active(void); +void lnet_check_routers(void); +void lnet_wait_router_start(void); void lnet_swap_pinginfo(struct lnet_ping_buffer *pbuf); int lnet_ping_info_validate(struct lnet_ping_info *pinfo); @@ -801,15 +792,22 @@ static inline void lnet_ping_buffer_addref(struct lnet_ping_buffer *pbuf) static inline void lnet_ping_buffer_decref(struct lnet_ping_buffer *pbuf) { - if (atomic_dec_and_test(&pbuf->pb_refcnt)) + if (atomic_dec_and_test(&pbuf->pb_refcnt)) { + wake_up_var(&pbuf->pb_refcnt); lnet_ping_buffer_free(pbuf); + } } -static inline int lnet_ping_buffer_numref(struct lnet_ping_buffer *pbuf) +static inline int lnet_push_target_resize_needed(void) { - return atomic_read(&pbuf->pb_refcnt); + return the_lnet.ln_push_target->pb_nnis < the_lnet.ln_push_target_nnis; } +int lnet_push_target_resize(void); +int lnet_push_target_post(struct lnet_ping_buffer *pbuf, + struct lnet_handle_md *mdh); +void lnet_peer_push_event(struct lnet_event *ev); + int lnet_parse_ip2nets(char **networksp, char *ip2nets); int lnet_parse_routes(char *route_str, int *im_a_router); int lnet_parse_networks(struct list_head *nilist, char *networks, @@ -819,16 +817,27 @@ bool lnet_net_unique(__u32 net_id, struct list_head *nilist, bool lnet_ni_unique_net(struct list_head *nilist, char *iface); void lnet_incr_dlc_seq(void); __u32 lnet_get_dlc_seq_locked(void); +int lnet_get_net_count(void); +extern unsigned int lnet_current_net_count; +struct lnet_peer_net *lnet_get_next_peer_net_locked(struct lnet_peer *lp, + __u32 prev_lpn_id); struct lnet_peer_ni *lnet_get_next_peer_ni_locked(struct lnet_peer *peer, struct lnet_peer_net *peer_net, struct lnet_peer_ni *prev); struct lnet_peer_ni *lnet_nid2peerni_locked(lnet_nid_t nid, lnet_nid_t pref, int cpt); struct lnet_peer_ni *lnet_nid2peerni_ex(lnet_nid_t nid, int cpt); +struct lnet_peer_ni *lnet_peer_get_ni_locked(struct lnet_peer *lp, + lnet_nid_t nid); struct lnet_peer_ni *lnet_find_peer_ni_locked(lnet_nid_t nid); +struct lnet_peer *lnet_find_peer(lnet_nid_t nid); void lnet_peer_net_added(struct lnet_net *net); lnet_nid_t lnet_peer_primary_nid_locked(lnet_nid_t nid); +int lnet_discover_peer_locked(struct lnet_peer_ni *lpni, int cpt, bool block); +int lnet_peer_discovery_start(void); +void lnet_peer_discovery_stop(void); +void lnet_push_update_to_peers(int force); void lnet_peer_tables_cleanup(struct lnet_net *net); void lnet_peer_uninit(void); int lnet_peer_tables_create(void); @@ -839,100 +848,169 @@ bool lnet_peer_is_pref_nid_locked(struct lnet_peer_ni *lpni, lnet_nid_t nid); int lnet_peer_ni_set_non_mr_pref_nid(struct lnet_peer_ni *lpni, lnet_nid_t nid); int lnet_add_peer_ni(lnet_nid_t key_nid, lnet_nid_t nid, bool mr); int lnet_del_peer_ni(lnet_nid_t key_nid, lnet_nid_t nid); -int lnet_get_peer_info(__u32 idx, lnet_nid_t *primary_nid, lnet_nid_t *nid, - bool *mr, - struct lnet_peer_ni_credit_info __user *peer_ni_info, - struct lnet_ioctl_element_stats __user *peer_ni_stats); +int lnet_get_peer_info(struct lnet_ioctl_peer_cfg *cfg, void __user *bulk); int lnet_get_peer_ni_info(__u32 peer_index, __u64 *nid, char alivness[LNET_MAX_STR_LEN], __u32 *cpt_iter, __u32 *refcount, __u32 *ni_peer_tx_credits, __u32 *peer_tx_credits, __u32 *peer_rtr_credits, __u32 *peer_min_rtr_credtis, __u32 *peer_tx_qnob); +int lnet_get_peer_ni_hstats(struct lnet_ioctl_peer_ni_hstats *stats); - -static inline __u32 -lnet_get_num_peer_nis(struct lnet_peer *peer) +static inline struct lnet_peer_net * +lnet_find_peer_net_locked(struct lnet_peer *peer, __u32 net_id) { - struct lnet_peer_net *lpn; - struct lnet_peer_ni *lpni; - __u32 count = 0; + struct lnet_peer_net *peer_net; - list_for_each_entry(lpn, &peer->lp_peer_nets, lpn_on_peer_list) - list_for_each_entry(lpni, &lpn->lpn_peer_nis, - lpni_on_peer_net_list) - count++; + list_for_each_entry(peer_net, &peer->lp_peer_nets, lpn_peer_nets) { + if (peer_net->lpn_net_id == net_id) + return peer_net; + } - return count; + return NULL; } static inline bool -lnet_is_peer_ni_healthy_locked(struct lnet_peer_ni *lpni) +lnet_peer_is_multi_rail(struct lnet_peer *lp) { - return lpni->lpni_healthy; + if (lp->lp_state & LNET_PEER_MULTI_RAIL) + return true; + return false; } -static inline void -lnet_set_peer_ni_health_locked(struct lnet_peer_ni *lpni, bool health) +static inline bool +lnet_peer_ni_is_configured(struct lnet_peer_ni *lpni) { - lpni->lpni_healthy = health; + if (lpni->lpni_peer_net->lpn_peer->lp_state & LNET_PEER_CONFIGURED) + return true; + return false; } static inline bool -lnet_is_peer_net_healthy_locked(struct lnet_peer_net *peer_net) +lnet_peer_ni_is_primary(struct lnet_peer_ni *lpni) { - struct lnet_peer_ni *lpni; + return lpni->lpni_nid == lpni->lpni_peer_net->lpn_peer->lp_primary_nid; +} - list_for_each_entry(lpni, &peer_net->lpn_peer_nis, - lpni_on_peer_net_list) { - if (lnet_is_peer_ni_healthy_locked(lpni)) - return true; - } +bool lnet_peer_is_uptodate(struct lnet_peer *lp); +bool lnet_peer_is_uptodate_locked(struct lnet_peer *lp); +bool lnet_is_discovery_disabled(struct lnet_peer *lp); +bool lnet_is_discovery_disabled_locked(struct lnet_peer *lp); +bool lnet_peer_gw_discovery(struct lnet_peer *lp); +static inline bool +lnet_peer_needs_push(struct lnet_peer *lp) +{ + if (!(lp->lp_state & LNET_PEER_MULTI_RAIL)) + return false; + if (lp->lp_state & LNET_PEER_FORCE_PUSH) + return true; + if (lp->lp_state & LNET_PEER_NO_DISCOVERY) + return false; + /* if discovery is not enabled then no need to push */ + if (lnet_peer_discovery_disabled) + return false; + if (lp->lp_node_seqno < atomic_read(&the_lnet.ln_ping_target_seqno)) + return true; return false; } +/* + * A peer is alive if it satisfies the following two conditions: + * 1. peer health >= LNET_MAX_HEALTH_VALUE * router_sensitivity_percentage + * 2. the cached NI status received when we discover the peer is UP + */ static inline bool -lnet_is_peer_healthy_locked(struct lnet_peer *peer) +lnet_is_peer_ni_alive(struct lnet_peer_ni *lpni) { - struct lnet_peer_net *peer_net; + bool halive = false; + + halive = (atomic_read(&lpni->lpni_healthv) >= + (LNET_MAX_HEALTH_VALUE * router_sensitivity_percentage / 100)); + + return halive && lpni->lpni_ns_status == LNET_NI_STATUS_UP; +} + +static inline void +lnet_update_peer_net_healthv(struct lnet_peer_ni *lpni) +{ + struct lnet_peer_net *lpn; + int best_healthv = 0; - list_for_each_entry(peer_net, &peer->lp_peer_nets, lpn_on_peer_list) { - if (lnet_is_peer_net_healthy_locked(peer_net)) - return true; + lpn = lpni->lpni_peer_net; + + list_for_each_entry(lpni, &lpn->lpn_peer_nis, lpni_peer_nis) { + int lpni_healthv = atomic_read(&lpni->lpni_healthv); + if (best_healthv < lpni_healthv) + best_healthv = lpni_healthv; } - return false; + lpn->lpn_healthv = best_healthv; } static inline void -lnet_peer_set_alive(struct lnet_peer_ni *lp) +lnet_set_lpni_healthv_locked(struct lnet_peer_ni *lpni, int value) { - lp->lpni_last_alive = lp->lpni_last_query = cfs_time_current(); - if (!lp->lpni_alive) - lnet_notify_locked(lp, 0, 1, lp->lpni_last_alive); + if (atomic_read(&lpni->lpni_healthv) == value) + return; + atomic_set(&lpni->lpni_healthv, value); + lnet_update_peer_net_healthv(lpni); } static inline bool -lnet_peer_is_multi_rail(struct lnet_peer *lp) +lnet_atomic_add_unless_max(atomic_t *v, int a, int u) { - if (lp->lp_state & LNET_PEER_MULTI_RAIL) - return true; - return false; + int c = atomic_read(v); + bool mod = false; + int old; + int m; + + if (c == u) + return mod; + + for (;;) { + if (c + a >= u) + m = u; + else + m = c + a; + old = atomic_cmpxchg(v, c, m); + + if (old == u) + break; + + if (old == c) { + mod = true; + break; + } + c = old; + } + + return mod; } -static inline bool -lnet_peer_ni_is_configured(struct lnet_peer_ni *lpni) +static inline void +lnet_inc_lpni_healthv_locked(struct lnet_peer_ni *lpni, int value) { - if (lpni->lpni_peer_net->lpn_peer->lp_state & LNET_PEER_CONFIGURED) - return true; - return false; + /* only adjust the net health if the lpni health value changed */ + if (lnet_atomic_add_unless_max(&lpni->lpni_healthv, value, + LNET_MAX_HEALTH_VALUE)) + lnet_update_peer_net_healthv(lpni); } -static inline bool -lnet_peer_ni_is_primary(struct lnet_peer_ni *lpni) +static inline void +lnet_inc_healthv(atomic_t *healthv, int value) { - return lpni->lpni_nid == lpni->lpni_peer_net->lpn_peer->lp_primary_nid; + lnet_atomic_add_unless_max(healthv, value, LNET_MAX_HEALTH_VALUE); } +void lnet_incr_stats(struct lnet_element_stats *stats, + enum lnet_msg_type msg_type, + enum lnet_stats_type stats_type); + +__u32 lnet_sum_stats(struct lnet_element_stats *stats, + enum lnet_stats_type stats_type); + +void lnet_usr_translate_stats(struct lnet_ioctl_element_msg_stats *msg_stats, + struct lnet_element_stats *stats); + #endif