X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=blobdiff_plain;f=lustre%2Fldlm%2Fldlm_resource.c;h=5b2ce0c0e6be7857df273305fb3a2ac7fc41fc65;hp=cbbb07472ce8fa49d0093b308d83308b92b1fb1b;hb=f07508d17b49574c7ea47a855c6e8af2b23c3add;hpb=651f2cdd2d8df1d4318f874993ab0706d16ce490;ds=sidebyside diff --git a/lustre/ldlm/ldlm_resource.c b/lustre/ldlm/ldlm_resource.c index cbbb074..5b2ce0c 100644 --- a/lustre/ldlm/ldlm_resource.c +++ b/lustre/ldlm/ldlm_resource.c @@ -43,6 +43,7 @@ struct kmem_cache *ldlm_resource_slab, *ldlm_lock_slab; struct kmem_cache *ldlm_interval_tree_slab; +struct kmem_cache *ldlm_inodebits_slab; int ldlm_srv_namespace_nr = 0; int ldlm_cli_namespace_nr = 0; @@ -110,14 +111,23 @@ static ssize_t seq_watermark_write(struct file *file, const char __user *buffer, size_t count, loff_t *off) { - __s64 value; + struct seq_file *m = file->private_data; + u64 value; __u64 watermark; - __u64 *data = ((struct seq_file *)file->private_data)->private; + __u64 *data = m->private; bool wm_low = (data == &ldlm_reclaim_threshold_mb) ? true : false; + char kernbuf[22] = ""; int rc; - rc = lprocfs_str_with_units_to_s64(buffer, count, &value, 'M'); - if (rc) { + if (count >= sizeof(kernbuf)) + return -EINVAL; + + if (copy_from_user(kernbuf, buffer, count)) + return -EFAULT; + kernbuf[count] = 0; + + rc = sysfs_memparse(kernbuf, count, &value, "MiB"); + if (rc < 0) { CERROR("Failed to set %s, rc = %d.\n", wm_low ? "lock_reclaim_threshold_mb" : "lock_limit_mb", rc); @@ -229,6 +239,7 @@ int ldlm_debugfs_setup(void) if (IS_ERR_OR_NULL(ldlm_debugfs_dir)) { CERROR("LDebugFS failed in ldlm-init\n"); rc = ldlm_debugfs_dir ? PTR_ERR(ldlm_debugfs_dir) : -ENOMEM; + ldlm_debugfs_dir = NULL; GOTO(err, rc); } @@ -239,7 +250,7 @@ int ldlm_debugfs_setup(void) CERROR("LProcFS failed in ldlm-init\n"); rc = ldlm_ns_debugfs_dir ? PTR_ERR(ldlm_ns_debugfs_dir) : -ENOMEM; - GOTO(err_type, rc); + GOTO(err, rc); } ldlm_svc_debugfs_dir = ldebugfs_register("services", @@ -249,24 +260,19 @@ int ldlm_debugfs_setup(void) CERROR("LProcFS failed in ldlm-init\n"); rc = ldlm_svc_debugfs_dir ? PTR_ERR(ldlm_svc_debugfs_dir) : -ENOMEM; - GOTO(err_ns, rc); + GOTO(err, rc); } rc = ldebugfs_add_vars(ldlm_debugfs_dir, ldlm_debugfs_list, NULL); if (rc != 0) { CERROR("LProcFS failed in ldlm-init\n"); - GOTO(err_svc, rc); + GOTO(err, rc); } RETURN(0); -err_svc: - ldebugfs_remove(&ldlm_svc_debugfs_dir); -err_ns: - ldebugfs_remove(&ldlm_ns_debugfs_dir); -err_type: - ldebugfs_remove(&ldlm_debugfs_dir); err: + debugfs_remove_recursive(ldlm_debugfs_dir); ldlm_svc_debugfs_dir = NULL; ldlm_ns_debugfs_dir = NULL; ldlm_debugfs_dir = NULL; @@ -275,14 +281,7 @@ err: void ldlm_debugfs_cleanup(void) { - if (!IS_ERR_OR_NULL(ldlm_svc_debugfs_dir)) - ldebugfs_remove(&ldlm_svc_debugfs_dir); - - if (!IS_ERR_OR_NULL(ldlm_ns_debugfs_dir)) - ldebugfs_remove(&ldlm_ns_debugfs_dir); - - if (!IS_ERR_OR_NULL(ldlm_debugfs_dir)) - ldebugfs_remove(&ldlm_debugfs_dir); + debugfs_remove_recursive(ldlm_debugfs_dir); ldlm_svc_debugfs_dir = NULL; ldlm_ns_debugfs_dir = NULL; @@ -295,12 +294,11 @@ static ssize_t resource_count_show(struct kobject *kobj, struct attribute *attr, struct ldlm_namespace *ns = container_of(kobj, struct ldlm_namespace, ns_kobj); __u64 res = 0; - struct cfs_hash_bd bd; int i; /* result is not strictly consistant */ - cfs_hash_for_each_bucket(ns->ns_rs_hash, &bd, i) - res += cfs_hash_bd_count_get(&bd); + for (i = 0; i < (1 << ns->ns_bucket_bits); i++) + res += atomic_read(&ns->ns_rs_buckets[i].nsb_count); return sprintf(buf, "%lld\n", res); } LUSTRE_RO_ATTR(resource_count); @@ -351,10 +349,10 @@ static ssize_t lru_size_store(struct kobject *kobj, struct attribute *attr, int err; if (strncmp(buffer, "clear", 5) == 0) { - CDEBUG(D_DLMTRACE, - "dropping all unused locks from namespace %s\n", - ldlm_ns_name(ns)); - if (ns_connect_lru_resize(ns)) { + CDEBUG(D_DLMTRACE, + "dropping all unused locks from namespace %s\n", + ldlm_ns_name(ns)); + if (ns_connect_lru_resize(ns)) { /* Try to cancel all @ns_nr_unused locks. */ ldlm_cancel_lru(ns, ns->ns_nr_unused, 0, LDLM_LRU_FLAG_PASSED | @@ -372,8 +370,8 @@ static ssize_t lru_size_store(struct kobject *kobj, struct attribute *attr, err = kstrtoul(buffer, 10, &tmp); if (err != 0) { CERROR("lru_size: invalid value written\n"); - return -EINVAL; - } + return -EINVAL; + } lru_resize = (tmp == 0); if (ns_connect_lru_resize(ns)) { @@ -396,7 +394,7 @@ static ssize_t lru_size_store(struct kobject *kobj, struct attribute *attr, ldlm_ns_name(ns)); ns->ns_connect_flags &= ~OBD_CONNECT_LRU_RESIZE; } - } else { + } else { CDEBUG(D_DLMTRACE, "changing namespace %s max_unused from %u to %u\n", ldlm_ns_name(ns), ns->ns_max_unused, @@ -407,16 +405,16 @@ static ssize_t lru_size_store(struct kobject *kobj, struct attribute *attr, /* Make sure that LRU resize was originally supported before * turning it on here. */ - if (lru_resize && - (ns->ns_orig_connect_flags & OBD_CONNECT_LRU_RESIZE)) { - CDEBUG(D_DLMTRACE, - "enable lru_resize for namespace %s\n", - ldlm_ns_name(ns)); - ns->ns_connect_flags |= OBD_CONNECT_LRU_RESIZE; - } - } + if (lru_resize && + (ns->ns_orig_connect_flags & OBD_CONNECT_LRU_RESIZE)) { + CDEBUG(D_DLMTRACE, + "enable lru_resize for namespace %s\n", + ldlm_ns_name(ns)); + ns->ns_connect_flags |= OBD_CONNECT_LRU_RESIZE; + } + } - return count; + return count; } LUSTRE_RW_ATTR(lru_size); @@ -701,10 +699,10 @@ static struct kobj_type ldlm_ns_ktype = { static void ldlm_namespace_debugfs_unregister(struct ldlm_namespace *ns) { if (IS_ERR_OR_NULL(ns->ns_debugfs_entry)) - CERROR("dlm namespace %s has no procfs dir?\n", - ldlm_ns_name(ns)); + CERROR("dlm namespace %s has no procfs dir?\n", + ldlm_ns_name(ns)); else - ldebugfs_remove(&ns->ns_debugfs_entry); + debugfs_remove_recursive(ns->ns_debugfs_entry); if (ns->ns_stats != NULL) lprocfs_free_stats(&ns->ns_stats); @@ -756,62 +754,53 @@ static int ldlm_namespace_debugfs_register(struct ldlm_namespace *ns) #undef MAX_STRING_SIZE static unsigned ldlm_res_hop_hash(struct cfs_hash *hs, - const void *key, unsigned mask) + const void *key, unsigned int mask) { - const struct ldlm_res_id *id = key; - unsigned val = 0; - unsigned i; + const struct ldlm_res_id *id = key; + unsigned int val = 0; + unsigned int i; - for (i = 0; i < RES_NAME_SIZE; i++) - val += id->name[i]; - return val & mask; + for (i = 0; i < RES_NAME_SIZE; i++) + val += id->name[i]; + return val & mask; } -static unsigned ldlm_res_hop_fid_hash(struct cfs_hash *hs, - const void *key, unsigned mask) +static unsigned int ldlm_res_hop_fid_hash(const struct ldlm_res_id *id, unsigned int bits) { - const struct ldlm_res_id *id = key; - struct lu_fid fid; - __u32 hash; - __u32 val; + struct lu_fid fid; + __u32 hash; + __u32 val; - fid.f_seq = id->name[LUSTRE_RES_ID_SEQ_OFF]; - fid.f_oid = (__u32)id->name[LUSTRE_RES_ID_VER_OID_OFF]; - fid.f_ver = (__u32)(id->name[LUSTRE_RES_ID_VER_OID_OFF] >> 32); + fid.f_seq = id->name[LUSTRE_RES_ID_SEQ_OFF]; + fid.f_oid = (__u32)id->name[LUSTRE_RES_ID_VER_OID_OFF]; + fid.f_ver = (__u32)(id->name[LUSTRE_RES_ID_VER_OID_OFF] >> 32); hash = fid_flatten32(&fid); hash += (hash >> 4) + (hash << 12); /* mixing oid and seq */ if (id->name[LUSTRE_RES_ID_HSH_OFF] != 0) { val = id->name[LUSTRE_RES_ID_HSH_OFF]; - hash += (val >> 5) + (val << 11); } else { val = fid_oid(&fid); } - hash = hash_long(hash, hs->hs_bkt_bits); - /* give me another random factor */ - hash -= hash_long((unsigned long)hs, val % 11 + 3); - - hash <<= hs->hs_cur_bits - hs->hs_bkt_bits; - hash |= ldlm_res_hop_hash(hs, key, CFS_HASH_NBKT(hs) - 1); - - return hash & mask; + hash += (val >> 5) + (val << 11); + return cfs_hash_32(hash, bits); } static void *ldlm_res_hop_key(struct hlist_node *hnode) { - struct ldlm_resource *res; + struct ldlm_resource *res; res = hlist_entry(hnode, struct ldlm_resource, lr_hash); - return &res->lr_name; + return &res->lr_name; } static int ldlm_res_hop_keycmp(const void *key, struct hlist_node *hnode) { - struct ldlm_resource *res; + struct ldlm_resource *res; res = hlist_entry(hnode, struct ldlm_resource, lr_hash); - return ldlm_res_eq((const struct ldlm_res_id *)key, - (const struct ldlm_res_id *)&res->lr_name); + return ldlm_res_eq((const struct ldlm_res_id *)key, + (const struct ldlm_res_id *)&res->lr_name); } static void *ldlm_res_hop_object(struct hlist_node *hnode) @@ -822,91 +811,60 @@ static void *ldlm_res_hop_object(struct hlist_node *hnode) static void ldlm_res_hop_get_locked(struct cfs_hash *hs, struct hlist_node *hnode) { - struct ldlm_resource *res; + struct ldlm_resource *res; res = hlist_entry(hnode, struct ldlm_resource, lr_hash); - ldlm_resource_getref(res); + ldlm_resource_getref(res); } static void ldlm_res_hop_put(struct cfs_hash *hs, struct hlist_node *hnode) { - struct ldlm_resource *res; + struct ldlm_resource *res; res = hlist_entry(hnode, struct ldlm_resource, lr_hash); - ldlm_resource_putref(res); + ldlm_resource_putref(res); } static struct cfs_hash_ops ldlm_ns_hash_ops = { - .hs_hash = ldlm_res_hop_hash, - .hs_key = ldlm_res_hop_key, - .hs_keycmp = ldlm_res_hop_keycmp, - .hs_keycpy = NULL, - .hs_object = ldlm_res_hop_object, - .hs_get = ldlm_res_hop_get_locked, - .hs_put = ldlm_res_hop_put + .hs_hash = ldlm_res_hop_hash, + .hs_key = ldlm_res_hop_key, + .hs_keycmp = ldlm_res_hop_keycmp, + .hs_keycpy = NULL, + .hs_object = ldlm_res_hop_object, + .hs_get = ldlm_res_hop_get_locked, + .hs_put = ldlm_res_hop_put }; -static struct cfs_hash_ops ldlm_ns_fid_hash_ops = { - .hs_hash = ldlm_res_hop_fid_hash, - .hs_key = ldlm_res_hop_key, - .hs_keycmp = ldlm_res_hop_keycmp, - .hs_keycpy = NULL, - .hs_object = ldlm_res_hop_object, - .hs_get = ldlm_res_hop_get_locked, - .hs_put = ldlm_res_hop_put -}; - -typedef struct ldlm_ns_hash_def { - enum ldlm_ns_type nsd_type; +static struct { /** hash bucket bits */ unsigned nsd_bkt_bits; /** hash bits */ unsigned nsd_all_bits; - /** hash operations */ - struct cfs_hash_ops *nsd_hops; -} ldlm_ns_hash_def_t; - -static struct ldlm_ns_hash_def ldlm_ns_hash_defs[] = -{ - { - .nsd_type = LDLM_NS_TYPE_MDC, - .nsd_bkt_bits = 11, - .nsd_all_bits = 16, - .nsd_hops = &ldlm_ns_fid_hash_ops, - }, - { - .nsd_type = LDLM_NS_TYPE_MDT, - .nsd_bkt_bits = 14, - .nsd_all_bits = 21, - .nsd_hops = &ldlm_ns_fid_hash_ops, - }, - { - .nsd_type = LDLM_NS_TYPE_OSC, - .nsd_bkt_bits = 8, - .nsd_all_bits = 12, - .nsd_hops = &ldlm_ns_hash_ops, - }, - { - .nsd_type = LDLM_NS_TYPE_OST, - .nsd_bkt_bits = 11, - .nsd_all_bits = 17, - .nsd_hops = &ldlm_ns_hash_ops, - }, - { - .nsd_type = LDLM_NS_TYPE_MGC, - .nsd_bkt_bits = 4, - .nsd_all_bits = 4, - .nsd_hops = &ldlm_ns_hash_ops, - }, - { - .nsd_type = LDLM_NS_TYPE_MGT, - .nsd_bkt_bits = 4, - .nsd_all_bits = 4, - .nsd_hops = &ldlm_ns_hash_ops, - }, - { - .nsd_type = LDLM_NS_TYPE_UNKNOWN, - }, +} ldlm_ns_hash_defs[] = { + [LDLM_NS_TYPE_MDC] = { + .nsd_bkt_bits = 11, + .nsd_all_bits = 16, + }, + [LDLM_NS_TYPE_MDT] = { + .nsd_bkt_bits = 14, + .nsd_all_bits = 21, + }, + [LDLM_NS_TYPE_OSC] = { + .nsd_bkt_bits = 8, + .nsd_all_bits = 12, + }, + [LDLM_NS_TYPE_OST] = { + .nsd_bkt_bits = 11, + .nsd_all_bits = 17, + }, + [LDLM_NS_TYPE_MGC] = { + .nsd_bkt_bits = 3, + .nsd_all_bits = 4, + }, + [LDLM_NS_TYPE_MGT] = { + .nsd_bkt_bits = 3, + .nsd_all_bits = 4, + }, }; /** @@ -918,59 +876,66 @@ struct ldlm_namespace *ldlm_namespace_new(struct obd_device *obd, char *name, enum ldlm_ns_type ns_type) { struct ldlm_namespace *ns = NULL; - struct ldlm_ns_bucket *nsb; - struct ldlm_ns_hash_def *nsd; - struct cfs_hash_bd bd; int idx; int rc; + ENTRY; + LASSERT(obd != NULL); + + rc = ldlm_get_ref(); + if (rc) { + CERROR("ldlm_get_ref failed: %d\n", rc); + RETURN(NULL); + } + + if (ns_type >= ARRAY_SIZE(ldlm_ns_hash_defs) || + ldlm_ns_hash_defs[ns_type].nsd_bkt_bits == 0) { + CERROR("Unknown type %d for ns %s\n", ns_type, name); + GOTO(out_ref, NULL); + } - LASSERT(obd != NULL); - - rc = ldlm_get_ref(); - if (rc) { - CERROR("ldlm_get_ref failed: %d\n", rc); - RETURN(NULL); - } - - for (idx = 0;;idx++) { - nsd = &ldlm_ns_hash_defs[idx]; - if (nsd->nsd_type == LDLM_NS_TYPE_UNKNOWN) { - CERROR("Unknown type %d for ns %s\n", ns_type, name); - GOTO(out_ref, NULL); - } - - if (nsd->nsd_type == ns_type) - break; - } - - OBD_ALLOC_PTR(ns); - if (!ns) - GOTO(out_ref, NULL); - - ns->ns_rs_hash = cfs_hash_create(name, - nsd->nsd_all_bits, nsd->nsd_all_bits, - nsd->nsd_bkt_bits, sizeof(*nsb), - CFS_HASH_MIN_THETA, - CFS_HASH_MAX_THETA, - nsd->nsd_hops, - CFS_HASH_DEPTH | - CFS_HASH_BIGNAME | - CFS_HASH_SPIN_BKTLOCK | - CFS_HASH_NO_ITEMREF); - if (ns->ns_rs_hash == NULL) - GOTO(out_ns, NULL); - - cfs_hash_for_each_bucket(ns->ns_rs_hash, &bd, idx) { - nsb = cfs_hash_bd_extra_get(ns->ns_rs_hash, &bd); - at_init(&nsb->nsb_at_estimate, ldlm_enqueue_min, 0); - nsb->nsb_namespace = ns; + OBD_ALLOC_PTR(ns); + if (!ns) + GOTO(out_ref, NULL); + + ns->ns_rs_hash = cfs_hash_create(name, + ldlm_ns_hash_defs[ns_type].nsd_all_bits, + ldlm_ns_hash_defs[ns_type].nsd_all_bits, + ldlm_ns_hash_defs[ns_type].nsd_bkt_bits, + 0, + CFS_HASH_MIN_THETA, + CFS_HASH_MAX_THETA, + &ldlm_ns_hash_ops, + CFS_HASH_DEPTH | + CFS_HASH_BIGNAME | + CFS_HASH_SPIN_BKTLOCK | + CFS_HASH_NO_ITEMREF); + if (ns->ns_rs_hash == NULL) + GOTO(out_ns, NULL); + + ns->ns_bucket_bits = ldlm_ns_hash_defs[ns_type].nsd_all_bits - + ldlm_ns_hash_defs[ns_type].nsd_bkt_bits; + + OBD_ALLOC_LARGE(ns->ns_rs_buckets, + BIT(ns->ns_bucket_bits) * sizeof(ns->ns_rs_buckets[0])); + if (!ns->ns_rs_buckets) + goto out_hash; + + for (idx = 0; idx < (1 << ns->ns_bucket_bits); idx++) { + struct ldlm_ns_bucket *nsb = &ns->ns_rs_buckets[idx]; + + at_init(&nsb->nsb_at_estimate, ldlm_enqueue_min, 0); + nsb->nsb_namespace = ns; nsb->nsb_reclaim_start = 0; - } + atomic_set(&nsb->nsb_count, 0); + } - ns->ns_obd = obd; - ns->ns_appetite = apt; - ns->ns_client = client; + ns->ns_obd = obd; + ns->ns_appetite = apt; + ns->ns_client = client; + ns->ns_name = kstrdup(name, GFP_KERNEL); + if (!ns->ns_name) + goto out_hash; INIT_LIST_HEAD(&ns->ns_list_chain); INIT_LIST_HEAD(&ns->ns_unused_list); @@ -982,16 +947,16 @@ struct ldlm_namespace *ldlm_namespace_new(struct obd_device *obd, char *name, ns->ns_contention_time = NS_DEFAULT_CONTENTION_SECONDS; ns->ns_contended_locks = NS_DEFAULT_CONTENDED_LOCKS; - ns->ns_max_parallel_ast = LDLM_DEFAULT_PARALLEL_AST_LIMIT; - ns->ns_nr_unused = 0; - ns->ns_max_unused = LDLM_DEFAULT_LRU_SIZE; + ns->ns_max_parallel_ast = LDLM_DEFAULT_PARALLEL_AST_LIMIT; + ns->ns_nr_unused = 0; + ns->ns_max_unused = LDLM_DEFAULT_LRU_SIZE; ns->ns_max_age = ktime_set(LDLM_DEFAULT_MAX_ALIVE, 0); - ns->ns_ctime_age_limit = LDLM_CTIME_AGE_LIMIT; + ns->ns_ctime_age_limit = LDLM_CTIME_AGE_LIMIT; ns->ns_dirty_age_limit = LDLM_DIRTY_AGE_LIMIT; - ns->ns_timeouts = 0; - ns->ns_orig_connect_flags = 0; - ns->ns_connect_flags = 0; - ns->ns_stopping = 0; + ns->ns_timeouts = 0; + ns->ns_orig_connect_flags = 0; + ns->ns_connect_flags = 0; + ns->ns_stopping = 0; ns->ns_reclaim_start = 0; ns->ns_last_pos = &ns->ns_unused_list; @@ -1007,32 +972,33 @@ struct ldlm_namespace *ldlm_namespace_new(struct obd_device *obd, char *name, GOTO(out_sysfs, rc); } - idx = ldlm_namespace_nr_read(client); - rc = ldlm_pool_init(&ns->ns_pool, ns, idx, client); - if (rc) { - CERROR("Can't initialize lock pool, rc %d\n", rc); - GOTO(out_proc, rc); - } + idx = ldlm_namespace_nr_read(client); + rc = ldlm_pool_init(&ns->ns_pool, ns, idx, client); + if (rc) { + CERROR("Can't initialize lock pool, rc %d\n", rc); + GOTO(out_proc, rc); + } - ldlm_namespace_register(ns, client); - RETURN(ns); + ldlm_namespace_register(ns, client); + RETURN(ns); out_proc: ldlm_namespace_debugfs_unregister(ns); out_sysfs: ldlm_namespace_sysfs_unregister(ns); ldlm_namespace_cleanup(ns, 0); out_hash: - cfs_hash_putref(ns->ns_rs_hash); + OBD_FREE_LARGE(ns->ns_rs_buckets, + BIT(ns->ns_bucket_bits) * sizeof(ns->ns_rs_buckets[0])); + kfree(ns->ns_name); + cfs_hash_putref(ns->ns_rs_hash); out_ns: OBD_FREE_PTR(ns); out_ref: - ldlm_put_ref(); - RETURN(NULL); + ldlm_put_ref(); + RETURN(NULL); } EXPORT_SYMBOL(ldlm_namespace_new); -extern struct ldlm_lock *ldlm_lock_get(struct ldlm_lock *lock); - /** * Cancel and destroy all locks on a resource. * @@ -1048,8 +1014,8 @@ static void cleanup_resource(struct ldlm_resource *res, struct list_head *q, int rc = 0, client = ns_is_client(ldlm_res_to_ns(res)); bool local_only = !!(flags & LDLM_FL_LOCAL_ONLY); - do { - struct ldlm_lock *lock = NULL; + do { + struct ldlm_lock *lock = NULL; /* First, we look for non-cleaned-yet lock * all cleaned locks are marked by CLEANED flag. */ @@ -1066,56 +1032,58 @@ static void cleanup_resource(struct ldlm_resource *res, struct list_head *q, break; } - if (lock == NULL) { - unlock_res(res); - break; - } + if (lock == NULL) { + unlock_res(res); + break; + } - /* Set CBPENDING so nothing in the cancellation path + /* Set CBPENDING so nothing in the cancellation path * can match this lock. */ ldlm_set_cbpending(lock); ldlm_set_failed(lock); - lock->l_flags |= flags; + lock->l_flags |= flags; - /* ... without sending a CANCEL message for local_only. */ - if (local_only) + /* ... without sending a CANCEL message for local_only. */ + if (local_only) ldlm_set_local_only(lock); - if (local_only && (lock->l_readers || lock->l_writers)) { - /* This is a little bit gross, but much better than the - * alternative: pretend that we got a blocking AST from - * the server, so that when the lock is decref'd, it - * will go away ... */ - unlock_res(res); - LDLM_DEBUG(lock, "setting FL_LOCAL_ONLY"); + if (local_only && (lock->l_readers || lock->l_writers)) { + /* + * This is a little bit gross, but much better than the + * alternative: pretend that we got a blocking AST from + * the server, so that when the lock is decref'd, it + * will go away ... + */ + unlock_res(res); + LDLM_DEBUG(lock, "setting FL_LOCAL_ONLY"); if (lock->l_flags & LDLM_FL_FAIL_LOC) { set_current_state(TASK_UNINTERRUPTIBLE); schedule_timeout(cfs_time_seconds(4)); set_current_state(TASK_RUNNING); } - if (lock->l_completion_ast) + if (lock->l_completion_ast) lock->l_completion_ast(lock, LDLM_FL_FAILED, NULL); - LDLM_LOCK_RELEASE(lock); - continue; - } + LDLM_LOCK_RELEASE(lock); + continue; + } - if (client) { - struct lustre_handle lockh; + if (client) { + struct lustre_handle lockh; - unlock_res(res); - ldlm_lock2handle(lock, &lockh); + unlock_res(res); + ldlm_lock2handle(lock, &lockh); rc = ldlm_cli_cancel(&lockh, LCF_LOCAL); - if (rc) - CERROR("ldlm_cli_cancel: %d\n", rc); - } else { - unlock_res(res); - LDLM_DEBUG(lock, "Freeing a lock still held by a " - "client node"); + if (rc) + CERROR("ldlm_cli_cancel: %d\n", rc); + } else { + unlock_res(res); + LDLM_DEBUG(lock, + "Freeing a lock still held by a client node"); ldlm_lock_cancel(lock); - } - LDLM_LOCK_RELEASE(lock); - } while (1); + } + LDLM_LOCK_RELEASE(lock); + } while (1); } static int ldlm_resource_clean(struct cfs_hash *hs, struct cfs_hash_bd *bd, @@ -1141,7 +1109,8 @@ static int ldlm_resource_complain(struct cfs_hash *hs, struct cfs_hash_bd *bd, ldlm_ns_name(ldlm_res_to_ns(res)), PLDLMRES(res), res, atomic_read(&res->lr_refcount) - 1); - ldlm_resource_dump(D_ERROR, res); + /* Use D_NETERROR since it is in the default mask */ + ldlm_resource_dump(D_NETERROR, res); unlock_res(res); return 0; } @@ -1155,10 +1124,10 @@ static int ldlm_resource_complain(struct cfs_hash *hs, struct cfs_hash_bd *bd, */ int ldlm_namespace_cleanup(struct ldlm_namespace *ns, __u64 flags) { - if (ns == NULL) { - CDEBUG(D_INFO, "NULL ns, skipping cleanup\n"); - return ELDLM_OK; - } + if (ns == NULL) { + CDEBUG(D_INFO, "NULL ns, skipping cleanup\n"); + return ELDLM_OK; + } cfs_hash_for_each_nolock(ns->ns_rs_hash, ldlm_resource_clean, &flags, 0); @@ -1181,22 +1150,24 @@ static int __ldlm_namespace_free(struct ldlm_namespace *ns, int force) ldlm_namespace_cleanup(ns, force ? LDLM_FL_LOCAL_ONLY : 0); if (atomic_read(&ns->ns_bref) > 0) { - struct l_wait_info lwi = LWI_INTR(LWI_ON_SIGNAL_NOOP, NULL); int rc; CDEBUG(D_DLMTRACE, "dlm namespace %s free waiting on refcount %d\n", ldlm_ns_name(ns), atomic_read(&ns->ns_bref)); force_wait: if (force) - lwi = LWI_TIMEOUT(msecs_to_jiffies(obd_timeout * - MSEC_PER_SEC) / 4, NULL, NULL); - - rc = l_wait_event(ns->ns_waitq, - atomic_read(&ns->ns_bref) == 0, &lwi); + rc = wait_event_idle_timeout( + ns->ns_waitq, + atomic_read(&ns->ns_bref) == 0, + cfs_time_seconds(1) / 4); + else + rc = l_wait_event_abortable( + ns->ns_waitq, atomic_read(&ns->ns_bref) == 0); /* Forced cleanups should be able to reclaim all references, * so it's safe to wait forever... we can't leak locks... */ - if (force && rc == -ETIMEDOUT) { + if (force && rc == 0) { + rc = -ETIMEDOUT; LCONSOLE_ERROR("Forced cleanup waiting for %s " "namespace with %d resources in use, " "(rc=%d)\n", ldlm_ns_name(ns), @@ -1231,35 +1202,36 @@ void ldlm_namespace_free_prior(struct ldlm_namespace *ns, struct obd_import *imp, int force) { - int rc; - ENTRY; - if (!ns) { - EXIT; - return; - } + int rc; + + ENTRY; + if (!ns) { + EXIT; + return; + } spin_lock(&ns->ns_lock); ns->ns_stopping = 1; spin_unlock(&ns->ns_lock); - /* - * Can fail with -EINTR when force == 0 in which case try harder. - */ - rc = __ldlm_namespace_free(ns, force); - if (rc != ELDLM_OK) { - if (imp) { - ptlrpc_disconnect_import(imp, 0); - ptlrpc_invalidate_import(imp); - } - - /* - * With all requests dropped and the import inactive - * we are gaurenteed all reference will be dropped. - */ - rc = __ldlm_namespace_free(ns, 1); - LASSERT(rc == 0); - } - EXIT; + /* + * Can fail with -EINTR when force == 0 in which case try harder. + */ + rc = __ldlm_namespace_free(ns, force); + if (rc != ELDLM_OK) { + if (imp) { + ptlrpc_disconnect_import(imp, 0); + ptlrpc_invalidate_import(imp); + } + + /* + * With all requests dropped and the import inactive + * we are gaurenteed all reference will be dropped. + */ + rc = __ldlm_namespace_free(ns, 1); + LASSERT(rc == 0); + } + EXIT; } EXPORT_SYMBOL(ldlm_namespace_free_prior); @@ -1270,11 +1242,11 @@ EXPORT_SYMBOL(ldlm_namespace_free_prior); */ void ldlm_namespace_free_post(struct ldlm_namespace *ns) { - ENTRY; - if (!ns) { - EXIT; - return; - } + ENTRY; + if (!ns) { + EXIT; + return; + } /* Make sure that nobody can find this ns in its list. */ ldlm_namespace_unregister(ns, ns->ns_client); @@ -1286,9 +1258,13 @@ void ldlm_namespace_free_post(struct ldlm_namespace *ns) ldlm_namespace_debugfs_unregister(ns); ldlm_namespace_sysfs_unregister(ns); cfs_hash_putref(ns->ns_rs_hash); + OBD_FREE_LARGE(ns->ns_rs_buckets, + BIT(ns->ns_bucket_bits) * sizeof(ns->ns_rs_buckets[0])); + kfree(ns->ns_name); /* Namespace \a ns should be not on list at this time, otherwise * this will cause issues related to using freed \a ns in poold - * thread. */ + * thread. + */ LASSERT(list_empty(&ns->ns_list_chain)); OBD_FREE_PTR(ns); ldlm_put_ref(); @@ -1315,11 +1291,11 @@ EXPORT_SYMBOL(ldlm_namespace_free_post); * held. */ void ldlm_namespace_free(struct ldlm_namespace *ns, - struct obd_import *imp, - int force) + struct obd_import *imp, + int force) { - ldlm_namespace_free_prior(ns, imp, force); - ldlm_namespace_free_post(ns); + ldlm_namespace_free_prior(ns, imp, force); + ldlm_namespace_free_post(ns); } EXPORT_SYMBOL(ldlm_namespace_free); @@ -1393,29 +1369,59 @@ struct ldlm_namespace *ldlm_namespace_first_locked(enum ldlm_side client) struct ldlm_namespace, ns_list_chain); } +static bool ldlm_resource_extent_new(struct ldlm_resource *res) +{ + int idx; + + OBD_SLAB_ALLOC(res->lr_itree, ldlm_interval_tree_slab, + sizeof(*res->lr_itree) * LCK_MODE_NUM); + if (res->lr_itree == NULL) + return false; + /* Initialize interval trees for each lock mode. */ + for (idx = 0; idx < LCK_MODE_NUM; idx++) { + res->lr_itree[idx].lit_size = 0; + res->lr_itree[idx].lit_mode = 1 << idx; + res->lr_itree[idx].lit_root = NULL; + } + return true; +} + +static bool ldlm_resource_inodebits_new(struct ldlm_resource *res) +{ + int i; + + OBD_ALLOC_PTR(res->lr_ibits_queues); + if (res->lr_ibits_queues == NULL) + return false; + for (i = 0; i < MDS_INODELOCK_NUMBITS; i++) + INIT_LIST_HEAD(&res->lr_ibits_queues->liq_waiting[i]); + return true; +} + /** Create and initialize new resource. */ static struct ldlm_resource *ldlm_resource_new(enum ldlm_type ldlm_type) { struct ldlm_resource *res; - int idx; + bool rc; OBD_SLAB_ALLOC_PTR_GFP(res, ldlm_resource_slab, GFP_NOFS); if (res == NULL) return NULL; - if (ldlm_type == LDLM_EXTENT) { - OBD_SLAB_ALLOC(res->lr_itree, ldlm_interval_tree_slab, - sizeof(*res->lr_itree) * LCK_MODE_NUM); - if (res->lr_itree == NULL) { - OBD_SLAB_FREE_PTR(res, ldlm_resource_slab); - return NULL; - } - /* Initialize interval trees for each lock mode. */ - for (idx = 0; idx < LCK_MODE_NUM; idx++) { - res->lr_itree[idx].lit_size = 0; - res->lr_itree[idx].lit_mode = 1 << idx; - res->lr_itree[idx].lit_root = NULL; - } + switch (ldlm_type) { + case LDLM_EXTENT: + rc = ldlm_resource_extent_new(res); + break; + case LDLM_IBITS: + rc = ldlm_resource_inodebits_new(res); + break; + default: + rc = true; + break; + } + if (!rc) { + OBD_SLAB_FREE_PTR(res, ldlm_resource_slab); + return NULL; } INIT_LIST_HEAD(&res->lr_granted); @@ -1433,6 +1439,20 @@ static struct ldlm_resource *ldlm_resource_new(enum ldlm_type ldlm_type) return res; } +static void ldlm_resource_free(struct ldlm_resource *res) +{ + if (res->lr_type == LDLM_EXTENT) { + if (res->lr_itree != NULL) + OBD_SLAB_FREE(res->lr_itree, ldlm_interval_tree_slab, + sizeof(*res->lr_itree) * LCK_MODE_NUM); + } else if (res->lr_type == LDLM_IBITS) { + if (res->lr_ibits_queues != NULL) + OBD_FREE_PTR(res->lr_ibits_queues); + } + + OBD_SLAB_FREE(res, ldlm_resource_slab, sizeof *res); +} + /** * Return a reference to resource with given name, creating it if necessary. * Args: namespace with ns_lock unlocked @@ -1449,16 +1469,17 @@ ldlm_resource_get(struct ldlm_namespace *ns, struct ldlm_resource *parent, struct cfs_hash_bd bd; __u64 version; int ns_refcount = 0; + int hash; - LASSERT(ns != NULL); - LASSERT(parent == NULL); - LASSERT(ns->ns_rs_hash != NULL); - LASSERT(name->name[0] != 0); + LASSERT(ns != NULL); + LASSERT(parent == NULL); + LASSERT(ns->ns_rs_hash != NULL); + LASSERT(name->name[0] != 0); - cfs_hash_bd_get_and_lock(ns->ns_rs_hash, (void *)name, &bd, 0); - hnode = cfs_hash_bd_lookup_locked(ns->ns_rs_hash, &bd, (void *)name); - if (hnode != NULL) { - cfs_hash_bd_unlock(ns->ns_rs_hash, &bd, 0); + cfs_hash_bd_get_and_lock(ns->ns_rs_hash, (void *)name, &bd, 0); + hnode = cfs_hash_bd_lookup_locked(ns->ns_rs_hash, &bd, (void *)name); + if (hnode != NULL) { + cfs_hash_bd_unlock(ns->ns_rs_hash, &bd, 0); GOTO(found, res); } @@ -1474,7 +1495,8 @@ ldlm_resource_get(struct ldlm_namespace *ns, struct ldlm_resource *parent, if (res == NULL) return ERR_PTR(-ENOMEM); - res->lr_ns_bucket = cfs_hash_bd_extra_get(ns->ns_rs_hash, &bd); + hash = ldlm_res_hop_fid_hash(name, ns->ns_bucket_bits); + res->lr_ns_bucket = &ns->ns_rs_buckets[hash]; res->lr_name = *name; res->lr_type = type; @@ -1487,20 +1509,17 @@ ldlm_resource_get(struct ldlm_namespace *ns, struct ldlm_resource *parent, cfs_hash_bd_unlock(ns->ns_rs_hash, &bd, 1); /* Clean lu_ref for failed resource. */ lu_ref_fini(&res->lr_reference); - if (res->lr_itree != NULL) - OBD_SLAB_FREE(res->lr_itree, ldlm_interval_tree_slab, - sizeof(*res->lr_itree) * LCK_MODE_NUM); - OBD_SLAB_FREE(res, ldlm_resource_slab, sizeof *res); + ldlm_resource_free(res); found: res = hlist_entry(hnode, struct ldlm_resource, lr_hash); return res; } /* We won! Let's add the resource. */ - cfs_hash_bd_add_locked(ns->ns_rs_hash, &bd, &res->lr_hash); - if (cfs_hash_bd_count_get(&bd) == 1) + cfs_hash_bd_add_locked(ns->ns_rs_hash, &bd, &res->lr_hash); + if (atomic_inc_return(&res->lr_ns_bucket->nsb_count) == 1) ns_refcount = ldlm_namespace_get_return(ns); - cfs_hash_bd_unlock(ns->ns_rs_hash, &bd, 1); + cfs_hash_bd_unlock(ns->ns_rs_hash, &bd, 1); OBD_FAIL_TIMEOUT(OBD_FAIL_LDLM_CREATE_RESOURCE, 2); @@ -1546,7 +1565,7 @@ static void __ldlm_resource_putref_final(struct cfs_hash_bd *bd, cfs_hash_bd_del_locked(nsb->nsb_namespace->ns_rs_hash, bd, &res->lr_hash); lu_ref_fini(&res->lr_reference); - if (cfs_hash_bd_count_get(bd) == 0) + if (atomic_dec_and_test(&nsb->nsb_count)) ldlm_namespace_put(nsb->nsb_namespace); } @@ -1566,10 +1585,7 @@ int ldlm_resource_putref(struct ldlm_resource *res) cfs_hash_bd_unlock(ns->ns_rs_hash, &bd, 1); if (ns->ns_lvbo && ns->ns_lvbo->lvbo_free) ns->ns_lvbo->lvbo_free(res); - if (res->lr_itree != NULL) - OBD_SLAB_FREE(res->lr_itree, ldlm_interval_tree_slab, - sizeof(*res->lr_itree) * LCK_MODE_NUM); - OBD_SLAB_FREE(res, ldlm_resource_slab, sizeof *res); + ldlm_resource_free(res); return 1; } return 0; @@ -1594,6 +1610,9 @@ void ldlm_resource_add_lock(struct ldlm_resource *res, struct list_head *head, LASSERT(list_empty(&lock->l_res_link)); list_add_tail(&lock->l_res_link, head); + + if (res->lr_type == LDLM_IBITS) + ldlm_inodebits_add_lock(res, head, lock); } /** @@ -1602,13 +1621,13 @@ void ldlm_resource_add_lock(struct ldlm_resource *res, struct list_head *head, * Obtain resource description from the lock we are inserting after. */ void ldlm_resource_insert_lock_after(struct ldlm_lock *original, - struct ldlm_lock *new) + struct ldlm_lock *new) { - struct ldlm_resource *res = original->l_resource; + struct ldlm_resource *res = original->l_resource; - check_res_locked(res); + check_res_locked(res); - ldlm_resource_dump(D_INFO, res); + ldlm_resource_dump(D_INFO, res); LDLM_DEBUG(new, "About to insert this lock after %p: ", original); if (ldlm_is_destroyed(new)) { @@ -1624,21 +1643,28 @@ void ldlm_resource_insert_lock_after(struct ldlm_lock *original, void ldlm_resource_unlink_lock(struct ldlm_lock *lock) { - int type = lock->l_resource->lr_type; - - check_res_locked(lock->l_resource); - if (type == LDLM_IBITS || type == LDLM_PLAIN) - ldlm_unlink_lock_skiplist(lock); - else if (type == LDLM_EXTENT) - ldlm_extent_unlink_lock(lock); + int type = lock->l_resource->lr_type; + + check_res_locked(lock->l_resource); + switch (type) { + case LDLM_PLAIN: + ldlm_unlink_lock_skiplist(lock); + break; + case LDLM_EXTENT: + ldlm_extent_unlink_lock(lock); + break; + case LDLM_IBITS: + ldlm_inodebits_unlink_lock(lock); + break; + } list_del_init(&lock->l_res_link); } EXPORT_SYMBOL(ldlm_resource_unlink_lock); void ldlm_res2desc(struct ldlm_resource *res, struct ldlm_resource_desc *desc) { - desc->lr_type = res->lr_type; - desc->lr_name = res->lr_name; + desc->lr_type = res->lr_type; + desc->lr_name = res->lr_name; } /** @@ -1667,14 +1693,14 @@ void ldlm_dump_all_namespaces(enum ldlm_side client, int level) static int ldlm_res_hash_dump(struct cfs_hash *hs, struct cfs_hash_bd *bd, struct hlist_node *hnode, void *arg) { - struct ldlm_resource *res = cfs_hash_object(hs, hnode); - int level = (int)(unsigned long)arg; + struct ldlm_resource *res = cfs_hash_object(hs, hnode); + int level = (int)(unsigned long)arg; - lock_res(res); - ldlm_resource_dump(level, res); - unlock_res(res); + lock_res(res); + ldlm_resource_dump(level, res); + unlock_res(res); - return 0; + return 0; } /** @@ -1709,7 +1735,7 @@ void ldlm_resource_dump(int level, struct ldlm_resource *res) struct ldlm_lock *lock; unsigned int granted = 0; - CLASSERT(RES_NAME_SIZE == 4); + BUILD_BUG_ON(RES_NAME_SIZE != 4); if (!((libcfs_debug | D_ERROR) & level)) return; @@ -1721,15 +1747,16 @@ void ldlm_resource_dump(int level, struct ldlm_resource *res) CDEBUG(level, "Granted locks (in reverse order):\n"); list_for_each_entry_reverse(lock, &res->lr_granted, l_res_link) { - LDLM_DEBUG_LIMIT(level, lock, "###"); - if (!(level & D_CANTMASK) && - ++granted > ldlm_dump_granted_max) { - CDEBUG(level, "only dump %d granted locks to " - "avoid DDOS.\n", granted); - break; - } - } - } + LDLM_DEBUG_LIMIT(level, lock, "###"); + if (!(level & D_CANTMASK) && + ++granted > ldlm_dump_granted_max) { + CDEBUG(level, + "only dump %d granted locks to avoid DDOS.\n", + granted); + break; + } + } + } if (!list_empty(&res->lr_waiting)) { CDEBUG(level, "Waiting locks:\n");