4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 only,
8 * as published by the Free Software Foundation.
10 * This program is distributed in the hope that it will be useful, but
11 * WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * General Public License version 2 for more details (a copy is included
14 * in the LICENSE file that accompanied this code).
16 * You should have received a copy of the GNU General Public License
17 * version 2 along with this program; If not, see
18 * http://www.sun.com/software/products/lustre/docs/GPLv2.pdf
20 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
21 * CA 95054 USA or visit www.sun.com if you need additional information or
27 * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
28 * Use is subject to license terms.
30 * Copyright (c) 2012, 2013, Intel Corporation.
33 * This file is part of Lustre, http://www.lustre.org/
34 * Lustre is a trademark of Sun Microsystems, Inc.
36 * lustre/lvfs/lvfs_lib.c
38 * Lustre filesystem abstraction routines
40 * Author: Andreas Dilger <adilger@clusterfs.com>
43 #include <linux/module.h>
45 #include <liblustre.h>
47 #include <lustre_lib.h>
48 #include <lprocfs_status.h>
51 void lprocfs_counter_add(struct lprocfs_stats *stats, int idx, long amount)
53 struct lprocfs_counter *percpu_cntr;
54 struct lprocfs_counter_header *header;
56 unsigned long flags = 0;
61 LASSERTF(0 <= idx && idx < stats->ls_num,
62 "idx %d, ls_num %hu\n", idx, stats->ls_num);
64 /* With per-client stats, statistics are allocated only for
65 * single CPU area, so the smp_id should be 0 always. */
66 smp_id = lprocfs_stats_lock(stats, LPROCFS_GET_SMP_ID, &flags);
70 header = &stats->ls_cnt_header[idx];
71 percpu_cntr = lprocfs_stats_counter_get(stats, smp_id, idx);
72 percpu_cntr->lc_count++;
74 if (header->lc_config & LPROCFS_CNTR_AVGMINMAX) {
76 * lprocfs_counter_add() can be called in interrupt context,
77 * as memory allocation could trigger memory shrinker call
78 * ldlm_pool_shrink(), which calls lprocfs_counter_add().
81 * Only obd_memory uses LPROCFS_STATS_FLAG_IRQ_SAFE
82 * flag, because it needs accurate counting lest memory leak
83 * check reports error.
86 (stats->ls_flags & LPROCFS_STATS_FLAG_IRQ_SAFE) != 0)
87 percpu_cntr->lc_sum_irq += amount;
89 percpu_cntr->lc_sum += amount;
91 if (header->lc_config & LPROCFS_CNTR_STDDEV)
92 percpu_cntr->lc_sumsquare += (__s64)amount * amount;
93 if (amount < percpu_cntr->lc_min)
94 percpu_cntr->lc_min = amount;
95 if (amount > percpu_cntr->lc_max)
96 percpu_cntr->lc_max = amount;
98 lprocfs_stats_unlock(stats, LPROCFS_GET_SMP_ID, &flags);
100 EXPORT_SYMBOL(lprocfs_counter_add);
102 void lprocfs_counter_sub(struct lprocfs_stats *stats, int idx, long amount)
104 struct lprocfs_counter *percpu_cntr;
105 struct lprocfs_counter_header *header;
107 unsigned long flags = 0;
112 LASSERTF(0 <= idx && idx < stats->ls_num,
113 "idx %d, ls_num %hu\n", idx, stats->ls_num);
115 /* With per-client stats, statistics are allocated only for
116 * single CPU area, so the smp_id should be 0 always. */
117 smp_id = lprocfs_stats_lock(stats, LPROCFS_GET_SMP_ID, &flags);
121 header = &stats->ls_cnt_header[idx];
122 percpu_cntr = lprocfs_stats_counter_get(stats, smp_id, idx);
123 if (header->lc_config & LPROCFS_CNTR_AVGMINMAX) {
125 * Sometimes we use RCU callbacks to free memory which calls
126 * lprocfs_counter_sub(), and RCU callbacks may execute in
127 * softirq context - right now that's the only case we're in
128 * softirq context here, use separate counter for that.
131 * Only obd_memory uses LPROCFS_STATS_FLAG_IRQ_SAFE
132 * flag, because it needs accurate counting lest memory leak
133 * check reports error.
135 if (in_interrupt() &&
136 (stats->ls_flags & LPROCFS_STATS_FLAG_IRQ_SAFE) != 0)
137 percpu_cntr->lc_sum_irq -= amount;
139 percpu_cntr->lc_sum -= amount;
141 lprocfs_stats_unlock(stats, LPROCFS_GET_SMP_ID, &flags);
143 EXPORT_SYMBOL(lprocfs_counter_sub);
145 int lprocfs_stats_alloc_one(struct lprocfs_stats *stats, unsigned int cpuid)
147 struct lprocfs_counter *cntr;
148 unsigned int percpusize;
150 unsigned long flags = 0;
153 LASSERT(stats->ls_percpu[cpuid] == NULL);
154 LASSERT((stats->ls_flags & LPROCFS_STATS_FLAG_NOPERCPU) == 0);
156 percpusize = lprocfs_stats_counter_size(stats);
157 LIBCFS_ALLOC_ATOMIC(stats->ls_percpu[cpuid], percpusize);
158 if (stats->ls_percpu[cpuid] != NULL) {
160 if (unlikely(stats->ls_biggest_alloc_num <= cpuid)) {
161 if (stats->ls_flags & LPROCFS_STATS_FLAG_IRQ_SAFE)
162 spin_lock_irqsave(&stats->ls_lock, flags);
164 spin_lock(&stats->ls_lock);
165 if (stats->ls_biggest_alloc_num <= cpuid)
166 stats->ls_biggest_alloc_num = cpuid + 1;
167 if (stats->ls_flags & LPROCFS_STATS_FLAG_IRQ_SAFE) {
168 spin_unlock_irqrestore(&stats->ls_lock, flags);
170 spin_unlock(&stats->ls_lock);
173 /* initialize the ls_percpu[cpuid] non-zero counter */
174 for (i = 0; i < stats->ls_num; ++i) {
175 cntr = lprocfs_stats_counter_get(stats, cpuid, i);
176 cntr->lc_min = LC_MIN_INIT;
182 EXPORT_SYMBOL(lprocfs_stats_alloc_one);