1 /*
2 * GPL HEADER START
3 *
4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 only,
8 * as published by the Free Software Foundation.
9 *
10 * This program is distributed in the hope that it will be useful, but
11 * WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * General Public License version 2 for more details (a copy is included
14 * in the LICENSE file that accompanied this code).
15 *
16 * You should have received a copy of the GNU General Public License
17 * version 2 along with this program; If not, see
18 * http://www.sun.com/software/products/lustre/docs/GPLv2.pdf
19 *
20 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
21 * CA 95054 USA or visit www.sun.com if you need additional information or
22 * have any questions.
23 *
24 * GPL HEADER END
25 */
26 /*
27 * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved.
28 * Use is subject to license terms.
29 *
30 * Copyright (c) 2012, Intel Corporation.
31 */
32 /*
33 * This file is part of Lustre, http://www.lustre.org/
34 * Lustre is a trademark of Sun Microsystems, Inc.
35 */
36
37 #define DEBUG_SUBSYSTEM S_LNET
38 #define LUSTRE_TRACEFILE_PRIVATE
39
40 #include "../../../include/linux/libcfs/libcfs.h"
41 #include "../tracefile.h"
42
43 /* percents to share the total debug memory for each type */
44 static unsigned int pages_factor[CFS_TCD_TYPE_MAX] = {
45 80, /* 80% pages for CFS_TCD_TYPE_PROC */
46 10, /* 10% pages for CFS_TCD_TYPE_SOFTIRQ */
47 10 /* 10% pages for CFS_TCD_TYPE_IRQ */
48 };
49
50 char *cfs_trace_console_buffers[NR_CPUS][CFS_TCD_TYPE_MAX];
51
52 static DECLARE_RWSEM(cfs_tracefile_sem);
53
cfs_tracefile_init_arch(void)54 int cfs_tracefile_init_arch(void)
55 {
56 int i;
57 int j;
58 struct cfs_trace_cpu_data *tcd;
59
60 /* initialize trace_data */
61 memset(cfs_trace_data, 0, sizeof(cfs_trace_data));
62 for (i = 0; i < CFS_TCD_TYPE_MAX; i++) {
63 cfs_trace_data[i] =
64 kmalloc(sizeof(union cfs_trace_data_union) *
65 num_possible_cpus(), GFP_KERNEL);
66 if (cfs_trace_data[i] == NULL)
67 goto out;
68
69 }
70
71 /* arch related info initialized */
72 cfs_tcd_for_each(tcd, i, j) {
73 spin_lock_init(&tcd->tcd_lock);
74 tcd->tcd_pages_factor = pages_factor[i];
75 tcd->tcd_type = i;
76 tcd->tcd_cpu = j;
77 }
78
79 for (i = 0; i < num_possible_cpus(); i++)
80 for (j = 0; j < 3; j++) {
81 cfs_trace_console_buffers[i][j] =
82 kmalloc(CFS_TRACE_CONSOLE_BUFFER_SIZE,
83 GFP_KERNEL);
84
85 if (cfs_trace_console_buffers[i][j] == NULL)
86 goto out;
87 }
88
89 return 0;
90
91 out:
92 cfs_tracefile_fini_arch();
93 printk(KERN_ERR "lnet: Not enough memory\n");
94 return -ENOMEM;
95 }
96
cfs_tracefile_fini_arch(void)97 void cfs_tracefile_fini_arch(void)
98 {
99 int i;
100 int j;
101
102 for (i = 0; i < num_possible_cpus(); i++)
103 for (j = 0; j < 3; j++) {
104 kfree(cfs_trace_console_buffers[i][j]);
105 cfs_trace_console_buffers[i][j] = NULL;
106 }
107
108 for (i = 0; cfs_trace_data[i] != NULL; i++) {
109 kfree(cfs_trace_data[i]);
110 cfs_trace_data[i] = NULL;
111 }
112 }
113
cfs_tracefile_read_lock(void)114 void cfs_tracefile_read_lock(void)
115 {
116 down_read(&cfs_tracefile_sem);
117 }
118
cfs_tracefile_read_unlock(void)119 void cfs_tracefile_read_unlock(void)
120 {
121 up_read(&cfs_tracefile_sem);
122 }
123
cfs_tracefile_write_lock(void)124 void cfs_tracefile_write_lock(void)
125 {
126 down_write(&cfs_tracefile_sem);
127 }
128
cfs_tracefile_write_unlock(void)129 void cfs_tracefile_write_unlock(void)
130 {
131 up_write(&cfs_tracefile_sem);
132 }
133
cfs_trace_buf_idx_get(void)134 cfs_trace_buf_type_t cfs_trace_buf_idx_get(void)
135 {
136 if (in_irq())
137 return CFS_TCD_TYPE_IRQ;
138 else if (in_softirq())
139 return CFS_TCD_TYPE_SOFTIRQ;
140 else
141 return CFS_TCD_TYPE_PROC;
142 }
143
144 /*
145 * The walking argument indicates the locking comes from all tcd types
146 * iterator and we must lock it and dissable local irqs to avoid deadlocks
147 * with other interrupt locks that might be happening. See LU-1311
148 * for details.
149 */
cfs_trace_lock_tcd(struct cfs_trace_cpu_data * tcd,int walking)150 int cfs_trace_lock_tcd(struct cfs_trace_cpu_data *tcd, int walking)
151 __acquires(&tcd->tc_lock)
152 {
153 __LASSERT(tcd->tcd_type < CFS_TCD_TYPE_MAX);
154 if (tcd->tcd_type == CFS_TCD_TYPE_IRQ)
155 spin_lock_irqsave(&tcd->tcd_lock, tcd->tcd_lock_flags);
156 else if (tcd->tcd_type == CFS_TCD_TYPE_SOFTIRQ)
157 spin_lock_bh(&tcd->tcd_lock);
158 else if (unlikely(walking))
159 spin_lock_irq(&tcd->tcd_lock);
160 else
161 spin_lock(&tcd->tcd_lock);
162 return 1;
163 }
164
cfs_trace_unlock_tcd(struct cfs_trace_cpu_data * tcd,int walking)165 void cfs_trace_unlock_tcd(struct cfs_trace_cpu_data *tcd, int walking)
166 __releases(&tcd->tcd_lock)
167 {
168 __LASSERT(tcd->tcd_type < CFS_TCD_TYPE_MAX);
169 if (tcd->tcd_type == CFS_TCD_TYPE_IRQ)
170 spin_unlock_irqrestore(&tcd->tcd_lock, tcd->tcd_lock_flags);
171 else if (tcd->tcd_type == CFS_TCD_TYPE_SOFTIRQ)
172 spin_unlock_bh(&tcd->tcd_lock);
173 else if (unlikely(walking))
174 spin_unlock_irq(&tcd->tcd_lock);
175 else
176 spin_unlock(&tcd->tcd_lock);
177 }
178
cfs_tcd_owns_tage(struct cfs_trace_cpu_data * tcd,struct cfs_trace_page * tage)179 int cfs_tcd_owns_tage(struct cfs_trace_cpu_data *tcd,
180 struct cfs_trace_page *tage)
181 {
182 /*
183 * XXX nikita: do NOT call portals_debug_msg() (CDEBUG/ENTRY/EXIT)
184 * from here: this will lead to infinite recursion.
185 */
186 return tcd->tcd_cpu == tage->cpu;
187 }
188
189 void
cfs_set_ptldebug_header(struct ptldebug_header * header,struct libcfs_debug_msg_data * msgdata,unsigned long stack)190 cfs_set_ptldebug_header(struct ptldebug_header *header,
191 struct libcfs_debug_msg_data *msgdata,
192 unsigned long stack)
193 {
194 struct timespec64 ts;
195
196 ktime_get_real_ts64(&ts);
197
198 header->ph_subsys = msgdata->msg_subsys;
199 header->ph_mask = msgdata->msg_mask;
200 header->ph_cpu_id = smp_processor_id();
201 header->ph_type = cfs_trace_buf_idx_get();
202 /* y2038 safe since all user space treats this as unsigned, but
203 * will overflow in 2106 */
204 header->ph_sec = (u32)ts.tv_sec;
205 header->ph_usec = ts.tv_nsec / NSEC_PER_USEC;
206 header->ph_stack = stack;
207 header->ph_pid = current->pid;
208 header->ph_line_num = msgdata->msg_line;
209 header->ph_extern_pid = 0;
210 return;
211 }
212
213 static char *
dbghdr_to_err_string(struct ptldebug_header * hdr)214 dbghdr_to_err_string(struct ptldebug_header *hdr)
215 {
216 switch (hdr->ph_subsys) {
217 case S_LND:
218 case S_LNET:
219 return "LNetError";
220 default:
221 return "LustreError";
222 }
223 }
224
225 static char *
dbghdr_to_info_string(struct ptldebug_header * hdr)226 dbghdr_to_info_string(struct ptldebug_header *hdr)
227 {
228 switch (hdr->ph_subsys) {
229 case S_LND:
230 case S_LNET:
231 return "LNet";
232 default:
233 return "Lustre";
234 }
235 }
236
cfs_print_to_console(struct ptldebug_header * hdr,int mask,const char * buf,int len,const char * file,const char * fn)237 void cfs_print_to_console(struct ptldebug_header *hdr, int mask,
238 const char *buf, int len, const char *file,
239 const char *fn)
240 {
241 char *prefix = "Lustre", *ptype = NULL;
242
243 if ((mask & D_EMERG) != 0) {
244 prefix = dbghdr_to_err_string(hdr);
245 ptype = KERN_EMERG;
246 } else if ((mask & D_ERROR) != 0) {
247 prefix = dbghdr_to_err_string(hdr);
248 ptype = KERN_ERR;
249 } else if ((mask & D_WARNING) != 0) {
250 prefix = dbghdr_to_info_string(hdr);
251 ptype = KERN_WARNING;
252 } else if ((mask & (D_CONSOLE | libcfs_printk)) != 0) {
253 prefix = dbghdr_to_info_string(hdr);
254 ptype = KERN_INFO;
255 }
256
257 if ((mask & D_CONSOLE) != 0) {
258 printk("%s%s: %.*s", ptype, prefix, len, buf);
259 } else {
260 printk("%s%s: %d:%d:(%s:%d:%s()) %.*s", ptype, prefix,
261 hdr->ph_pid, hdr->ph_extern_pid, file, hdr->ph_line_num,
262 fn, len, buf);
263 }
264 return;
265 }
266
cfs_trace_max_debug_mb(void)267 int cfs_trace_max_debug_mb(void)
268 {
269 int total_mb = (totalram_pages >> (20 - PAGE_SHIFT));
270
271 return max(512, (total_mb * 80)/100);
272 }
273