4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 only,
8 * as published by the Free Software Foundation.
10 * This program is distributed in the hope that it will be useful, but
11 * WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * General Public License version 2 for more details (a copy is included
14 * in the LICENSE file that accompanied this code).
16 * You should have received a copy of the GNU General Public License
17 * version 2 along with this program; If not, see
18 * http://www.gnu.org/licenses/gpl-2.0.html
23 * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved.
24 * Use is subject to license terms.
26 * Copyright (c) 2012, Intel Corporation.
29 * This file is part of Lustre, http://www.lustre.org/
30 * Lustre is a trademark of Sun Microsystems, Inc.
33 #define DEBUG_SUBSYSTEM S_LNET
34 #define LUSTRE_TRACEFILE_PRIVATE
36 #include <linux/libcfs/libcfs.h>
37 #include "../tracefile.h"
39 /* percents to share the total debug memory for each type */
40 static unsigned int pages_factor[CFS_TCD_TYPE_MAX] = {
41 80, /* 80% pages for CFS_TCD_TYPE_PROC */
42 10, /* 10% pages for CFS_TCD_TYPE_SOFTIRQ */
43 10 /* 10% pages for CFS_TCD_TYPE_IRQ */
46 char *cfs_trace_console_buffers[NR_CPUS][CFS_TCD_TYPE_MAX];
48 static DECLARE_RWSEM(cfs_tracefile_sem);
50 int cfs_tracefile_init_arch(void)
54 struct cfs_trace_cpu_data *tcd;
56 /* initialize trace_data */
57 memset(cfs_trace_data, 0, sizeof(cfs_trace_data));
58 for (i = 0; i < CFS_TCD_TYPE_MAX; i++) {
60 kmalloc_array(num_possible_cpus(),
61 sizeof(union cfs_trace_data_union),
63 if (!cfs_trace_data[i])
67 /* arch related info initialized */
68 cfs_tcd_for_each(tcd, i, j) {
69 spin_lock_init(&tcd->tcd_lock);
70 tcd->tcd_pages_factor = pages_factor[i];
75 for (i = 0; i < num_possible_cpus(); i++)
76 for (j = 0; j < 3; j++) {
77 cfs_trace_console_buffers[i][j] =
78 kmalloc(CFS_TRACE_CONSOLE_BUFFER_SIZE,
81 if (!cfs_trace_console_buffers[i][j])
88 cfs_tracefile_fini_arch();
89 pr_err("lnet: Not enough memory\n");
93 void cfs_tracefile_fini_arch(void)
98 for (i = 0; i < num_possible_cpus(); i++)
99 for (j = 0; j < 3; j++) {
100 kfree(cfs_trace_console_buffers[i][j]);
101 cfs_trace_console_buffers[i][j] = NULL;
104 for (i = 0; cfs_trace_data[i]; i++) {
105 kfree(cfs_trace_data[i]);
106 cfs_trace_data[i] = NULL;
110 void cfs_tracefile_read_lock(void)
112 down_read(&cfs_tracefile_sem);
115 void cfs_tracefile_read_unlock(void)
117 up_read(&cfs_tracefile_sem);
120 void cfs_tracefile_write_lock(void)
122 down_write(&cfs_tracefile_sem);
125 void cfs_tracefile_write_unlock(void)
127 up_write(&cfs_tracefile_sem);
130 enum cfs_trace_buf_type cfs_trace_buf_idx_get(void)
133 return CFS_TCD_TYPE_IRQ;
135 return CFS_TCD_TYPE_SOFTIRQ;
136 return CFS_TCD_TYPE_PROC;
140 * The walking argument indicates the locking comes from all tcd types
141 * iterator and we must lock it and dissable local irqs to avoid deadlocks
142 * with other interrupt locks that might be happening. See LU-1311
145 int cfs_trace_lock_tcd(struct cfs_trace_cpu_data *tcd, int walking)
146 __acquires(&tcd->tc_lock)
148 __LASSERT(tcd->tcd_type < CFS_TCD_TYPE_MAX);
149 if (tcd->tcd_type == CFS_TCD_TYPE_IRQ)
150 spin_lock_irqsave(&tcd->tcd_lock, tcd->tcd_lock_flags);
151 else if (tcd->tcd_type == CFS_TCD_TYPE_SOFTIRQ)
152 spin_lock_bh(&tcd->tcd_lock);
153 else if (unlikely(walking))
154 spin_lock_irq(&tcd->tcd_lock);
156 spin_lock(&tcd->tcd_lock);
160 void cfs_trace_unlock_tcd(struct cfs_trace_cpu_data *tcd, int walking)
161 __releases(&tcd->tcd_lock)
163 __LASSERT(tcd->tcd_type < CFS_TCD_TYPE_MAX);
164 if (tcd->tcd_type == CFS_TCD_TYPE_IRQ)
165 spin_unlock_irqrestore(&tcd->tcd_lock, tcd->tcd_lock_flags);
166 else if (tcd->tcd_type == CFS_TCD_TYPE_SOFTIRQ)
167 spin_unlock_bh(&tcd->tcd_lock);
168 else if (unlikely(walking))
169 spin_unlock_irq(&tcd->tcd_lock);
171 spin_unlock(&tcd->tcd_lock);
175 cfs_set_ptldebug_header(struct ptldebug_header *header,
176 struct libcfs_debug_msg_data *msgdata,
179 struct timespec64 ts;
181 ktime_get_real_ts64(&ts);
183 header->ph_subsys = msgdata->msg_subsys;
184 header->ph_mask = msgdata->msg_mask;
185 header->ph_cpu_id = smp_processor_id();
186 header->ph_type = cfs_trace_buf_idx_get();
187 /* y2038 safe since all user space treats this as unsigned, but
188 * will overflow in 2106
190 header->ph_sec = (u32)ts.tv_sec;
191 header->ph_usec = ts.tv_nsec / NSEC_PER_USEC;
192 header->ph_stack = stack;
193 header->ph_pid = current->pid;
194 header->ph_line_num = msgdata->msg_line;
195 header->ph_extern_pid = 0;
199 dbghdr_to_err_string(struct ptldebug_header *hdr)
201 switch (hdr->ph_subsys) {
206 return "LustreError";
211 dbghdr_to_info_string(struct ptldebug_header *hdr)
213 switch (hdr->ph_subsys) {
222 void cfs_print_to_console(struct ptldebug_header *hdr, int mask,
223 const char *buf, int len, const char *file,
226 char *prefix = "Lustre", *ptype = NULL;
228 if (mask & D_EMERG) {
229 prefix = dbghdr_to_err_string(hdr);
231 } else if (mask & D_ERROR) {
232 prefix = dbghdr_to_err_string(hdr);
234 } else if (mask & D_WARNING) {
235 prefix = dbghdr_to_info_string(hdr);
236 ptype = KERN_WARNING;
237 } else if (mask & (D_CONSOLE | libcfs_printk)) {
238 prefix = dbghdr_to_info_string(hdr);
242 if (mask & D_CONSOLE) {
243 pr_info("%s%s: %.*s", ptype, prefix, len, buf);
245 pr_info("%s%s: %d:%d:(%s:%d:%s()) %.*s", ptype, prefix,
246 hdr->ph_pid, hdr->ph_extern_pid, file,
247 hdr->ph_line_num, fn, len, buf);
251 int cfs_trace_max_debug_mb(void)
253 int total_mb = (totalram_pages >> (20 - PAGE_SHIFT));
255 return max(512, (total_mb * 80) / 100);