1 // SPDX-License-Identifier: GPL-2.0-only
3 * QLogic Fibre Channel HBA Driver
4 * Copyright (c) 2003-2014 QLogic Corporation
8 #include <linux/debugfs.h>
9 #include <linux/seq_file.h>
11 static struct dentry *qla2x00_dfs_root;
12 static atomic_t qla2x00_dfs_root_count;
14 #define QLA_DFS_RPORT_DEVLOSS_TMO 1
17 qla_dfs_rport_get(struct fc_port *fp, int attr_id, u64 *val)
20 case QLA_DFS_RPORT_DEVLOSS_TMO:
21 /* Only supported for FC-NVMe devices that are registered. */
22 if (!(fp->nvme_flag & NVME_FLAG_REGISTERED))
24 *val = fp->nvme_remote_port->dev_loss_tmo;
33 qla_dfs_rport_set(struct fc_port *fp, int attr_id, u64 val)
36 case QLA_DFS_RPORT_DEVLOSS_TMO:
37 /* Only supported for FC-NVMe devices that are registered. */
38 if (!(fp->nvme_flag & NVME_FLAG_REGISTERED))
40 #if (IS_ENABLED(CONFIG_NVME_FC))
41 return nvme_fc_set_remoteport_devloss(fp->nvme_remote_port,
43 #else /* CONFIG_NVME_FC */
45 #endif /* CONFIG_NVME_FC */
52 #define DEFINE_QLA_DFS_RPORT_RW_ATTR(_attr_id, _attr) \
53 static int qla_dfs_rport_##_attr##_get(void *data, u64 *val) \
55 struct fc_port *fp = data; \
56 return qla_dfs_rport_get(fp, _attr_id, val); \
58 static int qla_dfs_rport_##_attr##_set(void *data, u64 val) \
60 struct fc_port *fp = data; \
61 return qla_dfs_rport_set(fp, _attr_id, val); \
63 DEFINE_DEBUGFS_ATTRIBUTE(qla_dfs_rport_##_attr##_fops, \
64 qla_dfs_rport_##_attr##_get, \
65 qla_dfs_rport_##_attr##_set, "%llu\n")
68 * Wrapper for getting fc_port fields.
70 * _attr : Attribute name.
71 * _get_val : Accessor macro to retrieve the value.
73 #define DEFINE_QLA_DFS_RPORT_FIELD_GET(_attr, _get_val) \
74 static int qla_dfs_rport_field_##_attr##_get(void *data, u64 *val) \
76 struct fc_port *fp = data; \
80 DEFINE_DEBUGFS_ATTRIBUTE(qla_dfs_rport_field_##_attr##_fops, \
81 qla_dfs_rport_field_##_attr##_get, \
84 #define DEFINE_QLA_DFS_RPORT_ACCESS(_attr, _get_val) \
85 DEFINE_QLA_DFS_RPORT_FIELD_GET(_attr, _get_val)
87 #define DEFINE_QLA_DFS_RPORT_FIELD(_attr) \
88 DEFINE_QLA_DFS_RPORT_FIELD_GET(_attr, fp->_attr)
90 DEFINE_QLA_DFS_RPORT_RW_ATTR(QLA_DFS_RPORT_DEVLOSS_TMO, dev_loss_tmo);
92 DEFINE_QLA_DFS_RPORT_FIELD(disc_state);
93 DEFINE_QLA_DFS_RPORT_FIELD(scan_state);
94 DEFINE_QLA_DFS_RPORT_FIELD(fw_login_state);
95 DEFINE_QLA_DFS_RPORT_FIELD(login_pause);
96 DEFINE_QLA_DFS_RPORT_FIELD(flags);
97 DEFINE_QLA_DFS_RPORT_FIELD(nvme_flag);
98 DEFINE_QLA_DFS_RPORT_FIELD(last_rscn_gen);
99 DEFINE_QLA_DFS_RPORT_FIELD(rscn_gen);
100 DEFINE_QLA_DFS_RPORT_FIELD(login_gen);
101 DEFINE_QLA_DFS_RPORT_FIELD(loop_id);
102 DEFINE_QLA_DFS_RPORT_FIELD_GET(port_id, fp->d_id.b24);
103 DEFINE_QLA_DFS_RPORT_FIELD_GET(sess_kref, kref_read(&fp->sess_kref));
106 qla2x00_dfs_create_rport(scsi_qla_host_t *vha, struct fc_port *fp)
110 #define QLA_CREATE_RPORT_FIELD_ATTR(_attr) \
111 debugfs_create_file(#_attr, 0400, fp->dfs_rport_dir, \
112 fp, &qla_dfs_rport_field_##_attr##_fops)
114 if (!vha->dfs_rport_root || fp->dfs_rport_dir)
117 sprintf(wwn, "pn-%016llx", wwn_to_u64(fp->port_name));
118 fp->dfs_rport_dir = debugfs_create_dir(wwn, vha->dfs_rport_root);
119 if (!fp->dfs_rport_dir)
121 if (NVME_TARGET(vha->hw, fp))
122 debugfs_create_file("dev_loss_tmo", 0600, fp->dfs_rport_dir,
123 fp, &qla_dfs_rport_dev_loss_tmo_fops);
125 QLA_CREATE_RPORT_FIELD_ATTR(disc_state);
126 QLA_CREATE_RPORT_FIELD_ATTR(scan_state);
127 QLA_CREATE_RPORT_FIELD_ATTR(fw_login_state);
128 QLA_CREATE_RPORT_FIELD_ATTR(login_pause);
129 QLA_CREATE_RPORT_FIELD_ATTR(flags);
130 QLA_CREATE_RPORT_FIELD_ATTR(nvme_flag);
131 QLA_CREATE_RPORT_FIELD_ATTR(last_rscn_gen);
132 QLA_CREATE_RPORT_FIELD_ATTR(rscn_gen);
133 QLA_CREATE_RPORT_FIELD_ATTR(login_gen);
134 QLA_CREATE_RPORT_FIELD_ATTR(loop_id);
135 QLA_CREATE_RPORT_FIELD_ATTR(port_id);
136 QLA_CREATE_RPORT_FIELD_ATTR(sess_kref);
140 qla2x00_dfs_remove_rport(scsi_qla_host_t *vha, struct fc_port *fp)
142 if (!vha->dfs_rport_root || !fp->dfs_rport_dir)
144 debugfs_remove_recursive(fp->dfs_rport_dir);
145 fp->dfs_rport_dir = NULL;
149 qla2x00_dfs_tgt_sess_show(struct seq_file *s, void *unused)
151 scsi_qla_host_t *vha = s->private;
152 struct qla_hw_data *ha = vha->hw;
154 struct fc_port *sess = NULL;
155 struct qla_tgt *tgt = vha->vha_tgt.qla_tgt;
157 seq_printf(s, "%s\n", vha->host_str);
159 seq_puts(s, "Port ID Port Name Handle\n");
161 spin_lock_irqsave(&ha->tgt.sess_lock, flags);
162 list_for_each_entry(sess, &vha->vp_fcports, list)
163 seq_printf(s, "%02x:%02x:%02x %8phC %d\n",
164 sess->d_id.b.domain, sess->d_id.b.area,
165 sess->d_id.b.al_pa, sess->port_name,
167 spin_unlock_irqrestore(&ha->tgt.sess_lock, flags);
173 DEFINE_SHOW_ATTRIBUTE(qla2x00_dfs_tgt_sess);
176 qla2x00_dfs_tgt_port_database_show(struct seq_file *s, void *unused)
178 scsi_qla_host_t *vha = s->private;
179 struct qla_hw_data *ha = vha->hw;
180 struct gid_list_info *gid_list;
181 dma_addr_t gid_list_dma;
185 uint16_t entries, loop_id;
187 seq_printf(s, "%s\n", vha->host_str);
188 gid_list = dma_alloc_coherent(&ha->pdev->dev,
189 qla2x00_gid_list_size(ha),
190 &gid_list_dma, GFP_KERNEL);
192 ql_dbg(ql_dbg_user, vha, 0x7018,
193 "DMA allocation failed for %u\n",
194 qla2x00_gid_list_size(ha));
198 rc = qla24xx_gidlist_wait(vha, gid_list, gid_list_dma,
200 if (rc != QLA_SUCCESS)
201 goto out_free_id_list;
203 id_iter = (char *)gid_list;
205 seq_puts(s, "Port Name Port ID Loop ID\n");
207 for (i = 0; i < entries; i++) {
208 struct gid_list_info *gid =
209 (struct gid_list_info *)id_iter;
210 loop_id = le16_to_cpu(gid->loop_id);
211 memset(&fc_port, 0, sizeof(fc_port_t));
213 fc_port.loop_id = loop_id;
215 rc = qla24xx_gpdb_wait(vha, &fc_port, 0);
216 seq_printf(s, "%8phC %02x%02x%02x %d\n",
217 fc_port.port_name, fc_port.d_id.b.domain,
218 fc_port.d_id.b.area, fc_port.d_id.b.al_pa,
220 id_iter += ha->gid_list_info_size;
223 dma_free_coherent(&ha->pdev->dev, qla2x00_gid_list_size(ha),
224 gid_list, gid_list_dma);
229 DEFINE_SHOW_ATTRIBUTE(qla2x00_dfs_tgt_port_database);
232 qla_dfs_fw_resource_cnt_show(struct seq_file *s, void *unused)
234 struct scsi_qla_host *vha = s->private;
235 uint16_t mb[MAX_IOCB_MB_REG];
237 struct qla_hw_data *ha = vha->hw;
238 u16 iocbs_used, i, exch_used;
240 rc = qla24xx_res_count_wait(vha, mb, SIZEOF_IOCB_MB_REG);
241 if (rc != QLA_SUCCESS) {
242 seq_printf(s, "Mailbox Command failed %d, mb %#x", rc, mb[0]);
244 seq_puts(s, "FW Resource count\n\n");
245 seq_printf(s, "Original TGT exchg count[%d]\n", mb[1]);
246 seq_printf(s, "Current TGT exchg count[%d]\n", mb[2]);
247 seq_printf(s, "Current Initiator Exchange count[%d]\n", mb[3]);
248 seq_printf(s, "Original Initiator Exchange count[%d]\n", mb[6]);
249 seq_printf(s, "Current IOCB count[%d]\n", mb[7]);
250 seq_printf(s, "Original IOCB count[%d]\n", mb[10]);
251 seq_printf(s, "MAX VP count[%d]\n", mb[11]);
252 seq_printf(s, "MAX FCF count[%d]\n", mb[12]);
253 seq_printf(s, "Current free pageable XCB buffer cnt[%d]\n",
255 seq_printf(s, "Original Initiator fast XCB buffer cnt[%d]\n",
257 seq_printf(s, "Current free Initiator fast XCB buffer cnt[%d]\n",
259 seq_printf(s, "Original Target fast XCB buffer cnt[%d]\n",
263 if (ql2xenforce_iocb_limit) {
264 /* lock is not require. It's an estimate. */
265 iocbs_used = ha->base_qpair->fwres.iocbs_used;
266 exch_used = ha->base_qpair->fwres.exch_used;
267 for (i = 0; i < ha->max_qpairs; i++) {
268 if (ha->queue_pair_map[i]) {
269 iocbs_used += ha->queue_pair_map[i]->fwres.iocbs_used;
270 exch_used += ha->queue_pair_map[i]->fwres.exch_used;
274 seq_printf(s, "Driver: estimate iocb used [%d] high water limit [%d]\n",
275 iocbs_used, ha->base_qpair->fwres.iocbs_limit);
277 seq_printf(s, "estimate exchange used[%d] high water limit [%d] n",
278 exch_used, ha->base_qpair->fwres.exch_limit);
284 DEFINE_SHOW_ATTRIBUTE(qla_dfs_fw_resource_cnt);
287 qla_dfs_tgt_counters_show(struct seq_file *s, void *unused)
289 struct scsi_qla_host *vha = s->private;
290 struct qla_qpair *qpair = vha->hw->base_qpair;
291 uint64_t qla_core_sbt_cmd, core_qla_que_buf, qla_core_ret_ctio,
292 core_qla_snd_status, qla_core_ret_sta_ctio, core_qla_free_cmd,
293 num_q_full_sent, num_alloc_iocb_failed, num_term_xchg_sent;
295 fc_port_t *fcport = NULL;
297 if (qla2x00_chip_is_down(vha))
300 qla_core_sbt_cmd = qpair->tgt_counters.qla_core_sbt_cmd;
301 core_qla_que_buf = qpair->tgt_counters.core_qla_que_buf;
302 qla_core_ret_ctio = qpair->tgt_counters.qla_core_ret_ctio;
303 core_qla_snd_status = qpair->tgt_counters.core_qla_snd_status;
304 qla_core_ret_sta_ctio = qpair->tgt_counters.qla_core_ret_sta_ctio;
305 core_qla_free_cmd = qpair->tgt_counters.core_qla_free_cmd;
306 num_q_full_sent = qpair->tgt_counters.num_q_full_sent;
307 num_alloc_iocb_failed = qpair->tgt_counters.num_alloc_iocb_failed;
308 num_term_xchg_sent = qpair->tgt_counters.num_term_xchg_sent;
310 for (i = 0; i < vha->hw->max_qpairs; i++) {
311 qpair = vha->hw->queue_pair_map[i];
314 qla_core_sbt_cmd += qpair->tgt_counters.qla_core_sbt_cmd;
315 core_qla_que_buf += qpair->tgt_counters.core_qla_que_buf;
316 qla_core_ret_ctio += qpair->tgt_counters.qla_core_ret_ctio;
317 core_qla_snd_status += qpair->tgt_counters.core_qla_snd_status;
318 qla_core_ret_sta_ctio +=
319 qpair->tgt_counters.qla_core_ret_sta_ctio;
320 core_qla_free_cmd += qpair->tgt_counters.core_qla_free_cmd;
321 num_q_full_sent += qpair->tgt_counters.num_q_full_sent;
322 num_alloc_iocb_failed +=
323 qpair->tgt_counters.num_alloc_iocb_failed;
324 num_term_xchg_sent += qpair->tgt_counters.num_term_xchg_sent;
327 seq_puts(s, "Target Counters\n");
328 seq_printf(s, "qla_core_sbt_cmd = %lld\n",
330 seq_printf(s, "qla_core_ret_sta_ctio = %lld\n",
331 qla_core_ret_sta_ctio);
332 seq_printf(s, "qla_core_ret_ctio = %lld\n",
334 seq_printf(s, "core_qla_que_buf = %lld\n",
336 seq_printf(s, "core_qla_snd_status = %lld\n",
337 core_qla_snd_status);
338 seq_printf(s, "core_qla_free_cmd = %lld\n",
340 seq_printf(s, "num alloc iocb failed = %lld\n",
341 num_alloc_iocb_failed);
342 seq_printf(s, "num term exchange sent = %lld\n",
344 seq_printf(s, "num Q full sent = %lld\n",
348 seq_printf(s, "DIF Inp Bytes = %lld\n",
349 vha->qla_stats.qla_dif_stats.dif_input_bytes);
350 seq_printf(s, "DIF Outp Bytes = %lld\n",
351 vha->qla_stats.qla_dif_stats.dif_output_bytes);
352 seq_printf(s, "DIF Inp Req = %lld\n",
353 vha->qla_stats.qla_dif_stats.dif_input_requests);
354 seq_printf(s, "DIF Outp Req = %lld\n",
355 vha->qla_stats.qla_dif_stats.dif_output_requests);
356 seq_printf(s, "DIF Guard err = %d\n",
357 vha->qla_stats.qla_dif_stats.dif_guard_err);
358 seq_printf(s, "DIF Ref tag err = %d\n",
359 vha->qla_stats.qla_dif_stats.dif_ref_tag_err);
360 seq_printf(s, "DIF App tag err = %d\n",
361 vha->qla_stats.qla_dif_stats.dif_app_tag_err);
364 seq_puts(s, "Initiator Error Counters\n");
365 seq_printf(s, "HW Error Count = %14lld\n",
367 seq_printf(s, "Link Down Count = %14lld\n",
368 vha->short_link_down_cnt);
369 seq_printf(s, "Interface Err Count = %14lld\n",
370 vha->interface_err_cnt);
371 seq_printf(s, "Cmd Timeout Count = %14lld\n",
372 vha->cmd_timeout_cnt);
373 seq_printf(s, "Reset Count = %14lld\n",
374 vha->reset_cmd_err_cnt);
377 list_for_each_entry(fcport, &vha->vp_fcports, list) {
381 seq_printf(s, "Target Num = %7d Link Down Count = %14lld\n",
382 fcport->rport->number, fcport->tgt_short_link_down_cnt);
389 DEFINE_SHOW_ATTRIBUTE(qla_dfs_tgt_counters);
392 qla2x00_dfs_fce_show(struct seq_file *s, void *unused)
394 scsi_qla_host_t *vha = s->private;
398 struct qla_hw_data *ha = vha->hw;
400 mutex_lock(&ha->fce_mutex);
402 seq_puts(s, "FCE Trace Buffer\n");
403 seq_printf(s, "In Pointer = %llx\n\n", (unsigned long long)ha->fce_wr);
404 seq_printf(s, "Base = %llx\n\n", (unsigned long long) ha->fce_dma);
405 seq_puts(s, "FCE Enable Registers\n");
406 seq_printf(s, "%08x %08x %08x %08x %08x %08x\n",
407 ha->fce_mb[0], ha->fce_mb[2], ha->fce_mb[3], ha->fce_mb[4],
408 ha->fce_mb[5], ha->fce_mb[6]);
410 fce = (uint32_t *) ha->fce;
411 fce_start = (unsigned long long) ha->fce_dma;
412 for (cnt = 0; cnt < fce_calc_size(ha->fce_bufs) / 4; cnt++) {
414 seq_printf(s, "\n%llx: ",
415 (unsigned long long)((cnt * 4) + fce_start));
418 seq_printf(s, "%08x", *fce++);
421 seq_puts(s, "\nEnd\n");
423 mutex_unlock(&ha->fce_mutex);
429 qla2x00_dfs_fce_open(struct inode *inode, struct file *file)
431 scsi_qla_host_t *vha = inode->i_private;
432 struct qla_hw_data *ha = vha->hw;
435 if (!ha->flags.fce_enabled)
438 mutex_lock(&ha->fce_mutex);
440 /* Pause tracing to flush FCE buffers. */
441 rval = qla2x00_disable_fce_trace(vha, &ha->fce_wr, &ha->fce_rd);
443 ql_dbg(ql_dbg_user, vha, 0x705c,
444 "DebugFS: Unable to disable FCE (%d).\n", rval);
446 ha->flags.fce_enabled = 0;
448 mutex_unlock(&ha->fce_mutex);
450 return single_open(file, qla2x00_dfs_fce_show, vha);
454 qla2x00_dfs_fce_release(struct inode *inode, struct file *file)
456 scsi_qla_host_t *vha = inode->i_private;
457 struct qla_hw_data *ha = vha->hw;
460 if (ha->flags.fce_enabled)
463 mutex_lock(&ha->fce_mutex);
465 /* Re-enable FCE tracing. */
466 ha->flags.fce_enabled = 1;
467 memset(ha->fce, 0, fce_calc_size(ha->fce_bufs));
468 rval = qla2x00_enable_fce_trace(vha, ha->fce_dma, ha->fce_bufs,
469 ha->fce_mb, &ha->fce_bufs);
471 ql_dbg(ql_dbg_user, vha, 0x700d,
472 "DebugFS: Unable to reinitialize FCE (%d).\n", rval);
473 ha->flags.fce_enabled = 0;
476 mutex_unlock(&ha->fce_mutex);
478 return single_release(inode, file);
481 static const struct file_operations dfs_fce_ops = {
482 .open = qla2x00_dfs_fce_open,
485 .release = qla2x00_dfs_fce_release,
489 qla_dfs_naqp_show(struct seq_file *s, void *unused)
491 struct scsi_qla_host *vha = s->private;
492 struct qla_hw_data *ha = vha->hw;
494 seq_printf(s, "%d\n", ha->tgt.num_act_qpairs);
499 * Helper macros for setting up debugfs entries.
500 * _name: The name of the debugfs entry
501 * _ctx_struct: The context that was passed when creating the debugfs file
503 * QLA_DFS_SETUP_RD could be used when there is only a show function.
504 * - show function take the name qla_dfs_<sysfs-name>_show
506 * QLA_DFS_SETUP_RW could be used when there are both show and write functions.
507 * - show function take the name qla_dfs_<sysfs-name>_show
508 * - write function take the name qla_dfs_<sysfs-name>_write
510 * To have a new debugfs entry, do:
511 * 1. Create a "struct dentry *" in the appropriate structure in the format
513 * 2. Setup debugfs entries using QLA_DFS_SETUP_RD / QLA_DFS_SETUP_RW
514 * 3. Create debugfs file in qla2x00_dfs_setup() using QLA_DFS_CREATE_FILE
515 * or QLA_DFS_ROOT_CREATE_FILE
516 * 4. Remove debugfs file in qla2x00_dfs_remove() using QLA_DFS_REMOVE_FILE
517 * or QLA_DFS_ROOT_REMOVE_FILE
519 * Example for creating "TEST" sysfs file:
520 * 1. struct qla_hw_data { ... struct dentry *dfs_TEST; }
521 * 2. QLA_DFS_SETUP_RD(TEST, scsi_qla_host_t);
522 * 3. In qla2x00_dfs_setup():
523 * QLA_DFS_CREATE_FILE(ha, TEST, 0600, ha->dfs_dir, vha);
524 * 4. In qla2x00_dfs_remove():
525 * QLA_DFS_REMOVE_FILE(ha, TEST);
527 #define QLA_DFS_SETUP_RD(_name, _ctx_struct) \
529 qla_dfs_##_name##_open(struct inode *inode, struct file *file) \
531 _ctx_struct *__ctx = inode->i_private; \
533 return single_open(file, qla_dfs_##_name##_show, __ctx); \
536 static const struct file_operations qla_dfs_##_name##_ops = { \
537 .open = qla_dfs_##_name##_open, \
539 .llseek = seq_lseek, \
540 .release = single_release, \
543 #define QLA_DFS_SETUP_RW(_name, _ctx_struct) \
545 qla_dfs_##_name##_open(struct inode *inode, struct file *file) \
547 _ctx_struct *__ctx = inode->i_private; \
549 return single_open(file, qla_dfs_##_name##_show, __ctx); \
552 static const struct file_operations qla_dfs_##_name##_ops = { \
553 .open = qla_dfs_##_name##_open, \
555 .llseek = seq_lseek, \
556 .release = single_release, \
557 .write = qla_dfs_##_name##_write, \
560 #define QLA_DFS_ROOT_CREATE_FILE(_name, _perm, _ctx) \
562 if (!qla_dfs_##_name) \
563 qla_dfs_##_name = debugfs_create_file(#_name, \
564 _perm, qla2x00_dfs_root, _ctx, \
565 &qla_dfs_##_name##_ops); \
568 #define QLA_DFS_ROOT_REMOVE_FILE(_name) \
570 if (qla_dfs_##_name) { \
571 debugfs_remove(qla_dfs_##_name); \
572 qla_dfs_##_name = NULL; \
576 #define QLA_DFS_CREATE_FILE(_struct, _name, _perm, _parent, _ctx) \
578 (_struct)->dfs_##_name = debugfs_create_file(#_name, \
579 _perm, _parent, _ctx, \
580 &qla_dfs_##_name##_ops) \
583 #define QLA_DFS_REMOVE_FILE(_struct, _name) \
585 if ((_struct)->dfs_##_name) { \
586 debugfs_remove((_struct)->dfs_##_name); \
587 (_struct)->dfs_##_name = NULL; \
592 qla_dfs_naqp_open(struct inode *inode, struct file *file)
594 struct scsi_qla_host *vha = inode->i_private;
596 return single_open(file, qla_dfs_naqp_show, vha);
600 qla_dfs_naqp_write(struct file *file, const char __user *buffer,
601 size_t count, loff_t *pos)
603 struct seq_file *s = file->private_data;
604 struct scsi_qla_host *vha = s->private;
605 struct qla_hw_data *ha = vha->hw;
608 unsigned long num_act_qp;
610 if (!(IS_QLA27XX(ha) || IS_QLA83XX(ha) || IS_QLA28XX(ha))) {
611 pr_err("host%ld: this adapter does not support Multi Q.",
616 if (!vha->flags.qpairs_available) {
617 pr_err("host%ld: Driver is not setup with Multi Q.",
621 buf = memdup_user_nul(buffer, count);
623 pr_err("host%ld: fail to copy user buffer.",
628 num_act_qp = simple_strtoul(buf, NULL, 0);
630 if (num_act_qp >= vha->hw->max_qpairs) {
631 pr_err("User set invalid number of qpairs %lu. Max = %d",
632 num_act_qp, vha->hw->max_qpairs);
637 if (num_act_qp != ha->tgt.num_act_qpairs) {
638 ha->tgt.num_act_qpairs = num_act_qp;
639 qlt_clr_qp_table(vha);
647 static const struct file_operations dfs_naqp_ops = {
648 .open = qla_dfs_naqp_open,
651 .release = single_release,
652 .write = qla_dfs_naqp_write,
657 qla2x00_dfs_setup(scsi_qla_host_t *vha)
659 struct qla_hw_data *ha = vha->hw;
661 if (!IS_QLA25XX(ha) && !IS_QLA81XX(ha) && !IS_QLA83XX(ha) &&
662 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
667 if (qla2x00_dfs_root)
670 atomic_set(&qla2x00_dfs_root_count, 0);
671 qla2x00_dfs_root = debugfs_create_dir(QLA2XXX_DRIVER_NAME, NULL);
677 mutex_init(&ha->fce_mutex);
678 ha->dfs_dir = debugfs_create_dir(vha->host_str, qla2x00_dfs_root);
680 atomic_inc(&qla2x00_dfs_root_count);
683 ha->dfs_fw_resource_cnt = debugfs_create_file("fw_resource_count",
684 S_IRUSR, ha->dfs_dir, vha, &qla_dfs_fw_resource_cnt_fops);
686 ha->dfs_tgt_counters = debugfs_create_file("tgt_counters", S_IRUSR,
687 ha->dfs_dir, vha, &qla_dfs_tgt_counters_fops);
689 ha->tgt.dfs_tgt_port_database = debugfs_create_file("tgt_port_database",
690 S_IRUSR, ha->dfs_dir, vha, &qla2x00_dfs_tgt_port_database_fops);
692 ha->dfs_fce = debugfs_create_file("fce", S_IRUSR, ha->dfs_dir, vha,
695 ha->tgt.dfs_tgt_sess = debugfs_create_file("tgt_sess",
696 S_IRUSR, ha->dfs_dir, vha, &qla2x00_dfs_tgt_sess_fops);
698 if (IS_QLA27XX(ha) || IS_QLA83XX(ha) || IS_QLA28XX(ha)) {
699 ha->tgt.dfs_naqp = debugfs_create_file("naqp",
700 0400, ha->dfs_dir, vha, &dfs_naqp_ops);
701 if (!ha->tgt.dfs_naqp) {
702 ql_log(ql_log_warn, vha, 0xd011,
703 "Unable to create debugFS naqp node.\n");
707 vha->dfs_rport_root = debugfs_create_dir("rports", ha->dfs_dir);
708 if (!vha->dfs_rport_root) {
709 ql_log(ql_log_warn, vha, 0xd012,
710 "Unable to create debugFS rports node.\n");
718 qla2x00_dfs_remove(scsi_qla_host_t *vha)
720 struct qla_hw_data *ha = vha->hw;
722 if (ha->tgt.dfs_naqp) {
723 debugfs_remove(ha->tgt.dfs_naqp);
724 ha->tgt.dfs_naqp = NULL;
727 if (ha->tgt.dfs_tgt_sess) {
728 debugfs_remove(ha->tgt.dfs_tgt_sess);
729 ha->tgt.dfs_tgt_sess = NULL;
732 if (ha->tgt.dfs_tgt_port_database) {
733 debugfs_remove(ha->tgt.dfs_tgt_port_database);
734 ha->tgt.dfs_tgt_port_database = NULL;
737 if (ha->dfs_fw_resource_cnt) {
738 debugfs_remove(ha->dfs_fw_resource_cnt);
739 ha->dfs_fw_resource_cnt = NULL;
742 if (ha->dfs_tgt_counters) {
743 debugfs_remove(ha->dfs_tgt_counters);
744 ha->dfs_tgt_counters = NULL;
748 debugfs_remove(ha->dfs_fce);
752 if (vha->dfs_rport_root) {
753 debugfs_remove_recursive(vha->dfs_rport_root);
754 vha->dfs_rport_root = NULL;
758 debugfs_remove(ha->dfs_dir);
760 atomic_dec(&qla2x00_dfs_root_count);
763 if (atomic_read(&qla2x00_dfs_root_count) == 0 &&
765 debugfs_remove(qla2x00_dfs_root);
766 qla2x00_dfs_root = NULL;