result = IRQ_HANDLED;
/* Network control and best effort queue RX/TX */
- if (info->nc_queue) {
+ if (info->nc_queues) {
for (q = RAVB_NC; q >= RAVB_BE; q--) {
if (ravb_queue_interrupt(ndev, q))
result = IRQ_HANDLED;
/* Receive error message handling */
priv->rx_over_errors = priv->stats[RAVB_BE].rx_over_errors;
- if (info->nc_queue)
+ if (info->nc_queues)
priv->rx_over_errors += priv->stats[RAVB_NC].rx_over_errors;
if (priv->rx_over_errors != ndev->stats.rx_over_errors)
ndev->stats.rx_over_errors = priv->rx_over_errors;
int i = 0;
int q;
- num_rx_q = info->nc_queue ? NUM_RX_QUEUE : 1;
+ num_rx_q = info->nc_queues ? NUM_RX_QUEUE : 1;
/* Device-specific stats */
for (q = RAVB_BE; q < num_rx_q; q++) {
struct net_device_stats *stats = &priv->stats[q];
/* Free all the skb's in the RX queue and the DMA buffers. */
ravb_ring_free(ndev, RAVB_BE);
- if (info->nc_queue)
+ if (info->nc_queues)
ravb_ring_free(ndev, RAVB_NC);
}
int error;
napi_enable(&priv->napi[RAVB_BE]);
- if (info->nc_queue)
+ if (info->nc_queues)
napi_enable(&priv->napi[RAVB_NC]);
if (!info->multi_irqs) {
out_free_irq:
free_irq(ndev->irq, ndev);
out_napi_off:
- if (info->nc_queue)
+ if (info->nc_queues)
napi_disable(&priv->napi[RAVB_NC]);
napi_disable(&priv->napi[RAVB_BE]);
return error;
}
ravb_ring_free(ndev, RAVB_BE);
- if (info->nc_queue)
+ if (info->nc_queues)
ravb_ring_free(ndev, RAVB_NC);
/* Device init */
nstats->rx_length_errors = stats0->rx_length_errors;
nstats->rx_missed_errors = stats0->rx_missed_errors;
nstats->rx_over_errors = stats0->rx_over_errors;
- if (info->nc_queue) {
+ if (info->nc_queues) {
stats1 = &priv->stats[RAVB_NC];
nstats->rx_packets += stats1->rx_packets;
}
free_irq(ndev->irq, ndev);
- if (info->nc_queue)
+ if (info->nc_queues)
napi_disable(&priv->napi[RAVB_NC]);
napi_disable(&priv->napi[RAVB_BE]);
/* Free all the skb's in the RX queue and the DMA buffers. */
ravb_ring_free(ndev, RAVB_BE);
- if (info->nc_queue)
+ if (info->nc_queues)
ravb_ring_free(ndev, RAVB_NC);
return 0;
.tx_counters = 1,
.multi_irqs = 1,
.ccc_gac = 1,
- .nc_queue = 1,
+ .nc_queues = 1,
.magic_pkt = 1,
};
.rx_max_buf_size = SZ_2K,
.aligned_tx = 1,
.gptp = 1,
- .nc_queue = 1,
+ .nc_queues = 1,
.magic_pkt = 1,
};
priv->pdev = pdev;
priv->num_tx_ring[RAVB_BE] = BE_TX_RING_SIZE;
priv->num_rx_ring[RAVB_BE] = BE_RX_RING_SIZE;
- if (info->nc_queue) {
+ if (info->nc_queues) {
priv->num_tx_ring[RAVB_NC] = NC_TX_RING_SIZE;
priv->num_rx_ring[RAVB_NC] = NC_RX_RING_SIZE;
}
}
netif_napi_add(ndev, &priv->napi[RAVB_BE], ravb_poll, 64);
- if (info->nc_queue)
+ if (info->nc_queues)
netif_napi_add(ndev, &priv->napi[RAVB_NC], ravb_poll, 64);
/* Network device register */
return 0;
out_napi_del:
- if (info->nc_queue)
+ if (info->nc_queues)
netif_napi_del(&priv->napi[RAVB_NC]);
netif_napi_del(&priv->napi[RAVB_BE]);
ravb_write(ndev, CCC_OPC_RESET, CCC);
pm_runtime_put_sync(&pdev->dev);
unregister_netdev(ndev);
- if (info->nc_queue)
+ if (info->nc_queues)
netif_napi_del(&priv->napi[RAVB_NC]);
netif_napi_del(&priv->napi[RAVB_BE]);
ravb_mdio_release(priv);
/* Only allow ECI interrupts */
synchronize_irq(priv->emac_irq);
- if (info->nc_queue)
+ if (info->nc_queues)
napi_disable(&priv->napi[RAVB_NC]);
napi_disable(&priv->napi[RAVB_BE]);
ravb_write(ndev, ECSIPR_MPDIP, ECSIPR);
const struct ravb_hw_info *info = priv->info;
int ret;
- if (info->nc_queue)
+ if (info->nc_queues)
napi_enable(&priv->napi[RAVB_NC]);
napi_enable(&priv->napi[RAVB_BE]);