mirror of
https://github.com/torvalds/linux.git
synced 2024-11-15 16:41:58 +00:00
bna: Function name changes and cleanups
Renamed following functions: bnad_cleanup_tx to bnad_destroy_tx bnad_free_all_txbufs to bnad_txq_cleanup bnad_free_txbufs to bnad_txcmpl_process bnad_tx to bnad_tx_complete bnad_cleanup_rx to bnad_destroy_rx bnad_reset_rcb to bnad_rcb_cleanup bnad_free_all_rxbufs to bnad_rxq_cleanup bnad_cq_cmpl_init to bnad_cq_cleanup bnad_alloc_n_post_rxbufs to bnad_rxq_post bnad_poll_cq to bnad_cq_process Signed-off-by: Jing Huang <huangj@brocade.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
d95d1081c3
commit
b3cc6e88d5
@ -101,7 +101,7 @@ bnad_remove_from_list(struct bnad *bnad)
|
|||||||
* Reinitialize completions in CQ, once Rx is taken down
|
* Reinitialize completions in CQ, once Rx is taken down
|
||||||
*/
|
*/
|
||||||
static void
|
static void
|
||||||
bnad_cq_cmpl_init(struct bnad *bnad, struct bna_ccb *ccb)
|
bnad_cq_cleanup(struct bnad *bnad, struct bna_ccb *ccb)
|
||||||
{
|
{
|
||||||
struct bna_cq_entry *cmpl, *next_cmpl;
|
struct bna_cq_entry *cmpl, *next_cmpl;
|
||||||
unsigned int wi_range, wis = 0, ccb_prod = 0;
|
unsigned int wi_range, wis = 0, ccb_prod = 0;
|
||||||
@ -154,7 +154,7 @@ bnad_pci_unmap_skb(struct device *pdev, struct bnad_skb_unmap *array,
|
|||||||
* so DMA unmap & freeing is fine.
|
* so DMA unmap & freeing is fine.
|
||||||
*/
|
*/
|
||||||
static void
|
static void
|
||||||
bnad_free_all_txbufs(struct bnad *bnad,
|
bnad_txq_cleanup(struct bnad *bnad,
|
||||||
struct bna_tcb *tcb)
|
struct bna_tcb *tcb)
|
||||||
{
|
{
|
||||||
u32 unmap_cons;
|
u32 unmap_cons;
|
||||||
@ -182,12 +182,12 @@ bnad_free_all_txbufs(struct bnad *bnad,
|
|||||||
/* Data Path Handlers */
|
/* Data Path Handlers */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* bnad_free_txbufs : Frees the Tx bufs on Tx completion
|
* bnad_txcmpl_process : Frees the Tx bufs on Tx completion
|
||||||
* Can be called in a) Interrupt context
|
* Can be called in a) Interrupt context
|
||||||
* b) Sending context
|
* b) Sending context
|
||||||
*/
|
*/
|
||||||
static u32
|
static u32
|
||||||
bnad_free_txbufs(struct bnad *bnad,
|
bnad_txcmpl_process(struct bnad *bnad,
|
||||||
struct bna_tcb *tcb)
|
struct bna_tcb *tcb)
|
||||||
{
|
{
|
||||||
u32 unmap_cons, sent_packets = 0, sent_bytes = 0;
|
u32 unmap_cons, sent_packets = 0, sent_bytes = 0;
|
||||||
@ -236,7 +236,7 @@ bnad_free_txbufs(struct bnad *bnad,
|
|||||||
}
|
}
|
||||||
|
|
||||||
static u32
|
static u32
|
||||||
bnad_tx(struct bnad *bnad, struct bna_tcb *tcb)
|
bnad_tx_complete(struct bnad *bnad, struct bna_tcb *tcb)
|
||||||
{
|
{
|
||||||
struct net_device *netdev = bnad->netdev;
|
struct net_device *netdev = bnad->netdev;
|
||||||
u32 sent = 0;
|
u32 sent = 0;
|
||||||
@ -244,7 +244,7 @@ bnad_tx(struct bnad *bnad, struct bna_tcb *tcb)
|
|||||||
if (test_and_set_bit(BNAD_TXQ_FREE_SENT, &tcb->flags))
|
if (test_and_set_bit(BNAD_TXQ_FREE_SENT, &tcb->flags))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
sent = bnad_free_txbufs(bnad, tcb);
|
sent = bnad_txcmpl_process(bnad, tcb);
|
||||||
if (sent) {
|
if (sent) {
|
||||||
if (netif_queue_stopped(netdev) &&
|
if (netif_queue_stopped(netdev) &&
|
||||||
netif_carrier_ok(netdev) &&
|
netif_carrier_ok(netdev) &&
|
||||||
@ -273,13 +273,13 @@ bnad_msix_tx(int irq, void *data)
|
|||||||
struct bna_tcb *tcb = (struct bna_tcb *)data;
|
struct bna_tcb *tcb = (struct bna_tcb *)data;
|
||||||
struct bnad *bnad = tcb->bnad;
|
struct bnad *bnad = tcb->bnad;
|
||||||
|
|
||||||
bnad_tx(bnad, tcb);
|
bnad_tx_complete(bnad, tcb);
|
||||||
|
|
||||||
return IRQ_HANDLED;
|
return IRQ_HANDLED;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
bnad_reset_rcb(struct bnad *bnad, struct bna_rcb *rcb)
|
bnad_rcb_cleanup(struct bnad *bnad, struct bna_rcb *rcb)
|
||||||
{
|
{
|
||||||
struct bnad_unmap_q *unmap_q = rcb->unmap_q;
|
struct bnad_unmap_q *unmap_q = rcb->unmap_q;
|
||||||
|
|
||||||
@ -291,7 +291,7 @@ bnad_reset_rcb(struct bnad *bnad, struct bna_rcb *rcb)
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
bnad_free_all_rxbufs(struct bnad *bnad, struct bna_rcb *rcb)
|
bnad_rxq_cleanup(struct bnad *bnad, struct bna_rcb *rcb)
|
||||||
{
|
{
|
||||||
struct bnad_unmap_q *unmap_q;
|
struct bnad_unmap_q *unmap_q;
|
||||||
struct bnad_skb_unmap *unmap_array;
|
struct bnad_skb_unmap *unmap_array;
|
||||||
@ -312,11 +312,11 @@ bnad_free_all_rxbufs(struct bnad *bnad, struct bna_rcb *rcb)
|
|||||||
DMA_FROM_DEVICE);
|
DMA_FROM_DEVICE);
|
||||||
dev_kfree_skb(skb);
|
dev_kfree_skb(skb);
|
||||||
}
|
}
|
||||||
bnad_reset_rcb(bnad, rcb);
|
bnad_rcb_cleanup(bnad, rcb);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
bnad_alloc_n_post_rxbufs(struct bnad *bnad, struct bna_rcb *rcb)
|
bnad_rxq_post(struct bnad *bnad, struct bna_rcb *rcb)
|
||||||
{
|
{
|
||||||
u16 to_alloc, alloced, unmap_prod, wi_range;
|
u16 to_alloc, alloced, unmap_prod, wi_range;
|
||||||
struct bnad_unmap_q *unmap_q = rcb->unmap_q;
|
struct bnad_unmap_q *unmap_q = rcb->unmap_q;
|
||||||
@ -377,14 +377,14 @@ bnad_refill_rxq(struct bnad *bnad, struct bna_rcb *rcb)
|
|||||||
if (!test_and_set_bit(BNAD_RXQ_REFILL, &rcb->flags)) {
|
if (!test_and_set_bit(BNAD_RXQ_REFILL, &rcb->flags)) {
|
||||||
if (BNA_QE_FREE_CNT(unmap_q, unmap_q->q_depth)
|
if (BNA_QE_FREE_CNT(unmap_q, unmap_q->q_depth)
|
||||||
>> BNAD_RXQ_REFILL_THRESHOLD_SHIFT)
|
>> BNAD_RXQ_REFILL_THRESHOLD_SHIFT)
|
||||||
bnad_alloc_n_post_rxbufs(bnad, rcb);
|
bnad_rxq_post(bnad, rcb);
|
||||||
smp_mb__before_clear_bit();
|
smp_mb__before_clear_bit();
|
||||||
clear_bit(BNAD_RXQ_REFILL, &rcb->flags);
|
clear_bit(BNAD_RXQ_REFILL, &rcb->flags);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static u32
|
static u32
|
||||||
bnad_poll_cq(struct bnad *bnad, struct bna_ccb *ccb, int budget)
|
bnad_cq_process(struct bnad *bnad, struct bna_ccb *ccb, int budget)
|
||||||
{
|
{
|
||||||
struct bna_cq_entry *cmpl, *next_cmpl;
|
struct bna_cq_entry *cmpl, *next_cmpl;
|
||||||
struct bna_rcb *rcb = NULL;
|
struct bna_rcb *rcb = NULL;
|
||||||
@ -584,7 +584,7 @@ bnad_isr(int irq, void *data)
|
|||||||
for (j = 0; j < bnad->num_txq_per_tx; j++) {
|
for (j = 0; j < bnad->num_txq_per_tx; j++) {
|
||||||
tcb = bnad->tx_info[i].tcb[j];
|
tcb = bnad->tx_info[i].tcb[j];
|
||||||
if (tcb && test_bit(BNAD_TXQ_TX_STARTED, &tcb->flags))
|
if (tcb && test_bit(BNAD_TXQ_TX_STARTED, &tcb->flags))
|
||||||
bnad_tx(bnad, bnad->tx_info[i].tcb[j]);
|
bnad_tx_complete(bnad, bnad->tx_info[i].tcb[j]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/* Rx processing */
|
/* Rx processing */
|
||||||
@ -895,7 +895,7 @@ bnad_tx_cleanup(struct delayed_work *work)
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
bnad_free_all_txbufs(bnad, tcb);
|
bnad_txq_cleanup(bnad, tcb);
|
||||||
|
|
||||||
unmap_q = tcb->unmap_q;
|
unmap_q = tcb->unmap_q;
|
||||||
unmap_q->producer_index = 0;
|
unmap_q->producer_index = 0;
|
||||||
@ -981,10 +981,10 @@ bnad_rx_cleanup(void *work)
|
|||||||
*/
|
*/
|
||||||
napi_disable(&rx_ctrl->napi);
|
napi_disable(&rx_ctrl->napi);
|
||||||
|
|
||||||
bnad_cq_cmpl_init(bnad, rx_ctrl->ccb);
|
bnad_cq_cleanup(bnad, rx_ctrl->ccb);
|
||||||
bnad_free_all_rxbufs(bnad, rx_ctrl->ccb->rcb[0]);
|
bnad_rxq_cleanup(bnad, rx_ctrl->ccb->rcb[0]);
|
||||||
if (rx_ctrl->ccb->rcb[1])
|
if (rx_ctrl->ccb->rcb[1])
|
||||||
bnad_free_all_rxbufs(bnad, rx_ctrl->ccb->rcb[1]);
|
bnad_rxq_cleanup(bnad, rx_ctrl->ccb->rcb[1]);
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_lock_irqsave(&bnad->bna_lock, flags);
|
spin_lock_irqsave(&bnad->bna_lock, flags);
|
||||||
@ -1048,7 +1048,7 @@ bnad_cb_rx_post(struct bnad *bnad, struct bna_rx *rx)
|
|||||||
if (!test_and_set_bit(BNAD_RXQ_REFILL, &rcb->flags)) {
|
if (!test_and_set_bit(BNAD_RXQ_REFILL, &rcb->flags)) {
|
||||||
if (BNA_QE_FREE_CNT(unmap_q, unmap_q->q_depth)
|
if (BNA_QE_FREE_CNT(unmap_q, unmap_q->q_depth)
|
||||||
>> BNAD_RXQ_REFILL_THRESHOLD_SHIFT)
|
>> BNAD_RXQ_REFILL_THRESHOLD_SHIFT)
|
||||||
bnad_alloc_n_post_rxbufs(bnad, rcb);
|
bnad_rxq_post(bnad, rcb);
|
||||||
smp_mb__before_clear_bit();
|
smp_mb__before_clear_bit();
|
||||||
clear_bit(BNAD_RXQ_REFILL, &rcb->flags);
|
clear_bit(BNAD_RXQ_REFILL, &rcb->flags);
|
||||||
}
|
}
|
||||||
@ -1672,7 +1672,7 @@ bnad_napi_poll_rx(struct napi_struct *napi, int budget)
|
|||||||
if (!netif_carrier_ok(bnad->netdev))
|
if (!netif_carrier_ok(bnad->netdev))
|
||||||
goto poll_exit;
|
goto poll_exit;
|
||||||
|
|
||||||
rcvd = bnad_poll_cq(bnad, rx_ctrl->ccb, budget);
|
rcvd = bnad_cq_process(bnad, rx_ctrl->ccb, budget);
|
||||||
if (rcvd >= budget)
|
if (rcvd >= budget)
|
||||||
return rcvd;
|
return rcvd;
|
||||||
|
|
||||||
@ -1714,7 +1714,7 @@ bnad_napi_delete(struct bnad *bnad, u32 rx_id)
|
|||||||
|
|
||||||
/* Should be held with conf_lock held */
|
/* Should be held with conf_lock held */
|
||||||
void
|
void
|
||||||
bnad_cleanup_tx(struct bnad *bnad, u32 tx_id)
|
bnad_destroy_tx(struct bnad *bnad, u32 tx_id)
|
||||||
{
|
{
|
||||||
struct bnad_tx_info *tx_info = &bnad->tx_info[tx_id];
|
struct bnad_tx_info *tx_info = &bnad->tx_info[tx_id];
|
||||||
struct bna_res_info *res_info = &bnad->tx_res_info[tx_id].res_info[0];
|
struct bna_res_info *res_info = &bnad->tx_res_info[tx_id].res_info[0];
|
||||||
@ -1865,7 +1865,7 @@ bnad_rx_ctrl_init(struct bnad *bnad, u32 rx_id)
|
|||||||
|
|
||||||
/* Called with mutex_lock(&bnad->conf_mutex) held */
|
/* Called with mutex_lock(&bnad->conf_mutex) held */
|
||||||
void
|
void
|
||||||
bnad_cleanup_rx(struct bnad *bnad, u32 rx_id)
|
bnad_destroy_rx(struct bnad *bnad, u32 rx_id)
|
||||||
{
|
{
|
||||||
struct bnad_rx_info *rx_info = &bnad->rx_info[rx_id];
|
struct bnad_rx_info *rx_info = &bnad->rx_info[rx_id];
|
||||||
struct bna_rx_config *rx_config = &bnad->rx_config[rx_id];
|
struct bna_rx_config *rx_config = &bnad->rx_config[rx_id];
|
||||||
@ -2003,7 +2003,7 @@ bnad_setup_rx(struct bnad *bnad, u32 rx_id)
|
|||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
err_return:
|
err_return:
|
||||||
bnad_cleanup_rx(bnad, rx_id);
|
bnad_destroy_rx(bnad, rx_id);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2488,7 +2488,7 @@ bnad_open(struct net_device *netdev)
|
|||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
cleanup_tx:
|
cleanup_tx:
|
||||||
bnad_cleanup_tx(bnad, 0);
|
bnad_destroy_tx(bnad, 0);
|
||||||
|
|
||||||
err_return:
|
err_return:
|
||||||
mutex_unlock(&bnad->conf_mutex);
|
mutex_unlock(&bnad->conf_mutex);
|
||||||
@ -2515,8 +2515,8 @@ bnad_stop(struct net_device *netdev)
|
|||||||
|
|
||||||
wait_for_completion(&bnad->bnad_completions.enet_comp);
|
wait_for_completion(&bnad->bnad_completions.enet_comp);
|
||||||
|
|
||||||
bnad_cleanup_tx(bnad, 0);
|
bnad_destroy_tx(bnad, 0);
|
||||||
bnad_cleanup_rx(bnad, 0);
|
bnad_destroy_rx(bnad, 0);
|
||||||
|
|
||||||
/* Synchronize mailbox IRQ */
|
/* Synchronize mailbox IRQ */
|
||||||
bnad_mbox_irq_sync(bnad);
|
bnad_mbox_irq_sync(bnad);
|
||||||
@ -2589,7 +2589,7 @@ bnad_start_xmit(struct sk_buff *skb, struct net_device *netdev)
|
|||||||
if ((u16) (*tcb->hw_consumer_index) !=
|
if ((u16) (*tcb->hw_consumer_index) !=
|
||||||
tcb->consumer_index &&
|
tcb->consumer_index &&
|
||||||
!test_and_set_bit(BNAD_TXQ_FREE_SENT, &tcb->flags)) {
|
!test_and_set_bit(BNAD_TXQ_FREE_SENT, &tcb->flags)) {
|
||||||
acked = bnad_free_txbufs(bnad, tcb);
|
acked = bnad_txcmpl_process(bnad, tcb);
|
||||||
if (likely(test_bit(BNAD_TXQ_TX_STARTED, &tcb->flags)))
|
if (likely(test_bit(BNAD_TXQ_TX_STARTED, &tcb->flags)))
|
||||||
bna_ib_ack(tcb->i_dbell, acked);
|
bna_ib_ack(tcb->i_dbell, acked);
|
||||||
smp_mb__before_clear_bit();
|
smp_mb__before_clear_bit();
|
||||||
|
@ -373,8 +373,8 @@ extern void bnad_rx_coalescing_timeo_set(struct bnad *bnad);
|
|||||||
|
|
||||||
extern int bnad_setup_rx(struct bnad *bnad, u32 rx_id);
|
extern int bnad_setup_rx(struct bnad *bnad, u32 rx_id);
|
||||||
extern int bnad_setup_tx(struct bnad *bnad, u32 tx_id);
|
extern int bnad_setup_tx(struct bnad *bnad, u32 tx_id);
|
||||||
extern void bnad_cleanup_tx(struct bnad *bnad, u32 tx_id);
|
extern void bnad_destroy_tx(struct bnad *bnad, u32 tx_id);
|
||||||
extern void bnad_cleanup_rx(struct bnad *bnad, u32 rx_id);
|
extern void bnad_destroy_rx(struct bnad *bnad, u32 rx_id);
|
||||||
|
|
||||||
/* Timer start/stop protos */
|
/* Timer start/stop protos */
|
||||||
extern void bnad_dim_timer_start(struct bnad *bnad);
|
extern void bnad_dim_timer_start(struct bnad *bnad);
|
||||||
|
@ -464,7 +464,7 @@ bnad_set_ringparam(struct net_device *netdev,
|
|||||||
for (i = 0; i < bnad->num_rx; i++) {
|
for (i = 0; i < bnad->num_rx; i++) {
|
||||||
if (!bnad->rx_info[i].rx)
|
if (!bnad->rx_info[i].rx)
|
||||||
continue;
|
continue;
|
||||||
bnad_cleanup_rx(bnad, i);
|
bnad_destroy_rx(bnad, i);
|
||||||
current_err = bnad_setup_rx(bnad, i);
|
current_err = bnad_setup_rx(bnad, i);
|
||||||
if (current_err && !err)
|
if (current_err && !err)
|
||||||
err = current_err;
|
err = current_err;
|
||||||
@ -492,7 +492,7 @@ bnad_set_ringparam(struct net_device *netdev,
|
|||||||
for (i = 0; i < bnad->num_tx; i++) {
|
for (i = 0; i < bnad->num_tx; i++) {
|
||||||
if (!bnad->tx_info[i].tx)
|
if (!bnad->tx_info[i].tx)
|
||||||
continue;
|
continue;
|
||||||
bnad_cleanup_tx(bnad, i);
|
bnad_destroy_tx(bnad, i);
|
||||||
current_err = bnad_setup_tx(bnad, i);
|
current_err = bnad_setup_tx(bnad, i);
|
||||||
if (current_err && !err)
|
if (current_err && !err)
|
||||||
err = current_err;
|
err = current_err;
|
||||||
@ -539,7 +539,7 @@ bnad_set_pauseparam(struct net_device *netdev,
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
bnad_get_strings(struct net_device *netdev, u32 stringset, u8 * string)
|
bnad_get_strings(struct net_device *netdev, u32 stringset, u8 *string)
|
||||||
{
|
{
|
||||||
struct bnad *bnad = netdev_priv(netdev);
|
struct bnad *bnad = netdev_priv(netdev);
|
||||||
int i, j, q_num;
|
int i, j, q_num;
|
||||||
|
Loading…
Reference in New Issue
Block a user