mirror of
https://github.com/Fishwaldo/Star64_linux.git
synced 2025-06-28 01:21:58 +00:00
sfc: Remove some unreachable error paths
Some functions return an error code which is always 0. Change their return types to void and simplify their callers accordingly. Signed-off-by: Ben Hutchings <bhutchings@solarflare.com> Signed-off-by: Jeff Garzik <jgarzik@redhat.com>
This commit is contained in:
parent
c1e5fcc980
commit
bc3c90a2b7
9 changed files with 39 additions and 95 deletions
|
@ -282,13 +282,13 @@ static int efx_probe_eventq(struct efx_channel *channel)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Prepare channel's event queue */
|
/* Prepare channel's event queue */
|
||||||
static int efx_init_eventq(struct efx_channel *channel)
|
static void efx_init_eventq(struct efx_channel *channel)
|
||||||
{
|
{
|
||||||
EFX_LOG(channel->efx, "chan %d init event queue\n", channel->channel);
|
EFX_LOG(channel->efx, "chan %d init event queue\n", channel->channel);
|
||||||
|
|
||||||
channel->eventq_read_ptr = 0;
|
channel->eventq_read_ptr = 0;
|
||||||
|
|
||||||
return falcon_init_eventq(channel);
|
falcon_init_eventq(channel);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void efx_fini_eventq(struct efx_channel *channel)
|
static void efx_fini_eventq(struct efx_channel *channel)
|
||||||
|
@ -354,12 +354,11 @@ static int efx_probe_channel(struct efx_channel *channel)
|
||||||
* to propagate configuration changes (mtu, checksum offload), or
|
* to propagate configuration changes (mtu, checksum offload), or
|
||||||
* to clear hardware error conditions
|
* to clear hardware error conditions
|
||||||
*/
|
*/
|
||||||
static int efx_init_channels(struct efx_nic *efx)
|
static void efx_init_channels(struct efx_nic *efx)
|
||||||
{
|
{
|
||||||
struct efx_tx_queue *tx_queue;
|
struct efx_tx_queue *tx_queue;
|
||||||
struct efx_rx_queue *rx_queue;
|
struct efx_rx_queue *rx_queue;
|
||||||
struct efx_channel *channel;
|
struct efx_channel *channel;
|
||||||
int rc = 0;
|
|
||||||
|
|
||||||
/* Calculate the rx buffer allocation parameters required to
|
/* Calculate the rx buffer allocation parameters required to
|
||||||
* support the current MTU, including padding for header
|
* support the current MTU, including padding for header
|
||||||
|
@ -374,36 +373,20 @@ static int efx_init_channels(struct efx_nic *efx)
|
||||||
efx_for_each_channel(channel, efx) {
|
efx_for_each_channel(channel, efx) {
|
||||||
EFX_LOG(channel->efx, "init chan %d\n", channel->channel);
|
EFX_LOG(channel->efx, "init chan %d\n", channel->channel);
|
||||||
|
|
||||||
rc = efx_init_eventq(channel);
|
efx_init_eventq(channel);
|
||||||
if (rc)
|
|
||||||
goto err;
|
|
||||||
|
|
||||||
efx_for_each_channel_tx_queue(tx_queue, channel) {
|
efx_for_each_channel_tx_queue(tx_queue, channel)
|
||||||
rc = efx_init_tx_queue(tx_queue);
|
efx_init_tx_queue(tx_queue);
|
||||||
if (rc)
|
|
||||||
goto err;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* The rx buffer allocation strategy is MTU dependent */
|
/* The rx buffer allocation strategy is MTU dependent */
|
||||||
efx_rx_strategy(channel);
|
efx_rx_strategy(channel);
|
||||||
|
|
||||||
efx_for_each_channel_rx_queue(rx_queue, channel) {
|
efx_for_each_channel_rx_queue(rx_queue, channel)
|
||||||
rc = efx_init_rx_queue(rx_queue);
|
efx_init_rx_queue(rx_queue);
|
||||||
if (rc)
|
|
||||||
goto err;
|
|
||||||
}
|
|
||||||
|
|
||||||
WARN_ON(channel->rx_pkt != NULL);
|
WARN_ON(channel->rx_pkt != NULL);
|
||||||
efx_rx_strategy(channel);
|
efx_rx_strategy(channel);
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
err:
|
|
||||||
EFX_ERR(efx, "failed to initialise channel %d\n",
|
|
||||||
channel ? channel->channel : -1);
|
|
||||||
efx_fini_channels(efx);
|
|
||||||
return rc;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* This enables event queue processing and packet transmission.
|
/* This enables event queue processing and packet transmission.
|
||||||
|
@ -1121,24 +1104,16 @@ static void efx_remove_all(struct efx_nic *efx)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* A convinience function to safely flush all the queues */
|
/* A convinience function to safely flush all the queues */
|
||||||
int efx_flush_queues(struct efx_nic *efx)
|
void efx_flush_queues(struct efx_nic *efx)
|
||||||
{
|
{
|
||||||
int rc;
|
|
||||||
|
|
||||||
EFX_ASSERT_RESET_SERIALISED(efx);
|
EFX_ASSERT_RESET_SERIALISED(efx);
|
||||||
|
|
||||||
efx_stop_all(efx);
|
efx_stop_all(efx);
|
||||||
|
|
||||||
efx_fini_channels(efx);
|
efx_fini_channels(efx);
|
||||||
rc = efx_init_channels(efx);
|
efx_init_channels(efx);
|
||||||
if (rc) {
|
|
||||||
efx_schedule_reset(efx, RESET_TYPE_DISABLE);
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
|
|
||||||
efx_start_all(efx);
|
efx_start_all(efx);
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**************************************************************************
|
/**************************************************************************
|
||||||
|
@ -1311,7 +1286,6 @@ static int efx_net_open(struct net_device *net_dev)
|
||||||
static int efx_net_stop(struct net_device *net_dev)
|
static int efx_net_stop(struct net_device *net_dev)
|
||||||
{
|
{
|
||||||
struct efx_nic *efx = netdev_priv(net_dev);
|
struct efx_nic *efx = netdev_priv(net_dev);
|
||||||
int rc;
|
|
||||||
|
|
||||||
EFX_LOG(efx, "closing %s on CPU %d\n", net_dev->name,
|
EFX_LOG(efx, "closing %s on CPU %d\n", net_dev->name,
|
||||||
raw_smp_processor_id());
|
raw_smp_processor_id());
|
||||||
|
@ -1319,9 +1293,7 @@ static int efx_net_stop(struct net_device *net_dev)
|
||||||
/* Stop the device and flush all the channels */
|
/* Stop the device and flush all the channels */
|
||||||
efx_stop_all(efx);
|
efx_stop_all(efx);
|
||||||
efx_fini_channels(efx);
|
efx_fini_channels(efx);
|
||||||
rc = efx_init_channels(efx);
|
efx_init_channels(efx);
|
||||||
if (rc)
|
|
||||||
efx_schedule_reset(efx, RESET_TYPE_DISABLE);
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@ -1404,16 +1376,10 @@ static int efx_change_mtu(struct net_device *net_dev, int new_mtu)
|
||||||
|
|
||||||
efx_fini_channels(efx);
|
efx_fini_channels(efx);
|
||||||
net_dev->mtu = new_mtu;
|
net_dev->mtu = new_mtu;
|
||||||
rc = efx_init_channels(efx);
|
efx_init_channels(efx);
|
||||||
if (rc)
|
|
||||||
goto fail;
|
|
||||||
|
|
||||||
efx_start_all(efx);
|
efx_start_all(efx);
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
fail:
|
|
||||||
efx_schedule_reset(efx, RESET_TYPE_DISABLE);
|
|
||||||
return rc;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int efx_set_mac_address(struct net_device *net_dev, void *data)
|
static int efx_set_mac_address(struct net_device *net_dev, void *data)
|
||||||
|
@ -1588,22 +1554,19 @@ static int efx_reset_up(struct efx_nic *efx, struct ethtool_cmd *ecmd)
|
||||||
{
|
{
|
||||||
int rc;
|
int rc;
|
||||||
|
|
||||||
rc = efx_init_channels(efx);
|
efx_init_channels(efx);
|
||||||
if (rc)
|
|
||||||
goto fail1;
|
|
||||||
|
|
||||||
/* Restore MAC and PHY settings. */
|
/* Restore MAC and PHY settings. */
|
||||||
rc = falcon_xmac_set_settings(efx, ecmd);
|
rc = falcon_xmac_set_settings(efx, ecmd);
|
||||||
if (rc) {
|
if (rc) {
|
||||||
EFX_ERR(efx, "could not restore PHY settings\n");
|
EFX_ERR(efx, "could not restore PHY settings\n");
|
||||||
goto fail2;
|
goto fail;
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
fail2:
|
fail:
|
||||||
efx_fini_channels(efx);
|
efx_fini_channels(efx);
|
||||||
fail1:
|
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2023,19 +1986,16 @@ static int efx_pci_probe_main(struct efx_nic *efx)
|
||||||
goto fail5;
|
goto fail5;
|
||||||
}
|
}
|
||||||
|
|
||||||
rc = efx_init_channels(efx);
|
efx_init_channels(efx);
|
||||||
if (rc)
|
|
||||||
goto fail6;
|
|
||||||
|
|
||||||
rc = falcon_init_interrupt(efx);
|
rc = falcon_init_interrupt(efx);
|
||||||
if (rc)
|
if (rc)
|
||||||
goto fail7;
|
goto fail6;
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
fail7:
|
|
||||||
efx_fini_channels(efx);
|
|
||||||
fail6:
|
fail6:
|
||||||
|
efx_fini_channels(efx);
|
||||||
efx_fini_port(efx);
|
efx_fini_port(efx);
|
||||||
fail5:
|
fail5:
|
||||||
fail4:
|
fail4:
|
||||||
|
|
|
@ -33,7 +33,7 @@ extern void efx_schedule_slow_fill(struct efx_rx_queue *rx_queue, int delay);
|
||||||
|
|
||||||
/* Channels */
|
/* Channels */
|
||||||
extern void efx_process_channel_now(struct efx_channel *channel);
|
extern void efx_process_channel_now(struct efx_channel *channel);
|
||||||
extern int efx_flush_queues(struct efx_nic *efx);
|
extern void efx_flush_queues(struct efx_nic *efx);
|
||||||
|
|
||||||
/* Ports */
|
/* Ports */
|
||||||
extern void efx_reconfigure_port(struct efx_nic *efx);
|
extern void efx_reconfigure_port(struct efx_nic *efx);
|
||||||
|
|
|
@ -503,10 +503,10 @@ static void efx_ethtool_self_test(struct net_device *net_dev,
|
||||||
if (offline) {
|
if (offline) {
|
||||||
/* Stop the kernel from sending packets during the test. */
|
/* Stop the kernel from sending packets during the test. */
|
||||||
efx_stop_queue(efx);
|
efx_stop_queue(efx);
|
||||||
rc = efx_flush_queues(efx);
|
efx_flush_queues(efx);
|
||||||
if (!rc)
|
|
||||||
rc = efx_offline_test(efx, &efx_tests,
|
rc = efx_offline_test(efx, &efx_tests,
|
||||||
efx->loopback_modes);
|
efx->loopback_modes);
|
||||||
efx_wake_queue(efx);
|
efx_wake_queue(efx);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -242,7 +242,7 @@ static struct i2c_algo_bit_data falcon_i2c_bit_operations = {
|
||||||
* falcon_alloc_special_buffer()) in Falcon's buffer table, allowing
|
* falcon_alloc_special_buffer()) in Falcon's buffer table, allowing
|
||||||
* it to be used for event queues, descriptor rings etc.
|
* it to be used for event queues, descriptor rings etc.
|
||||||
*/
|
*/
|
||||||
static int
|
static void
|
||||||
falcon_init_special_buffer(struct efx_nic *efx,
|
falcon_init_special_buffer(struct efx_nic *efx,
|
||||||
struct efx_special_buffer *buffer)
|
struct efx_special_buffer *buffer)
|
||||||
{
|
{
|
||||||
|
@ -266,8 +266,6 @@ falcon_init_special_buffer(struct efx_nic *efx,
|
||||||
BUF_OWNER_ID_FBUF, 0);
|
BUF_OWNER_ID_FBUF, 0);
|
||||||
falcon_write_sram(efx, &buf_desc, index);
|
falcon_write_sram(efx, &buf_desc, index);
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Unmaps a buffer from Falcon and clears the buffer table entries */
|
/* Unmaps a buffer from Falcon and clears the buffer table entries */
|
||||||
|
@ -449,16 +447,13 @@ int falcon_probe_tx(struct efx_tx_queue *tx_queue)
|
||||||
sizeof(efx_qword_t));
|
sizeof(efx_qword_t));
|
||||||
}
|
}
|
||||||
|
|
||||||
int falcon_init_tx(struct efx_tx_queue *tx_queue)
|
void falcon_init_tx(struct efx_tx_queue *tx_queue)
|
||||||
{
|
{
|
||||||
efx_oword_t tx_desc_ptr;
|
efx_oword_t tx_desc_ptr;
|
||||||
struct efx_nic *efx = tx_queue->efx;
|
struct efx_nic *efx = tx_queue->efx;
|
||||||
int rc;
|
|
||||||
|
|
||||||
/* Pin TX descriptor ring */
|
/* Pin TX descriptor ring */
|
||||||
rc = falcon_init_special_buffer(efx, &tx_queue->txd);
|
falcon_init_special_buffer(efx, &tx_queue->txd);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
|
|
||||||
/* Push TX descriptor ring to card */
|
/* Push TX descriptor ring to card */
|
||||||
EFX_POPULATE_OWORD_10(tx_desc_ptr,
|
EFX_POPULATE_OWORD_10(tx_desc_ptr,
|
||||||
|
@ -495,8 +490,6 @@ int falcon_init_tx(struct efx_tx_queue *tx_queue)
|
||||||
set_bit_le(tx_queue->queue, (void *)®);
|
set_bit_le(tx_queue->queue, (void *)®);
|
||||||
falcon_write(efx, ®, TX_CHKSM_CFG_REG_KER_A1);
|
falcon_write(efx, ®, TX_CHKSM_CFG_REG_KER_A1);
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int falcon_flush_tx_queue(struct efx_tx_queue *tx_queue)
|
static int falcon_flush_tx_queue(struct efx_tx_queue *tx_queue)
|
||||||
|
@ -639,11 +632,10 @@ int falcon_probe_rx(struct efx_rx_queue *rx_queue)
|
||||||
sizeof(efx_qword_t));
|
sizeof(efx_qword_t));
|
||||||
}
|
}
|
||||||
|
|
||||||
int falcon_init_rx(struct efx_rx_queue *rx_queue)
|
void falcon_init_rx(struct efx_rx_queue *rx_queue)
|
||||||
{
|
{
|
||||||
efx_oword_t rx_desc_ptr;
|
efx_oword_t rx_desc_ptr;
|
||||||
struct efx_nic *efx = rx_queue->efx;
|
struct efx_nic *efx = rx_queue->efx;
|
||||||
int rc;
|
|
||||||
bool is_b0 = falcon_rev(efx) >= FALCON_REV_B0;
|
bool is_b0 = falcon_rev(efx) >= FALCON_REV_B0;
|
||||||
bool iscsi_digest_en = is_b0;
|
bool iscsi_digest_en = is_b0;
|
||||||
|
|
||||||
|
@ -652,9 +644,7 @@ int falcon_init_rx(struct efx_rx_queue *rx_queue)
|
||||||
rx_queue->rxd.index + rx_queue->rxd.entries - 1);
|
rx_queue->rxd.index + rx_queue->rxd.entries - 1);
|
||||||
|
|
||||||
/* Pin RX descriptor ring */
|
/* Pin RX descriptor ring */
|
||||||
rc = falcon_init_special_buffer(efx, &rx_queue->rxd);
|
falcon_init_special_buffer(efx, &rx_queue->rxd);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
|
|
||||||
/* Push RX descriptor ring to card */
|
/* Push RX descriptor ring to card */
|
||||||
EFX_POPULATE_OWORD_10(rx_desc_ptr,
|
EFX_POPULATE_OWORD_10(rx_desc_ptr,
|
||||||
|
@ -671,7 +661,6 @@ int falcon_init_rx(struct efx_rx_queue *rx_queue)
|
||||||
RX_DESCQ_EN, 1);
|
RX_DESCQ_EN, 1);
|
||||||
falcon_write_table(efx, &rx_desc_ptr, efx->type->rxd_ptr_tbl_base,
|
falcon_write_table(efx, &rx_desc_ptr, efx->type->rxd_ptr_tbl_base,
|
||||||
rx_queue->queue);
|
rx_queue->queue);
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int falcon_flush_rx_queue(struct efx_rx_queue *rx_queue)
|
static int falcon_flush_rx_queue(struct efx_rx_queue *rx_queue)
|
||||||
|
@ -1205,20 +1194,17 @@ int falcon_probe_eventq(struct efx_channel *channel)
|
||||||
return falcon_alloc_special_buffer(efx, &channel->eventq, evq_size);
|
return falcon_alloc_special_buffer(efx, &channel->eventq, evq_size);
|
||||||
}
|
}
|
||||||
|
|
||||||
int falcon_init_eventq(struct efx_channel *channel)
|
void falcon_init_eventq(struct efx_channel *channel)
|
||||||
{
|
{
|
||||||
efx_oword_t evq_ptr;
|
efx_oword_t evq_ptr;
|
||||||
struct efx_nic *efx = channel->efx;
|
struct efx_nic *efx = channel->efx;
|
||||||
int rc;
|
|
||||||
|
|
||||||
EFX_LOG(efx, "channel %d event queue in special buffers %d-%d\n",
|
EFX_LOG(efx, "channel %d event queue in special buffers %d-%d\n",
|
||||||
channel->channel, channel->eventq.index,
|
channel->channel, channel->eventq.index,
|
||||||
channel->eventq.index + channel->eventq.entries - 1);
|
channel->eventq.index + channel->eventq.entries - 1);
|
||||||
|
|
||||||
/* Pin event queue buffer */
|
/* Pin event queue buffer */
|
||||||
rc = falcon_init_special_buffer(efx, &channel->eventq);
|
falcon_init_special_buffer(efx, &channel->eventq);
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
|
|
||||||
/* Fill event queue with all ones (i.e. empty events) */
|
/* Fill event queue with all ones (i.e. empty events) */
|
||||||
memset(channel->eventq.addr, 0xff, channel->eventq.len);
|
memset(channel->eventq.addr, 0xff, channel->eventq.len);
|
||||||
|
@ -1232,8 +1218,6 @@ int falcon_init_eventq(struct efx_channel *channel)
|
||||||
channel->channel);
|
channel->channel);
|
||||||
|
|
||||||
falcon_set_int_moderation(channel);
|
falcon_set_int_moderation(channel);
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void falcon_fini_eventq(struct efx_channel *channel)
|
void falcon_fini_eventq(struct efx_channel *channel)
|
||||||
|
|
|
@ -40,21 +40,21 @@ extern struct efx_nic_type falcon_b_nic_type;
|
||||||
|
|
||||||
/* TX data path */
|
/* TX data path */
|
||||||
extern int falcon_probe_tx(struct efx_tx_queue *tx_queue);
|
extern int falcon_probe_tx(struct efx_tx_queue *tx_queue);
|
||||||
extern int falcon_init_tx(struct efx_tx_queue *tx_queue);
|
extern void falcon_init_tx(struct efx_tx_queue *tx_queue);
|
||||||
extern void falcon_fini_tx(struct efx_tx_queue *tx_queue);
|
extern void falcon_fini_tx(struct efx_tx_queue *tx_queue);
|
||||||
extern void falcon_remove_tx(struct efx_tx_queue *tx_queue);
|
extern void falcon_remove_tx(struct efx_tx_queue *tx_queue);
|
||||||
extern void falcon_push_buffers(struct efx_tx_queue *tx_queue);
|
extern void falcon_push_buffers(struct efx_tx_queue *tx_queue);
|
||||||
|
|
||||||
/* RX data path */
|
/* RX data path */
|
||||||
extern int falcon_probe_rx(struct efx_rx_queue *rx_queue);
|
extern int falcon_probe_rx(struct efx_rx_queue *rx_queue);
|
||||||
extern int falcon_init_rx(struct efx_rx_queue *rx_queue);
|
extern void falcon_init_rx(struct efx_rx_queue *rx_queue);
|
||||||
extern void falcon_fini_rx(struct efx_rx_queue *rx_queue);
|
extern void falcon_fini_rx(struct efx_rx_queue *rx_queue);
|
||||||
extern void falcon_remove_rx(struct efx_rx_queue *rx_queue);
|
extern void falcon_remove_rx(struct efx_rx_queue *rx_queue);
|
||||||
extern void falcon_notify_rx_desc(struct efx_rx_queue *rx_queue);
|
extern void falcon_notify_rx_desc(struct efx_rx_queue *rx_queue);
|
||||||
|
|
||||||
/* Event data path */
|
/* Event data path */
|
||||||
extern int falcon_probe_eventq(struct efx_channel *channel);
|
extern int falcon_probe_eventq(struct efx_channel *channel);
|
||||||
extern int falcon_init_eventq(struct efx_channel *channel);
|
extern void falcon_init_eventq(struct efx_channel *channel);
|
||||||
extern void falcon_fini_eventq(struct efx_channel *channel);
|
extern void falcon_fini_eventq(struct efx_channel *channel);
|
||||||
extern void falcon_remove_eventq(struct efx_channel *channel);
|
extern void falcon_remove_eventq(struct efx_channel *channel);
|
||||||
extern int falcon_process_eventq(struct efx_channel *channel, int rx_quota);
|
extern int falcon_process_eventq(struct efx_channel *channel, int rx_quota);
|
||||||
|
|
|
@ -800,7 +800,7 @@ int efx_probe_rx_queue(struct efx_rx_queue *rx_queue)
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
int efx_init_rx_queue(struct efx_rx_queue *rx_queue)
|
void efx_init_rx_queue(struct efx_rx_queue *rx_queue)
|
||||||
{
|
{
|
||||||
struct efx_nic *efx = rx_queue->efx;
|
struct efx_nic *efx = rx_queue->efx;
|
||||||
unsigned int max_fill, trigger, limit;
|
unsigned int max_fill, trigger, limit;
|
||||||
|
@ -824,7 +824,7 @@ int efx_init_rx_queue(struct efx_rx_queue *rx_queue)
|
||||||
rx_queue->fast_fill_limit = limit;
|
rx_queue->fast_fill_limit = limit;
|
||||||
|
|
||||||
/* Set up RX descriptor ring */
|
/* Set up RX descriptor ring */
|
||||||
return falcon_init_rx(rx_queue);
|
falcon_init_rx(rx_queue);
|
||||||
}
|
}
|
||||||
|
|
||||||
void efx_fini_rx_queue(struct efx_rx_queue *rx_queue)
|
void efx_fini_rx_queue(struct efx_rx_queue *rx_queue)
|
||||||
|
|
|
@ -14,7 +14,7 @@
|
||||||
|
|
||||||
int efx_probe_rx_queue(struct efx_rx_queue *rx_queue);
|
int efx_probe_rx_queue(struct efx_rx_queue *rx_queue);
|
||||||
void efx_remove_rx_queue(struct efx_rx_queue *rx_queue);
|
void efx_remove_rx_queue(struct efx_rx_queue *rx_queue);
|
||||||
int efx_init_rx_queue(struct efx_rx_queue *rx_queue);
|
void efx_init_rx_queue(struct efx_rx_queue *rx_queue);
|
||||||
void efx_fini_rx_queue(struct efx_rx_queue *rx_queue);
|
void efx_fini_rx_queue(struct efx_rx_queue *rx_queue);
|
||||||
|
|
||||||
int efx_lro_init(struct net_lro_mgr *lro_mgr, struct efx_nic *efx);
|
int efx_lro_init(struct net_lro_mgr *lro_mgr, struct efx_nic *efx);
|
||||||
|
|
|
@ -443,7 +443,7 @@ int efx_probe_tx_queue(struct efx_tx_queue *tx_queue)
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
int efx_init_tx_queue(struct efx_tx_queue *tx_queue)
|
void efx_init_tx_queue(struct efx_tx_queue *tx_queue)
|
||||||
{
|
{
|
||||||
EFX_LOG(tx_queue->efx, "initialising TX queue %d\n", tx_queue->queue);
|
EFX_LOG(tx_queue->efx, "initialising TX queue %d\n", tx_queue->queue);
|
||||||
|
|
||||||
|
@ -454,7 +454,7 @@ int efx_init_tx_queue(struct efx_tx_queue *tx_queue)
|
||||||
BUG_ON(tx_queue->stopped);
|
BUG_ON(tx_queue->stopped);
|
||||||
|
|
||||||
/* Set up TX descriptor ring */
|
/* Set up TX descriptor ring */
|
||||||
return falcon_init_tx(tx_queue);
|
falcon_init_tx(tx_queue);
|
||||||
}
|
}
|
||||||
|
|
||||||
void efx_release_tx_buffers(struct efx_tx_queue *tx_queue)
|
void efx_release_tx_buffers(struct efx_tx_queue *tx_queue)
|
||||||
|
|
|
@ -15,7 +15,7 @@
|
||||||
|
|
||||||
int efx_probe_tx_queue(struct efx_tx_queue *tx_queue);
|
int efx_probe_tx_queue(struct efx_tx_queue *tx_queue);
|
||||||
void efx_remove_tx_queue(struct efx_tx_queue *tx_queue);
|
void efx_remove_tx_queue(struct efx_tx_queue *tx_queue);
|
||||||
int efx_init_tx_queue(struct efx_tx_queue *tx_queue);
|
void efx_init_tx_queue(struct efx_tx_queue *tx_queue);
|
||||||
void efx_fini_tx_queue(struct efx_tx_queue *tx_queue);
|
void efx_fini_tx_queue(struct efx_tx_queue *tx_queue);
|
||||||
|
|
||||||
int efx_hard_start_xmit(struct sk_buff *skb, struct net_device *net_dev);
|
int efx_hard_start_xmit(struct sk_buff *skb, struct net_device *net_dev);
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue