Commit 6a606b90 authored by Alex Elder's avatar Alex Elder Committed by David S. Miller

net: ipa: allocate transaction in replenish loop

When replenishing, have ipa_endpoint_replenish() allocate a
transaction, and pass that to ipa_endpoint_replenish_one() to fill.
Then, if that produces no error, commit the transaction within the
replenish loop as well.  In this way we can distinguish between
transaction failures and buffer allocation/mapping failures.

Failure to allocate a transaction simply means the hardware already
has as many receive buffers as it can hold.  In that case we can
break out of the replenish loop because there's nothing more to do.

If we fail to allocate or map pages for the receive buffer, just
try again later.
Signed-off-by: default avatarAlex Elder <elder@linaro.org>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent b9dbabc5
......@@ -1036,24 +1036,19 @@ static void ipa_endpoint_status(struct ipa_endpoint *endpoint)
iowrite32(val, ipa->reg_virt + offset);
}
static int
ipa_endpoint_replenish_one(struct ipa_endpoint *endpoint, bool doorbell)
static int ipa_endpoint_replenish_one(struct ipa_endpoint *endpoint,
struct gsi_trans *trans)
{
struct gsi_trans *trans;
struct page *page;
u32 buffer_size;
u32 offset;
u32 len;
int ret;
trans = ipa_endpoint_trans_alloc(endpoint, 1);
if (!trans)
return -ENOMEM;
buffer_size = endpoint->data->rx.buffer_size;
page = dev_alloc_pages(get_order(buffer_size));
if (!page)
goto err_trans_free;
return -ENOMEM;
/* Offset the buffer to make space for skb headroom */
offset = NET_SKB_PAD;
......@@ -1061,19 +1056,11 @@ ipa_endpoint_replenish_one(struct ipa_endpoint *endpoint, bool doorbell)
ret = gsi_trans_page_add(trans, page, len, offset);
if (ret)
goto err_free_pages;
trans->data = page; /* transaction owns page now */
gsi_trans_commit(trans, doorbell);
return 0;
err_free_pages:
__free_pages(page, get_order(buffer_size));
err_trans_free:
gsi_trans_free(trans);
else
trans->data = page; /* transaction owns page now */
return -ENOMEM;
return ret;
}
/**
......@@ -1089,6 +1076,7 @@ ipa_endpoint_replenish_one(struct ipa_endpoint *endpoint, bool doorbell)
*/
static void ipa_endpoint_replenish(struct ipa_endpoint *endpoint)
{
struct gsi_trans *trans;
struct gsi *gsi;
u32 backlog;
......@@ -1100,15 +1088,18 @@ static void ipa_endpoint_replenish(struct ipa_endpoint *endpoint)
return;
while (atomic_dec_not_zero(&endpoint->replenish_backlog)) {
bool doorbell;
trans = ipa_endpoint_trans_alloc(endpoint, 1);
if (!trans)
break;
if (ipa_endpoint_replenish_one(endpoint, trans))
goto try_again_later;
if (++endpoint->replenish_ready == IPA_REPLENISH_BATCH)
endpoint->replenish_ready = 0;
/* Ring the doorbell if we've got a full batch */
doorbell = !endpoint->replenish_ready;
if (ipa_endpoint_replenish_one(endpoint, doorbell))
goto try_again_later;
gsi_trans_commit(trans, !endpoint->replenish_ready);
}
clear_bit(IPA_REPLENISH_ACTIVE, endpoint->replenish_flags);
......@@ -1116,6 +1107,7 @@ static void ipa_endpoint_replenish(struct ipa_endpoint *endpoint)
return;
try_again_later:
gsi_trans_free(trans);
clear_bit(IPA_REPLENISH_ACTIVE, endpoint->replenish_flags);
/* The last one didn't succeed, so fix the backlog */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment