Commit 4293d5f5 authored by Mitch Williams's avatar Mitch Williams Committed by Jeff Kirsher

i40e: simplify txd use count calculation

The i40e_txd_use_count function was fast but confusing. In the comments,
it even admits that it's ugly. So replace it with a new function that is
(very) slightly faster and has extensive commenting to help the thicker
among us (including the author, who will forget in a week) understand
how it works.

Change-ID: Ifb533f13786a0bf39cb29f77969a5be2c83d9a87
Signed-off-by: default avatarMitch Williams <mitch.a.williams@intel.com>
Signed-off-by: default avatarAlexander Duyck <alexander.h.duyck@intel.com>
Tested-by: default avatarAndrew Bowers <andrewx.bowers@intel.com>
Signed-off-by: default avatarJeff Kirsher <jeffrey.t.kirsher@intel.com>
parent 7ec9ba11
...@@ -173,26 +173,37 @@ static inline bool i40e_test_staterr(union i40e_rx_desc *rx_desc, ...@@ -173,26 +173,37 @@ static inline bool i40e_test_staterr(union i40e_rx_desc *rx_desc,
#define I40E_MAX_DATA_PER_TXD_ALIGNED \ #define I40E_MAX_DATA_PER_TXD_ALIGNED \
(I40E_MAX_DATA_PER_TXD & ~(I40E_MAX_READ_REQ_SIZE - 1)) (I40E_MAX_DATA_PER_TXD & ~(I40E_MAX_READ_REQ_SIZE - 1))
/* This ugly bit of math is equivalent to DIV_ROUNDUP(size, X) where X is /**
* the value I40E_MAX_DATA_PER_TXD_ALIGNED. It is needed due to the fact * i40e_txd_use_count - estimate the number of descriptors needed for Tx
* that 12K is not a power of 2 and division is expensive. It is used to * @size: transmit request size in bytes
* approximate the number of descriptors used per linear buffer. Note *
* that this will overestimate in some cases as it doesn't account for the * Due to hardware alignment restrictions (4K alignment), we need to
* fact that we will add up to 4K - 1 in aligning the 12K buffer, however * assume that we can have no more than 12K of data per descriptor, even
* the error should not impact things much as large buffers usually mean * though each descriptor can take up to 16K - 1 bytes of aligned memory.
* we will use fewer descriptors then there are frags in an skb. * Thus, we need to divide by 12K. But division is slow! Instead,
* we decompose the operation into shifts and one relatively cheap
* multiply operation.
*
* To divide by 12K, we first divide by 4K, then divide by 3:
* To divide by 4K, shift right by 12 bits
* To divide by 3, multiply by 85, then divide by 256
* (Divide by 256 is done by shifting right by 8 bits)
* Finally, we add one to round up. Because 256 isn't an exact multiple of
* 3, we'll underestimate near each multiple of 12K. This is actually more
* accurate as we have 4K - 1 of wiggle room that we can fit into the last
* segment. For our purposes this is accurate out to 1M which is orders of
* magnitude greater than our largest possible GSO size.
*
* This would then be implemented as:
* return (((size >> 12) * 85) >> 8) + 1;
*
* Since multiplication and division are commutative, we can reorder
* operations into:
* return ((size * 85) >> 20) + 1;
*/ */
static inline unsigned int i40e_txd_use_count(unsigned int size) static inline unsigned int i40e_txd_use_count(unsigned int size)
{ {
const unsigned int max = I40E_MAX_DATA_PER_TXD_ALIGNED; return ((size * 85) >> 20) + 1;
const unsigned int reciprocal = ((1ull << 32) - 1 + (max / 2)) / max;
unsigned int adjust = ~(u32)0;
/* if we rounded up on the reciprocal pull down the adjustment */
if ((max * reciprocal) > adjust)
adjust = ~(u32)(reciprocal - 1);
return (u32)((((u64)size * reciprocal) + adjust) >> 32);
} }
/* Tx Descriptors needed, worst case */ /* Tx Descriptors needed, worst case */
......
...@@ -173,26 +173,37 @@ static inline bool i40e_test_staterr(union i40e_rx_desc *rx_desc, ...@@ -173,26 +173,37 @@ static inline bool i40e_test_staterr(union i40e_rx_desc *rx_desc,
#define I40E_MAX_DATA_PER_TXD_ALIGNED \ #define I40E_MAX_DATA_PER_TXD_ALIGNED \
(I40E_MAX_DATA_PER_TXD & ~(I40E_MAX_READ_REQ_SIZE - 1)) (I40E_MAX_DATA_PER_TXD & ~(I40E_MAX_READ_REQ_SIZE - 1))
/* This ugly bit of math is equivalent to DIV_ROUNDUP(size, X) where X is /**
* the value I40E_MAX_DATA_PER_TXD_ALIGNED. It is needed due to the fact * i40e_txd_use_count - estimate the number of descriptors needed for Tx
* that 12K is not a power of 2 and division is expensive. It is used to * @size: transmit request size in bytes
* approximate the number of descriptors used per linear buffer. Note *
* that this will overestimate in some cases as it doesn't account for the * Due to hardware alignment restrictions (4K alignment), we need to
* fact that we will add up to 4K - 1 in aligning the 12K buffer, however * assume that we can have no more than 12K of data per descriptor, even
* the error should not impact things much as large buffers usually mean * though each descriptor can take up to 16K - 1 bytes of aligned memory.
* we will use fewer descriptors then there are frags in an skb. * Thus, we need to divide by 12K. But division is slow! Instead,
* we decompose the operation into shifts and one relatively cheap
* multiply operation.
*
* To divide by 12K, we first divide by 4K, then divide by 3:
* To divide by 4K, shift right by 12 bits
* To divide by 3, multiply by 85, then divide by 256
* (Divide by 256 is done by shifting right by 8 bits)
* Finally, we add one to round up. Because 256 isn't an exact multiple of
* 3, we'll underestimate near each multiple of 12K. This is actually more
* accurate as we have 4K - 1 of wiggle room that we can fit into the last
* segment. For our purposes this is accurate out to 1M which is orders of
* magnitude greater than our largest possible GSO size.
*
* This would then be implemented as:
* return (((size >> 12) * 85) >> 8) + 1;
*
* Since multiplication and division are commutative, we can reorder
* operations into:
* return ((size * 85) >> 20) + 1;
*/ */
static inline unsigned int i40e_txd_use_count(unsigned int size) static inline unsigned int i40e_txd_use_count(unsigned int size)
{ {
const unsigned int max = I40E_MAX_DATA_PER_TXD_ALIGNED; return ((size * 85) >> 20) + 1;
const unsigned int reciprocal = ((1ull << 32) - 1 + (max / 2)) / max;
unsigned int adjust = ~(u32)0;
/* if we rounded up on the reciprocal pull down the adjustment */
if ((max * reciprocal) > adjust)
adjust = ~(u32)(reciprocal - 1);
return (u32)((((u64)size * reciprocal) + adjust) >> 32);
} }
/* Tx Descriptors needed, worst case */ /* Tx Descriptors needed, worst case */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment