Commit 86c0c196 authored by Camelia Groza's avatar Camelia Groza Committed by Jakub Kicinski

dpaa_eth: add basic XDP support

Implement the XDP_DROP and XDP_PASS actions.

Avoid draining and reconfiguring the buffer pool at each XDP
setup/teardown by increasing the frame headroom and reserving
XDP_PACKET_HEADROOM bytes from the start. Since we always reserve an
entire page per buffer, this change only impacts Jumbo frame scenarios
where the maximum linear frame size is reduced by 256 bytes. Multi
buffer Scatter/Gather frames are now used instead in these scenarios.

Allow XDP programs to access the entire buffer.

The data in the received frame's headroom can be overwritten by the XDP
program. Extract the relevant fields from the headroom while they are
still available, before running the XDP program.

Since the headroom might be resized before the frame is passed up to the
stack, remove the check for a fixed headroom value when building an skb.

Allow the meta data to be updated and pass the information up the stack.

Scatter/Gather frames are dropped when XDP is enabled.
Acked-by: default avatarMadalin Bucur <madalin.bucur@oss.nxp.com>
Signed-off-by: default avatarCamelia Groza <camelia.groza@nxp.com>
Reviewed-by: default avatarMaciej Fijalkowski <maciej.fijalkowski@intel.com>
Signed-off-by: default avatarJakub Kicinski <kuba@kernel.org>
parent fb9afd96
...@@ -53,6 +53,8 @@ ...@@ -53,6 +53,8 @@
#include <linux/dma-mapping.h> #include <linux/dma-mapping.h>
#include <linux/sort.h> #include <linux/sort.h>
#include <linux/phy_fixed.h> #include <linux/phy_fixed.h>
#include <linux/bpf.h>
#include <linux/bpf_trace.h>
#include <soc/fsl/bman.h> #include <soc/fsl/bman.h>
#include <soc/fsl/qman.h> #include <soc/fsl/qman.h>
#include "fman.h" #include "fman.h"
...@@ -177,7 +179,7 @@ MODULE_PARM_DESC(tx_timeout, "The Tx timeout in ms"); ...@@ -177,7 +179,7 @@ MODULE_PARM_DESC(tx_timeout, "The Tx timeout in ms");
#define DPAA_HWA_SIZE (DPAA_PARSE_RESULTS_SIZE + DPAA_TIME_STAMP_SIZE \ #define DPAA_HWA_SIZE (DPAA_PARSE_RESULTS_SIZE + DPAA_TIME_STAMP_SIZE \
+ DPAA_HASH_RESULTS_SIZE) + DPAA_HASH_RESULTS_SIZE)
#define DPAA_RX_PRIV_DATA_DEFAULT_SIZE (DPAA_TX_PRIV_DATA_SIZE + \ #define DPAA_RX_PRIV_DATA_DEFAULT_SIZE (DPAA_TX_PRIV_DATA_SIZE + \
dpaa_rx_extra_headroom) XDP_PACKET_HEADROOM - DPAA_HWA_SIZE)
#ifdef CONFIG_DPAA_ERRATUM_A050385 #ifdef CONFIG_DPAA_ERRATUM_A050385
#define DPAA_RX_PRIV_DATA_A050385_SIZE (DPAA_A050385_ALIGN - DPAA_HWA_SIZE) #define DPAA_RX_PRIV_DATA_A050385_SIZE (DPAA_A050385_ALIGN - DPAA_HWA_SIZE)
#define DPAA_RX_PRIV_DATA_SIZE (fman_has_errata_a050385() ? \ #define DPAA_RX_PRIV_DATA_SIZE (fman_has_errata_a050385() ? \
...@@ -1733,7 +1735,6 @@ static struct sk_buff *contig_fd_to_skb(const struct dpaa_priv *priv, ...@@ -1733,7 +1735,6 @@ static struct sk_buff *contig_fd_to_skb(const struct dpaa_priv *priv,
SKB_DATA_ALIGN(sizeof(struct skb_shared_info))); SKB_DATA_ALIGN(sizeof(struct skb_shared_info)));
if (WARN_ONCE(!skb, "Build skb failure on Rx\n")) if (WARN_ONCE(!skb, "Build skb failure on Rx\n"))
goto free_buffer; goto free_buffer;
WARN_ON(fd_off != priv->rx_headroom);
skb_reserve(skb, fd_off); skb_reserve(skb, fd_off);
skb_put(skb, qm_fd_get_length(fd)); skb_put(skb, qm_fd_get_length(fd));
...@@ -2349,12 +2350,62 @@ static enum qman_cb_dqrr_result rx_error_dqrr(struct qman_portal *portal, ...@@ -2349,12 +2350,62 @@ static enum qman_cb_dqrr_result rx_error_dqrr(struct qman_portal *portal,
return qman_cb_dqrr_consume; return qman_cb_dqrr_consume;
} }
static u32 dpaa_run_xdp(struct dpaa_priv *priv, struct qm_fd *fd, void *vaddr,
unsigned int *xdp_meta_len)
{
ssize_t fd_off = qm_fd_get_offset(fd);
struct bpf_prog *xdp_prog;
struct xdp_buff xdp;
u32 xdp_act;
rcu_read_lock();
xdp_prog = READ_ONCE(priv->xdp_prog);
if (!xdp_prog) {
rcu_read_unlock();
return XDP_PASS;
}
xdp.data = vaddr + fd_off;
xdp.data_meta = xdp.data;
xdp.data_hard_start = xdp.data - XDP_PACKET_HEADROOM;
xdp.data_end = xdp.data + qm_fd_get_length(fd);
xdp.frame_sz = DPAA_BP_RAW_SIZE;
xdp_act = bpf_prog_run_xdp(xdp_prog, &xdp);
/* Update the length and the offset of the FD */
qm_fd_set_contig(fd, xdp.data - vaddr, xdp.data_end - xdp.data);
switch (xdp_act) {
case XDP_PASS:
*xdp_meta_len = xdp.data - xdp.data_meta;
break;
default:
bpf_warn_invalid_xdp_action(xdp_act);
fallthrough;
case XDP_ABORTED:
trace_xdp_exception(priv->net_dev, xdp_prog, xdp_act);
fallthrough;
case XDP_DROP:
/* Free the buffer */
free_pages((unsigned long)vaddr, 0);
break;
}
rcu_read_unlock();
return xdp_act;
}
static enum qman_cb_dqrr_result rx_default_dqrr(struct qman_portal *portal, static enum qman_cb_dqrr_result rx_default_dqrr(struct qman_portal *portal,
struct qman_fq *fq, struct qman_fq *fq,
const struct qm_dqrr_entry *dq, const struct qm_dqrr_entry *dq,
bool sched_napi) bool sched_napi)
{ {
bool ts_valid = false, hash_valid = false;
struct skb_shared_hwtstamps *shhwtstamps; struct skb_shared_hwtstamps *shhwtstamps;
unsigned int skb_len, xdp_meta_len = 0;
struct rtnl_link_stats64 *percpu_stats; struct rtnl_link_stats64 *percpu_stats;
struct dpaa_percpu_priv *percpu_priv; struct dpaa_percpu_priv *percpu_priv;
const struct qm_fd *fd = &dq->fd; const struct qm_fd *fd = &dq->fd;
...@@ -2362,12 +2413,14 @@ static enum qman_cb_dqrr_result rx_default_dqrr(struct qman_portal *portal, ...@@ -2362,12 +2413,14 @@ static enum qman_cb_dqrr_result rx_default_dqrr(struct qman_portal *portal,
enum qm_fd_format fd_format; enum qm_fd_format fd_format;
struct net_device *net_dev; struct net_device *net_dev;
u32 fd_status, hash_offset; u32 fd_status, hash_offset;
struct qm_sg_entry *sgt;
struct dpaa_bp *dpaa_bp; struct dpaa_bp *dpaa_bp;
struct dpaa_priv *priv; struct dpaa_priv *priv;
unsigned int skb_len;
struct sk_buff *skb; struct sk_buff *skb;
int *count_ptr; int *count_ptr;
u32 xdp_act;
void *vaddr; void *vaddr;
u32 hash;
u64 ns; u64 ns;
fd_status = be32_to_cpu(fd->status); fd_status = be32_to_cpu(fd->status);
...@@ -2423,35 +2476,67 @@ static enum qman_cb_dqrr_result rx_default_dqrr(struct qman_portal *portal, ...@@ -2423,35 +2476,67 @@ static enum qman_cb_dqrr_result rx_default_dqrr(struct qman_portal *portal,
count_ptr = this_cpu_ptr(dpaa_bp->percpu_count); count_ptr = this_cpu_ptr(dpaa_bp->percpu_count);
(*count_ptr)--; (*count_ptr)--;
if (likely(fd_format == qm_fd_contig)) /* Extract the timestamp stored in the headroom before running XDP */
if (priv->rx_tstamp) {
if (!fman_port_get_tstamp(priv->mac_dev->port[RX], vaddr, &ns))
ts_valid = true;
else
WARN_ONCE(1, "fman_port_get_tstamp failed!\n");
}
/* Extract the hash stored in the headroom before running XDP */
if (net_dev->features & NETIF_F_RXHASH && priv->keygen_in_use &&
!fman_port_get_hash_result_offset(priv->mac_dev->port[RX],
&hash_offset)) {
hash = be32_to_cpu(*(u32 *)(vaddr + hash_offset));
hash_valid = true;
}
if (likely(fd_format == qm_fd_contig)) {
xdp_act = dpaa_run_xdp(priv, (struct qm_fd *)fd, vaddr,
&xdp_meta_len);
if (xdp_act != XDP_PASS) {
percpu_stats->rx_packets++;
percpu_stats->rx_bytes += qm_fd_get_length(fd);
return qman_cb_dqrr_consume;
}
skb = contig_fd_to_skb(priv, fd); skb = contig_fd_to_skb(priv, fd);
else } else {
/* XDP doesn't support S/G frames. Return the fragments to the
* buffer pool and release the SGT.
*/
if (READ_ONCE(priv->xdp_prog)) {
WARN_ONCE(1, "S/G frames not supported under XDP\n");
sgt = vaddr + qm_fd_get_offset(fd);
dpaa_release_sgt_members(sgt);
free_pages((unsigned long)vaddr, 0);
return qman_cb_dqrr_consume;
}
skb = sg_fd_to_skb(priv, fd); skb = sg_fd_to_skb(priv, fd);
}
if (!skb) if (!skb)
return qman_cb_dqrr_consume; return qman_cb_dqrr_consume;
if (priv->rx_tstamp) { if (xdp_meta_len)
skb_metadata_set(skb, xdp_meta_len);
/* Set the previously extracted timestamp */
if (ts_valid) {
shhwtstamps = skb_hwtstamps(skb); shhwtstamps = skb_hwtstamps(skb);
memset(shhwtstamps, 0, sizeof(*shhwtstamps)); memset(shhwtstamps, 0, sizeof(*shhwtstamps));
shhwtstamps->hwtstamp = ns_to_ktime(ns);
if (!fman_port_get_tstamp(priv->mac_dev->port[RX], vaddr, &ns))
shhwtstamps->hwtstamp = ns_to_ktime(ns);
else
dev_warn(net_dev->dev.parent, "fman_port_get_tstamp failed!\n");
} }
skb->protocol = eth_type_trans(skb, net_dev); skb->protocol = eth_type_trans(skb, net_dev);
if (net_dev->features & NETIF_F_RXHASH && priv->keygen_in_use && /* Set the previously extracted hash */
!fman_port_get_hash_result_offset(priv->mac_dev->port[RX], if (hash_valid) {
&hash_offset)) {
enum pkt_hash_types type; enum pkt_hash_types type;
/* if L4 exists, it was used in the hash generation */ /* if L4 exists, it was used in the hash generation */
type = be32_to_cpu(fd->status) & FM_FD_STAT_L4CV ? type = be32_to_cpu(fd->status) & FM_FD_STAT_L4CV ?
PKT_HASH_TYPE_L4 : PKT_HASH_TYPE_L3; PKT_HASH_TYPE_L4 : PKT_HASH_TYPE_L3;
skb_set_hash(skb, be32_to_cpu(*(u32 *)(vaddr + hash_offset)), skb_set_hash(skb, hash, type);
type);
} }
skb_len = skb->len; skb_len = skb->len;
...@@ -2671,6 +2756,55 @@ static int dpaa_eth_stop(struct net_device *net_dev) ...@@ -2671,6 +2756,55 @@ static int dpaa_eth_stop(struct net_device *net_dev)
return err; return err;
} }
static int dpaa_setup_xdp(struct net_device *net_dev, struct netdev_bpf *bpf)
{
struct dpaa_priv *priv = netdev_priv(net_dev);
struct bpf_prog *old_prog;
int err, max_contig_data;
bool up;
max_contig_data = priv->dpaa_bp->size - priv->rx_headroom;
/* S/G fragments are not supported in XDP-mode */
if (bpf->prog &&
(net_dev->mtu + VLAN_ETH_HLEN + ETH_FCS_LEN > max_contig_data)) {
NL_SET_ERR_MSG_MOD(bpf->extack, "MTU too large for XDP");
dev_warn(net_dev->dev.parent,
"The maximum MTU for XDP is %d\n",
max_contig_data - VLAN_ETH_HLEN - ETH_FCS_LEN);
return -EINVAL;
}
up = netif_running(net_dev);
if (up)
dpaa_eth_stop(net_dev);
old_prog = xchg(&priv->xdp_prog, bpf->prog);
if (old_prog)
bpf_prog_put(old_prog);
if (up) {
err = dpaa_open(net_dev);
if (err) {
NL_SET_ERR_MSG_MOD(bpf->extack, "dpaa_open() failed");
return err;
}
}
return 0;
}
static int dpaa_xdp(struct net_device *net_dev, struct netdev_bpf *xdp)
{
switch (xdp->command) {
case XDP_SETUP_PROG:
return dpaa_setup_xdp(net_dev, xdp);
default:
return -EINVAL;
}
}
static int dpaa_ts_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) static int dpaa_ts_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
{ {
struct dpaa_priv *priv = netdev_priv(dev); struct dpaa_priv *priv = netdev_priv(dev);
...@@ -2737,6 +2871,7 @@ static const struct net_device_ops dpaa_ops = { ...@@ -2737,6 +2871,7 @@ static const struct net_device_ops dpaa_ops = {
.ndo_set_rx_mode = dpaa_set_rx_mode, .ndo_set_rx_mode = dpaa_set_rx_mode,
.ndo_do_ioctl = dpaa_ioctl, .ndo_do_ioctl = dpaa_ioctl,
.ndo_setup_tc = dpaa_setup_tc, .ndo_setup_tc = dpaa_setup_tc,
.ndo_bpf = dpaa_xdp,
}; };
static int dpaa_napi_add(struct net_device *net_dev) static int dpaa_napi_add(struct net_device *net_dev)
......
...@@ -196,6 +196,8 @@ struct dpaa_priv { ...@@ -196,6 +196,8 @@ struct dpaa_priv {
bool tx_tstamp; /* Tx timestamping enabled */ bool tx_tstamp; /* Tx timestamping enabled */
bool rx_tstamp; /* Rx timestamping enabled */ bool rx_tstamp; /* Rx timestamping enabled */
struct bpf_prog *xdp_prog;
}; };
/* from dpaa_ethtool.c */ /* from dpaa_ethtool.c */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment