mirror of
https://github.com/openwrt/openwrt.git
synced 2024-12-29 10:08:59 +00:00
f6a7ce2501
Fixes: 6262d3eb06
("mac80211: sync mesh fast xmit patch with upstream requested changes")
Signed-off-by: Felix Fietkau <nbd@nbd.name>
851 lines
26 KiB
Diff
851 lines
26 KiB
Diff
From: Felix Fietkau <nbd@nbd.name>
|
|
Date: Sun, 26 Feb 2023 13:53:08 +0100
|
|
Subject: [PATCH] wifi: mac80211: mesh fast xmit support
|
|
|
|
Previously, fast xmit only worked on interface types where initially a
|
|
sta lookup is performed, and a cached header can be attached to the sta,
|
|
requiring only some fields to be updated at runtime.
|
|
|
|
This technique is not directly applicable for a mesh device type due
|
|
to the dynamic nature of the topology and protocol. There are more
|
|
addresses that need to be filled, and there is an extra header with a
|
|
dynamic length based on the addressing mode.
|
|
|
|
Change the code to cache entries contain a copy of the mesh subframe header +
|
|
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
|
|
contains the information for building the 802.11 header.
|
|
|
|
Add a mesh specific early fast xmit call, which looks up a cached entry and
|
|
adds only the mesh subframe header, before passing it over to the generic
|
|
fast xmit code.
|
|
|
|
To ensure the changes in network are reflected in these cached headers,
|
|
flush affected cached entries on path changes, as well as other conditions
|
|
that currently trigger a fast xmit check in other modes (key changes etc.)
|
|
|
|
This code is loosely based on a previous implementation by:
|
|
Sriram R <quic_srirrama@quicinc.com>
|
|
|
|
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
|
|
Signed-off-by: Felix Fietkau <nbd@nbd.name>
|
|
---
|
|
|
|
--- a/net/mac80211/ieee80211_i.h
|
|
+++ b/net/mac80211/ieee80211_i.h
|
|
@@ -37,6 +37,7 @@
|
|
extern const struct cfg80211_ops mac80211_config_ops;
|
|
|
|
struct ieee80211_local;
|
|
+struct ieee80211_mesh_fast_tx;
|
|
|
|
/* Maximum number of broadcast/multicast frames to buffer when some of the
|
|
* associated stations are using power saving. */
|
|
@@ -655,6 +656,19 @@ struct mesh_table {
|
|
atomic_t entries; /* Up to MAX_MESH_NEIGHBOURS */
|
|
};
|
|
|
|
+/**
|
|
+ * struct mesh_tx_cache - mesh fast xmit header cache
|
|
+ *
|
|
+ * @rht: hash table containing struct ieee80211_mesh_fast_tx, using skb DA as key
|
|
+ * @walk_head: linked list containing all ieee80211_mesh_fast_tx objects
|
|
+ * @walk_lock: lock protecting walk_head and rht
|
|
+ */
|
|
+struct mesh_tx_cache {
|
|
+ struct rhashtable rht;
|
|
+ struct hlist_head walk_head;
|
|
+ spinlock_t walk_lock;
|
|
+};
|
|
+
|
|
struct ieee80211_if_mesh {
|
|
struct timer_list housekeeping_timer;
|
|
struct timer_list mesh_path_timer;
|
|
@@ -733,6 +747,7 @@ struct ieee80211_if_mesh {
|
|
struct mesh_table mpp_paths; /* Store paths for MPP&MAP */
|
|
int mesh_paths_generation;
|
|
int mpp_paths_generation;
|
|
+ struct mesh_tx_cache tx_cache;
|
|
};
|
|
|
|
#ifdef CPTCFG_MAC80211_MESH
|
|
@@ -1998,6 +2013,11 @@ int ieee80211_tx_control_port(struct wip
|
|
int link_id, u64 *cookie);
|
|
int ieee80211_probe_mesh_link(struct wiphy *wiphy, struct net_device *dev,
|
|
const u8 *buf, size_t len);
|
|
+void __ieee80211_xmit_fast(struct ieee80211_sub_if_data *sdata,
|
|
+ struct sta_info *sta,
|
|
+ struct ieee80211_fast_tx *fast_tx,
|
|
+ struct sk_buff *skb, bool ampdu,
|
|
+ const u8 *da, const u8 *sa);
|
|
|
|
/* HT */
|
|
void ieee80211_apply_htcap_overrides(struct ieee80211_sub_if_data *sdata,
|
|
--- a/net/mac80211/mesh.c
|
|
+++ b/net/mac80211/mesh.c
|
|
@@ -10,6 +10,7 @@
|
|
#include <asm/unaligned.h>
|
|
#include "ieee80211_i.h"
|
|
#include "mesh.h"
|
|
+#include "wme.h"
|
|
#include "driver-ops.h"
|
|
|
|
static int mesh_allocated;
|
|
@@ -698,6 +699,95 @@ ieee80211_mesh_update_bss_params(struct
|
|
__le32_to_cpu(he_oper->he_oper_params);
|
|
}
|
|
|
|
+bool ieee80211_mesh_xmit_fast(struct ieee80211_sub_if_data *sdata,
|
|
+ struct sk_buff *skb, u32 ctrl_flags)
|
|
+{
|
|
+ struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
|
|
+ struct ieee80211_mesh_fast_tx *entry;
|
|
+ struct ieee80211s_hdr *meshhdr;
|
|
+ u8 sa[ETH_ALEN] __aligned(2);
|
|
+ struct tid_ampdu_tx *tid_tx;
|
|
+ struct sta_info *sta;
|
|
+ bool copy_sa = false;
|
|
+ u16 ethertype;
|
|
+ u8 tid;
|
|
+
|
|
+ if (ctrl_flags & IEEE80211_TX_CTRL_SKIP_MPATH_LOOKUP)
|
|
+ return false;
|
|
+
|
|
+ if (ifmsh->mshcfg.dot11MeshNolearn)
|
|
+ return false;
|
|
+
|
|
+ /* Add support for these cases later */
|
|
+ if (ifmsh->ps_peers_light_sleep || ifmsh->ps_peers_deep_sleep)
|
|
+ return false;
|
|
+
|
|
+ if (is_multicast_ether_addr(skb->data))
|
|
+ return false;
|
|
+
|
|
+ ethertype = (skb->data[12] << 8) | skb->data[13];
|
|
+ if (ethertype < ETH_P_802_3_MIN)
|
|
+ return false;
|
|
+
|
|
+ if (skb->sk && skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS)
|
|
+ return false;
|
|
+
|
|
+ if (skb->ip_summed == CHECKSUM_PARTIAL) {
|
|
+ skb_set_transport_header(skb, skb_checksum_start_offset(skb));
|
|
+ if (skb_checksum_help(skb))
|
|
+ return false;
|
|
+ }
|
|
+
|
|
+ entry = mesh_fast_tx_get(sdata, skb->data);
|
|
+ if (!entry)
|
|
+ return false;
|
|
+
|
|
+ if (skb_headroom(skb) < entry->hdrlen + entry->fast_tx.hdr_len)
|
|
+ return false;
|
|
+
|
|
+ sta = rcu_dereference(entry->mpath->next_hop);
|
|
+ if (!sta)
|
|
+ return false;
|
|
+
|
|
+ tid = skb->priority & IEEE80211_QOS_CTL_TAG1D_MASK;
|
|
+ tid_tx = rcu_dereference(sta->ampdu_mlme.tid_tx[tid]);
|
|
+ if (tid_tx) {
|
|
+ if (!test_bit(HT_AGG_STATE_OPERATIONAL, &tid_tx->state))
|
|
+ return false;
|
|
+ if (tid_tx->timeout)
|
|
+ tid_tx->last_tx = jiffies;
|
|
+ }
|
|
+
|
|
+ skb = skb_share_check(skb, GFP_ATOMIC);
|
|
+ if (!skb)
|
|
+ return true;
|
|
+
|
|
+ skb_set_queue_mapping(skb, ieee80211_select_queue(sdata, sta, skb));
|
|
+
|
|
+ meshhdr = (struct ieee80211s_hdr *)entry->hdr;
|
|
+ if ((meshhdr->flags & MESH_FLAGS_AE) == MESH_FLAGS_AE_A5_A6) {
|
|
+ /* preserve SA from eth header for 6-addr frames */
|
|
+ ether_addr_copy(sa, skb->data + ETH_ALEN);
|
|
+ copy_sa = true;
|
|
+ }
|
|
+
|
|
+ memcpy(skb_push(skb, entry->hdrlen - 2 * ETH_ALEN), entry->hdr,
|
|
+ entry->hdrlen);
|
|
+
|
|
+ meshhdr = (struct ieee80211s_hdr *)skb->data;
|
|
+ put_unaligned_le32(atomic_inc_return(&sdata->u.mesh.mesh_seqnum),
|
|
+ &meshhdr->seqnum);
|
|
+ meshhdr->ttl = sdata->u.mesh.mshcfg.dot11MeshTTL;
|
|
+ if (copy_sa)
|
|
+ ether_addr_copy(meshhdr->eaddr2, sa);
|
|
+
|
|
+ skb_push(skb, 2 * ETH_ALEN);
|
|
+ __ieee80211_xmit_fast(sdata, sta, &entry->fast_tx, skb, tid_tx,
|
|
+ entry->mpath->dst, sdata->vif.addr);
|
|
+
|
|
+ return true;
|
|
+}
|
|
+
|
|
/**
|
|
* ieee80211_fill_mesh_addresses - fill addresses of a locally originated mesh frame
|
|
* @hdr: 802.11 frame header
|
|
@@ -780,6 +870,8 @@ static void ieee80211_mesh_housekeeping(
|
|
changed = mesh_accept_plinks_update(sdata);
|
|
ieee80211_mbss_info_change_notify(sdata, changed);
|
|
|
|
+ mesh_fast_tx_gc(sdata);
|
|
+
|
|
mod_timer(&ifmsh->housekeeping_timer,
|
|
round_jiffies(jiffies +
|
|
IEEE80211_MESH_HOUSEKEEPING_INTERVAL));
|
|
--- a/net/mac80211/mesh.h
|
|
+++ b/net/mac80211/mesh.h
|
|
@@ -122,11 +122,41 @@ struct mesh_path {
|
|
u8 rann_snd_addr[ETH_ALEN];
|
|
u32 rann_metric;
|
|
unsigned long last_preq_to_root;
|
|
+ unsigned long fast_tx_check;
|
|
bool is_root;
|
|
bool is_gate;
|
|
u32 path_change_count;
|
|
};
|
|
|
|
+#define MESH_FAST_TX_CACHE_MAX_SIZE 512
|
|
+#define MESH_FAST_TX_CACHE_THRESHOLD_SIZE 384
|
|
+#define MESH_FAST_TX_CACHE_TIMEOUT 8000 /* msecs */
|
|
+
|
|
+/**
|
|
+ * struct ieee80211_mesh_fast_tx - cached mesh fast tx entry
|
|
+ * @rhash: rhashtable pointer
|
|
+ * @addr_key: The Ethernet DA which is the key for this entry
|
|
+ * @fast_tx: base fast_tx data
|
|
+ * @hdr: cached mesh and rfc1042 headers
|
|
+ * @hdrlen: length of mesh + rfc1042
|
|
+ * @walk_list: list containing all the fast tx entries
|
|
+ * @mpath: mesh path corresponding to the Mesh DA
|
|
+ * @mppath: MPP entry corresponding to this DA
|
|
+ * @timestamp: Last used time of this entry
|
|
+ */
|
|
+struct ieee80211_mesh_fast_tx {
|
|
+ struct rhash_head rhash;
|
|
+ u8 addr_key[ETH_ALEN] __aligned(2);
|
|
+
|
|
+ struct ieee80211_fast_tx fast_tx;
|
|
+ u8 hdr[sizeof(struct ieee80211s_hdr) + sizeof(rfc1042_header)];
|
|
+ u16 hdrlen;
|
|
+
|
|
+ struct mesh_path *mpath, *mppath;
|
|
+ struct hlist_node walk_list;
|
|
+ unsigned long timestamp;
|
|
+};
|
|
+
|
|
/* Recent multicast cache */
|
|
/* RMC_BUCKETS must be a power of 2, maximum 256 */
|
|
#define RMC_BUCKETS 256
|
|
@@ -298,6 +328,20 @@ void mesh_path_discard_frame(struct ieee
|
|
void mesh_path_tx_root_frame(struct ieee80211_sub_if_data *sdata);
|
|
|
|
bool mesh_action_is_path_sel(struct ieee80211_mgmt *mgmt);
|
|
+struct ieee80211_mesh_fast_tx *
|
|
+mesh_fast_tx_get(struct ieee80211_sub_if_data *sdata, const u8 *addr);
|
|
+bool ieee80211_mesh_xmit_fast(struct ieee80211_sub_if_data *sdata,
|
|
+ struct sk_buff *skb, u32 ctrl_flags);
|
|
+void mesh_fast_tx_cache(struct ieee80211_sub_if_data *sdata,
|
|
+ struct sk_buff *skb, struct mesh_path *mpath);
|
|
+void mesh_fast_tx_gc(struct ieee80211_sub_if_data *sdata);
|
|
+void mesh_fast_tx_flush_addr(struct ieee80211_sub_if_data *sdata,
|
|
+ const u8 *addr);
|
|
+void mesh_fast_tx_flush_mpath(struct mesh_path *mpath);
|
|
+void mesh_fast_tx_flush_sta(struct ieee80211_sub_if_data *sdata,
|
|
+ struct sta_info *sta);
|
|
+void mesh_path_refresh(struct ieee80211_sub_if_data *sdata,
|
|
+ struct mesh_path *mpath, const u8 *addr);
|
|
|
|
#ifdef CPTCFG_MAC80211_MESH
|
|
static inline
|
|
--- a/net/mac80211/mesh_hwmp.c
|
|
+++ b/net/mac80211/mesh_hwmp.c
|
|
@@ -394,6 +394,7 @@ static u32 hwmp_route_info_get(struct ie
|
|
u32 orig_sn, orig_metric;
|
|
unsigned long orig_lifetime, exp_time;
|
|
u32 last_hop_metric, new_metric;
|
|
+ bool flush_mpath = false;
|
|
bool process = true;
|
|
u8 hopcount;
|
|
|
|
@@ -491,8 +492,10 @@ static u32 hwmp_route_info_get(struct ie
|
|
}
|
|
|
|
if (fresh_info) {
|
|
- if (rcu_access_pointer(mpath->next_hop) != sta)
|
|
+ if (rcu_access_pointer(mpath->next_hop) != sta) {
|
|
mpath->path_change_count++;
|
|
+ flush_mpath = true;
|
|
+ }
|
|
mesh_path_assign_nexthop(mpath, sta);
|
|
mpath->flags |= MESH_PATH_SN_VALID;
|
|
mpath->metric = new_metric;
|
|
@@ -502,6 +505,8 @@ static u32 hwmp_route_info_get(struct ie
|
|
mpath->hop_count = hopcount;
|
|
mesh_path_activate(mpath);
|
|
spin_unlock_bh(&mpath->state_lock);
|
|
+ if (flush_mpath)
|
|
+ mesh_fast_tx_flush_mpath(mpath);
|
|
ewma_mesh_fail_avg_init(&sta->mesh->fail_avg);
|
|
/* init it at a low value - 0 start is tricky */
|
|
ewma_mesh_fail_avg_add(&sta->mesh->fail_avg, 1);
|
|
@@ -539,8 +544,10 @@ static u32 hwmp_route_info_get(struct ie
|
|
}
|
|
|
|
if (fresh_info) {
|
|
- if (rcu_access_pointer(mpath->next_hop) != sta)
|
|
+ if (rcu_access_pointer(mpath->next_hop) != sta) {
|
|
mpath->path_change_count++;
|
|
+ flush_mpath = true;
|
|
+ }
|
|
mesh_path_assign_nexthop(mpath, sta);
|
|
mpath->metric = last_hop_metric;
|
|
mpath->exp_time = time_after(mpath->exp_time, exp_time)
|
|
@@ -548,6 +555,8 @@ static u32 hwmp_route_info_get(struct ie
|
|
mpath->hop_count = 1;
|
|
mesh_path_activate(mpath);
|
|
spin_unlock_bh(&mpath->state_lock);
|
|
+ if (flush_mpath)
|
|
+ mesh_fast_tx_flush_mpath(mpath);
|
|
ewma_mesh_fail_avg_init(&sta->mesh->fail_avg);
|
|
/* init it at a low value - 0 start is tricky */
|
|
ewma_mesh_fail_avg_add(&sta->mesh->fail_avg, 1);
|
|
@@ -1215,6 +1224,20 @@ static int mesh_nexthop_lookup_nolearn(s
|
|
return 0;
|
|
}
|
|
|
|
+void mesh_path_refresh(struct ieee80211_sub_if_data *sdata,
|
|
+ struct mesh_path *mpath, const u8 *addr)
|
|
+{
|
|
+ if (mpath->flags & (MESH_PATH_REQ_QUEUED | MESH_PATH_FIXED |
|
|
+ MESH_PATH_RESOLVING))
|
|
+ return;
|
|
+
|
|
+ if (time_after(jiffies,
|
|
+ mpath->exp_time -
|
|
+ msecs_to_jiffies(sdata->u.mesh.mshcfg.path_refresh_time)) &&
|
|
+ (!addr || ether_addr_equal(sdata->vif.addr, addr)))
|
|
+ mesh_queue_preq(mpath, PREQ_Q_F_START | PREQ_Q_F_REFRESH);
|
|
+}
|
|
+
|
|
/**
|
|
* mesh_nexthop_lookup - put the appropriate next hop on a mesh frame. Calling
|
|
* this function is considered "using" the associated mpath, so preempt a path
|
|
@@ -1242,19 +1265,15 @@ int mesh_nexthop_lookup(struct ieee80211
|
|
if (!mpath || !(mpath->flags & MESH_PATH_ACTIVE))
|
|
return -ENOENT;
|
|
|
|
- if (time_after(jiffies,
|
|
- mpath->exp_time -
|
|
- msecs_to_jiffies(sdata->u.mesh.mshcfg.path_refresh_time)) &&
|
|
- ether_addr_equal(sdata->vif.addr, hdr->addr4) &&
|
|
- !(mpath->flags & MESH_PATH_RESOLVING) &&
|
|
- !(mpath->flags & MESH_PATH_FIXED))
|
|
- mesh_queue_preq(mpath, PREQ_Q_F_START | PREQ_Q_F_REFRESH);
|
|
+ mesh_path_refresh(sdata, mpath, hdr->addr4);
|
|
|
|
next_hop = rcu_dereference(mpath->next_hop);
|
|
if (next_hop) {
|
|
memcpy(hdr->addr1, next_hop->sta.addr, ETH_ALEN);
|
|
memcpy(hdr->addr2, sdata->vif.addr, ETH_ALEN);
|
|
ieee80211_mps_set_frame_flags(sdata, next_hop, hdr);
|
|
+ if (ieee80211_hw_check(&sdata->local->hw, SUPPORT_FAST_XMIT))
|
|
+ mesh_fast_tx_cache(sdata, skb, mpath);
|
|
return 0;
|
|
}
|
|
|
|
--- a/net/mac80211/mesh_pathtbl.c
|
|
+++ b/net/mac80211/mesh_pathtbl.c
|
|
@@ -14,6 +14,7 @@
|
|
#include "wme.h"
|
|
#include "ieee80211_i.h"
|
|
#include "mesh.h"
|
|
+#include <linux/rhashtable.h>
|
|
|
|
static void mesh_path_free_rcu(struct mesh_table *tbl, struct mesh_path *mpath);
|
|
|
|
@@ -32,6 +33,41 @@ static const struct rhashtable_params me
|
|
.hashfn = mesh_table_hash,
|
|
};
|
|
|
|
+static const struct rhashtable_params fast_tx_rht_params = {
|
|
+ .nelem_hint = 10,
|
|
+ .automatic_shrinking = true,
|
|
+ .key_len = ETH_ALEN,
|
|
+ .key_offset = offsetof(struct ieee80211_mesh_fast_tx, addr_key),
|
|
+ .head_offset = offsetof(struct ieee80211_mesh_fast_tx, rhash),
|
|
+ .hashfn = mesh_table_hash,
|
|
+};
|
|
+
|
|
+static void __mesh_fast_tx_entry_free(void *ptr, void *tblptr)
|
|
+{
|
|
+ struct ieee80211_mesh_fast_tx *entry = ptr;
|
|
+
|
|
+ kfree_rcu(entry, fast_tx.rcu_head);
|
|
+}
|
|
+
|
|
+static void mesh_fast_tx_deinit(struct ieee80211_sub_if_data *sdata)
|
|
+{
|
|
+ struct mesh_tx_cache *cache;
|
|
+
|
|
+ cache = &sdata->u.mesh.tx_cache;
|
|
+ rhashtable_free_and_destroy(&cache->rht,
|
|
+ __mesh_fast_tx_entry_free, NULL);
|
|
+}
|
|
+
|
|
+static void mesh_fast_tx_init(struct ieee80211_sub_if_data *sdata)
|
|
+{
|
|
+ struct mesh_tx_cache *cache;
|
|
+
|
|
+ cache = &sdata->u.mesh.tx_cache;
|
|
+ rhashtable_init(&cache->rht, &fast_tx_rht_params);
|
|
+ INIT_HLIST_HEAD(&cache->walk_head);
|
|
+ spin_lock_init(&cache->walk_lock);
|
|
+}
|
|
+
|
|
static inline bool mpath_expired(struct mesh_path *mpath)
|
|
{
|
|
return (mpath->flags & MESH_PATH_ACTIVE) &&
|
|
@@ -381,6 +417,243 @@ struct mesh_path *mesh_path_new(struct i
|
|
return new_mpath;
|
|
}
|
|
|
|
+static void mesh_fast_tx_entry_free(struct mesh_tx_cache *cache,
|
|
+ struct ieee80211_mesh_fast_tx *entry)
|
|
+{
|
|
+ hlist_del_rcu(&entry->walk_list);
|
|
+ rhashtable_remove_fast(&cache->rht, &entry->rhash, fast_tx_rht_params);
|
|
+ kfree_rcu(entry, fast_tx.rcu_head);
|
|
+}
|
|
+
|
|
+struct ieee80211_mesh_fast_tx *
|
|
+mesh_fast_tx_get(struct ieee80211_sub_if_data *sdata, const u8 *addr)
|
|
+{
|
|
+ struct ieee80211_mesh_fast_tx *entry;
|
|
+ struct mesh_tx_cache *cache;
|
|
+
|
|
+ cache = &sdata->u.mesh.tx_cache;
|
|
+ entry = rhashtable_lookup(&cache->rht, addr, fast_tx_rht_params);
|
|
+ if (!entry)
|
|
+ return NULL;
|
|
+
|
|
+ if (!(entry->mpath->flags & MESH_PATH_ACTIVE) ||
|
|
+ mpath_expired(entry->mpath)) {
|
|
+ spin_lock_bh(&cache->walk_lock);
|
|
+ entry = rhashtable_lookup(&cache->rht, addr, fast_tx_rht_params);
|
|
+ if (entry)
|
|
+ mesh_fast_tx_entry_free(cache, entry);
|
|
+ spin_unlock_bh(&cache->walk_lock);
|
|
+ return NULL;
|
|
+ }
|
|
+
|
|
+ mesh_path_refresh(sdata, entry->mpath, NULL);
|
|
+ if (entry->mppath)
|
|
+ entry->mppath->exp_time = jiffies;
|
|
+ entry->timestamp = jiffies;
|
|
+
|
|
+ return entry;
|
|
+}
|
|
+
|
|
+void mesh_fast_tx_cache(struct ieee80211_sub_if_data *sdata,
|
|
+ struct sk_buff *skb, struct mesh_path *mpath)
|
|
+{
|
|
+ struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data;
|
|
+ struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
|
|
+ struct ieee80211_mesh_fast_tx *entry, *prev;
|
|
+ struct ieee80211_mesh_fast_tx build = {};
|
|
+ struct ieee80211s_hdr *meshhdr;
|
|
+ struct mesh_tx_cache *cache;
|
|
+ struct ieee80211_key *key;
|
|
+ struct mesh_path *mppath;
|
|
+ struct sta_info *sta;
|
|
+ u8 *qc;
|
|
+
|
|
+ if (sdata->noack_map ||
|
|
+ !ieee80211_is_data_qos(hdr->frame_control))
|
|
+ return;
|
|
+
|
|
+ build.fast_tx.hdr_len = ieee80211_hdrlen(hdr->frame_control);
|
|
+ meshhdr = (struct ieee80211s_hdr *)(skb->data + build.fast_tx.hdr_len);
|
|
+ build.hdrlen = ieee80211_get_mesh_hdrlen(meshhdr);
|
|
+
|
|
+ cache = &sdata->u.mesh.tx_cache;
|
|
+ if (atomic_read(&cache->rht.nelems) >= MESH_FAST_TX_CACHE_MAX_SIZE)
|
|
+ return;
|
|
+
|
|
+ sta = rcu_dereference(mpath->next_hop);
|
|
+ if (!sta)
|
|
+ return;
|
|
+
|
|
+ if ((meshhdr->flags & MESH_FLAGS_AE) == MESH_FLAGS_AE_A5_A6) {
|
|
+ /* This is required to keep the mppath alive */
|
|
+ mppath = mpp_path_lookup(sdata, meshhdr->eaddr1);
|
|
+ if (!mppath)
|
|
+ return;
|
|
+ build.mppath = mppath;
|
|
+ } else if (ieee80211_has_a4(hdr->frame_control)) {
|
|
+ mppath = mpath;
|
|
+ } else {
|
|
+ return;
|
|
+ }
|
|
+
|
|
+ /* rate limit, in case fast xmit can't be enabled */
|
|
+ if (mppath->fast_tx_check == jiffies)
|
|
+ return;
|
|
+
|
|
+ mppath->fast_tx_check = jiffies;
|
|
+
|
|
+ /*
|
|
+ * Same use of the sta lock as in ieee80211_check_fast_xmit, in order
|
|
+ * to protect against concurrent sta key updates.
|
|
+ */
|
|
+ spin_lock_bh(&sta->lock);
|
|
+ key = rcu_access_pointer(sta->ptk[sta->ptk_idx]);
|
|
+ if (!key)
|
|
+ key = rcu_access_pointer(sdata->default_unicast_key);
|
|
+ build.fast_tx.key = key;
|
|
+
|
|
+ if (key) {
|
|
+ bool gen_iv, iv_spc;
|
|
+
|
|
+ gen_iv = key->conf.flags & IEEE80211_KEY_FLAG_GENERATE_IV;
|
|
+ iv_spc = key->conf.flags & IEEE80211_KEY_FLAG_PUT_IV_SPACE;
|
|
+
|
|
+ if (!(key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE) ||
|
|
+ (key->flags & KEY_FLAG_TAINTED))
|
|
+ goto unlock_sta;
|
|
+
|
|
+ switch (key->conf.cipher) {
|
|
+ case WLAN_CIPHER_SUITE_CCMP:
|
|
+ case WLAN_CIPHER_SUITE_CCMP_256:
|
|
+ if (gen_iv)
|
|
+ build.fast_tx.pn_offs = build.fast_tx.hdr_len;
|
|
+ if (gen_iv || iv_spc)
|
|
+ build.fast_tx.hdr_len += IEEE80211_CCMP_HDR_LEN;
|
|
+ break;
|
|
+ case WLAN_CIPHER_SUITE_GCMP:
|
|
+ case WLAN_CIPHER_SUITE_GCMP_256:
|
|
+ if (gen_iv)
|
|
+ build.fast_tx.pn_offs = build.fast_tx.hdr_len;
|
|
+ if (gen_iv || iv_spc)
|
|
+ build.fast_tx.hdr_len += IEEE80211_GCMP_HDR_LEN;
|
|
+ break;
|
|
+ default:
|
|
+ goto unlock_sta;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ memcpy(build.addr_key, mppath->dst, ETH_ALEN);
|
|
+ build.timestamp = jiffies;
|
|
+ build.fast_tx.band = info->band;
|
|
+ build.fast_tx.da_offs = offsetof(struct ieee80211_hdr, addr3);
|
|
+ build.fast_tx.sa_offs = offsetof(struct ieee80211_hdr, addr4);
|
|
+ build.mpath = mpath;
|
|
+ memcpy(build.hdr, meshhdr, build.hdrlen);
|
|
+ memcpy(build.hdr + build.hdrlen, rfc1042_header, sizeof(rfc1042_header));
|
|
+ build.hdrlen += sizeof(rfc1042_header);
|
|
+ memcpy(build.fast_tx.hdr, hdr, build.fast_tx.hdr_len);
|
|
+
|
|
+ hdr = (struct ieee80211_hdr *)build.fast_tx.hdr;
|
|
+ if (build.fast_tx.key)
|
|
+ hdr->frame_control |= cpu_to_le16(IEEE80211_FCTL_PROTECTED);
|
|
+
|
|
+ qc = ieee80211_get_qos_ctl(hdr);
|
|
+ qc[1] |= IEEE80211_QOS_CTL_MESH_CONTROL_PRESENT >> 8;
|
|
+
|
|
+ entry = kmemdup(&build, sizeof(build), GFP_ATOMIC);
|
|
+ if (!entry)
|
|
+ goto unlock_sta;
|
|
+
|
|
+ spin_lock(&cache->walk_lock);
|
|
+ prev = rhashtable_lookup_get_insert_fast(&cache->rht,
|
|
+ &entry->rhash,
|
|
+ fast_tx_rht_params);
|
|
+ if (unlikely(IS_ERR(prev))) {
|
|
+ kfree(entry);
|
|
+ goto unlock_cache;
|
|
+ }
|
|
+
|
|
+ /*
|
|
+ * replace any previous entry in the hash table, in case we're
|
|
+ * replacing it with a different type (e.g. mpath -> mpp)
|
|
+ */
|
|
+ if (unlikely(prev)) {
|
|
+ rhashtable_replace_fast(&cache->rht, &prev->rhash,
|
|
+ &entry->rhash, fast_tx_rht_params);
|
|
+ hlist_del_rcu(&prev->walk_list);
|
|
+ kfree_rcu(prev, fast_tx.rcu_head);
|
|
+ }
|
|
+
|
|
+ hlist_add_head(&entry->walk_list, &cache->walk_head);
|
|
+
|
|
+unlock_cache:
|
|
+ spin_unlock(&cache->walk_lock);
|
|
+unlock_sta:
|
|
+ spin_unlock_bh(&sta->lock);
|
|
+}
|
|
+
|
|
+void mesh_fast_tx_gc(struct ieee80211_sub_if_data *sdata)
|
|
+{
|
|
+ unsigned long timeout = msecs_to_jiffies(MESH_FAST_TX_CACHE_TIMEOUT);
|
|
+ struct mesh_tx_cache *cache;
|
|
+ struct ieee80211_mesh_fast_tx *entry;
|
|
+ struct hlist_node *n;
|
|
+
|
|
+ cache = &sdata->u.mesh.tx_cache;
|
|
+ if (atomic_read(&cache->rht.nelems) < MESH_FAST_TX_CACHE_THRESHOLD_SIZE)
|
|
+ return;
|
|
+
|
|
+ spin_lock_bh(&cache->walk_lock);
|
|
+ hlist_for_each_entry_safe(entry, n, &cache->walk_head, walk_list)
|
|
+ if (!time_is_after_jiffies(entry->timestamp + timeout))
|
|
+ mesh_fast_tx_entry_free(cache, entry);
|
|
+ spin_unlock_bh(&cache->walk_lock);
|
|
+}
|
|
+
|
|
+void mesh_fast_tx_flush_mpath(struct mesh_path *mpath)
|
|
+{
|
|
+ struct ieee80211_sub_if_data *sdata = mpath->sdata;
|
|
+ struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache;
|
|
+ struct ieee80211_mesh_fast_tx *entry;
|
|
+ struct hlist_node *n;
|
|
+
|
|
+ cache = &sdata->u.mesh.tx_cache;
|
|
+ spin_lock_bh(&cache->walk_lock);
|
|
+ hlist_for_each_entry_safe(entry, n, &cache->walk_head, walk_list)
|
|
+ if (entry->mpath == mpath)
|
|
+ mesh_fast_tx_entry_free(cache, entry);
|
|
+ spin_unlock_bh(&cache->walk_lock);
|
|
+}
|
|
+
|
|
+void mesh_fast_tx_flush_sta(struct ieee80211_sub_if_data *sdata,
|
|
+ struct sta_info *sta)
|
|
+{
|
|
+ struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache;
|
|
+ struct ieee80211_mesh_fast_tx *entry;
|
|
+ struct hlist_node *n;
|
|
+
|
|
+ cache = &sdata->u.mesh.tx_cache;
|
|
+ spin_lock_bh(&cache->walk_lock);
|
|
+ hlist_for_each_entry_safe(entry, n, &cache->walk_head, walk_list)
|
|
+ if (rcu_access_pointer(entry->mpath->next_hop) == sta)
|
|
+ mesh_fast_tx_entry_free(cache, entry);
|
|
+ spin_unlock_bh(&cache->walk_lock);
|
|
+}
|
|
+
|
|
+void mesh_fast_tx_flush_addr(struct ieee80211_sub_if_data *sdata,
|
|
+ const u8 *addr)
|
|
+{
|
|
+ struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache;
|
|
+ struct ieee80211_mesh_fast_tx *entry;
|
|
+
|
|
+ cache = &sdata->u.mesh.tx_cache;
|
|
+ spin_lock_bh(&cache->walk_lock);
|
|
+ entry = rhashtable_lookup(&cache->rht, addr, fast_tx_rht_params);
|
|
+ if (entry)
|
|
+ mesh_fast_tx_entry_free(cache, entry);
|
|
+ spin_unlock_bh(&cache->walk_lock);
|
|
+}
|
|
+
|
|
/**
|
|
* mesh_path_add - allocate and add a new path to the mesh path table
|
|
* @dst: destination address of the path (ETH_ALEN length)
|
|
@@ -464,6 +737,8 @@ int mpp_path_add(struct ieee80211_sub_if
|
|
|
|
if (ret)
|
|
kfree(new_mpath);
|
|
+ else
|
|
+ mesh_fast_tx_flush_addr(sdata, dst);
|
|
|
|
sdata->u.mesh.mpp_paths_generation++;
|
|
return ret;
|
|
@@ -523,6 +798,10 @@ static void __mesh_path_del(struct mesh_
|
|
{
|
|
hlist_del_rcu(&mpath->walk_list);
|
|
rhashtable_remove_fast(&tbl->rhead, &mpath->rhash, mesh_rht_params);
|
|
+ if (tbl == &mpath->sdata->u.mesh.mpp_paths)
|
|
+ mesh_fast_tx_flush_addr(mpath->sdata, mpath->dst);
|
|
+ else
|
|
+ mesh_fast_tx_flush_mpath(mpath);
|
|
mesh_path_free_rcu(tbl, mpath);
|
|
}
|
|
|
|
@@ -747,6 +1026,7 @@ void mesh_path_fix_nexthop(struct mesh_p
|
|
mpath->exp_time = 0;
|
|
mpath->flags = MESH_PATH_FIXED | MESH_PATH_SN_VALID;
|
|
mesh_path_activate(mpath);
|
|
+ mesh_fast_tx_flush_mpath(mpath);
|
|
spin_unlock_bh(&mpath->state_lock);
|
|
ewma_mesh_fail_avg_init(&next_hop->mesh->fail_avg);
|
|
/* init it at a low value - 0 start is tricky */
|
|
@@ -758,6 +1038,7 @@ void mesh_pathtbl_init(struct ieee80211_
|
|
{
|
|
mesh_table_init(&sdata->u.mesh.mesh_paths);
|
|
mesh_table_init(&sdata->u.mesh.mpp_paths);
|
|
+ mesh_fast_tx_init(sdata);
|
|
}
|
|
|
|
static
|
|
@@ -785,6 +1066,7 @@ void mesh_path_expire(struct ieee80211_s
|
|
|
|
void mesh_pathtbl_unregister(struct ieee80211_sub_if_data *sdata)
|
|
{
|
|
+ mesh_fast_tx_deinit(sdata);
|
|
mesh_table_free(&sdata->u.mesh.mesh_paths);
|
|
mesh_table_free(&sdata->u.mesh.mpp_paths);
|
|
}
|
|
--- a/net/mac80211/rx.c
|
|
+++ b/net/mac80211/rx.c
|
|
@@ -2791,6 +2791,7 @@ ieee80211_rx_mesh_data(struct ieee80211_
|
|
if (mesh_hdr->flags & MESH_FLAGS_AE) {
|
|
struct mesh_path *mppath;
|
|
char *proxied_addr;
|
|
+ bool update = false;
|
|
|
|
if (multicast)
|
|
proxied_addr = mesh_hdr->eaddr1;
|
|
@@ -2806,11 +2807,18 @@ ieee80211_rx_mesh_data(struct ieee80211_
|
|
mpp_path_add(sdata, proxied_addr, eth->h_source);
|
|
} else {
|
|
spin_lock_bh(&mppath->state_lock);
|
|
- if (!ether_addr_equal(mppath->mpp, eth->h_source))
|
|
+ if (!ether_addr_equal(mppath->mpp, eth->h_source)) {
|
|
memcpy(mppath->mpp, eth->h_source, ETH_ALEN);
|
|
+ update = true;
|
|
+ }
|
|
mppath->exp_time = jiffies;
|
|
spin_unlock_bh(&mppath->state_lock);
|
|
}
|
|
+
|
|
+ /* flush fast xmit cache if the address path changed */
|
|
+ if (update)
|
|
+ mesh_fast_tx_flush_addr(sdata, proxied_addr);
|
|
+
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
--- a/net/mac80211/tx.c
|
|
+++ b/net/mac80211/tx.c
|
|
@@ -3021,6 +3021,9 @@ void ieee80211_check_fast_xmit(struct st
|
|
if (!ieee80211_hw_check(&local->hw, SUPPORT_FAST_XMIT))
|
|
return;
|
|
|
|
+ if (ieee80211_vif_is_mesh(&sdata->vif))
|
|
+ mesh_fast_tx_flush_sta(sdata, sta);
|
|
+
|
|
/* Locking here protects both the pointer itself, and against concurrent
|
|
* invocations winning data access races to, e.g., the key pointer that
|
|
* is used.
|
|
@@ -3402,6 +3405,9 @@ static bool ieee80211_amsdu_aggregate(st
|
|
if (sdata->vif.offload_flags & IEEE80211_OFFLOAD_ENCAP_ENABLED)
|
|
return false;
|
|
|
|
+ if (ieee80211_vif_is_mesh(&sdata->vif))
|
|
+ return false;
|
|
+
|
|
if (skb_is_gso(skb))
|
|
return false;
|
|
|
|
@@ -3634,10 +3640,11 @@ free:
|
|
return NULL;
|
|
}
|
|
|
|
-static void __ieee80211_xmit_fast(struct ieee80211_sub_if_data *sdata,
|
|
- struct sta_info *sta,
|
|
- struct ieee80211_fast_tx *fast_tx,
|
|
- struct sk_buff *skb, u8 tid, bool ampdu)
|
|
+void __ieee80211_xmit_fast(struct ieee80211_sub_if_data *sdata,
|
|
+ struct sta_info *sta,
|
|
+ struct ieee80211_fast_tx *fast_tx,
|
|
+ struct sk_buff *skb, bool ampdu,
|
|
+ const u8 *da, const u8 *sa)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_hdr *hdr = (void *)fast_tx->hdr;
|
|
@@ -3646,7 +3653,6 @@ static void __ieee80211_xmit_fast(struct
|
|
ieee80211_tx_result r;
|
|
int hw_headroom = sdata->local->hw.extra_tx_headroom;
|
|
int extra_head = fast_tx->hdr_len - (ETH_HLEN - 2);
|
|
- struct ethhdr eth;
|
|
|
|
skb = skb_share_check(skb, GFP_ATOMIC);
|
|
if (unlikely(!skb))
|
|
@@ -3666,11 +3672,10 @@ static void __ieee80211_xmit_fast(struct
|
|
ENCRYPT_NO)))
|
|
goto free;
|
|
|
|
- memcpy(ð, skb->data, ETH_HLEN - 2);
|
|
hdr = skb_push(skb, extra_head);
|
|
memcpy(skb->data, fast_tx->hdr, fast_tx->hdr_len);
|
|
- memcpy(skb->data + fast_tx->da_offs, eth.h_dest, ETH_ALEN);
|
|
- memcpy(skb->data + fast_tx->sa_offs, eth.h_source, ETH_ALEN);
|
|
+ memcpy(skb->data + fast_tx->da_offs, da, ETH_ALEN);
|
|
+ memcpy(skb->data + fast_tx->sa_offs, sa, ETH_ALEN);
|
|
|
|
info = IEEE80211_SKB_CB(skb);
|
|
memset(info, 0, sizeof(*info));
|
|
@@ -3689,7 +3694,8 @@ static void __ieee80211_xmit_fast(struct
|
|
#endif
|
|
|
|
if (hdr->frame_control & cpu_to_le16(IEEE80211_STYPE_QOS_DATA)) {
|
|
- tid = skb->priority & IEEE80211_QOS_CTL_TAG1D_MASK;
|
|
+ u8 tid = skb->priority & IEEE80211_QOS_CTL_TAG1D_MASK;
|
|
+
|
|
*ieee80211_get_qos_ctl(hdr) = tid;
|
|
}
|
|
|
|
@@ -3732,6 +3738,7 @@ static bool ieee80211_xmit_fast(struct i
|
|
struct ieee80211_hdr *hdr = (void *)fast_tx->hdr;
|
|
struct tid_ampdu_tx *tid_tx = NULL;
|
|
struct sk_buff *next;
|
|
+ struct ethhdr eth;
|
|
u8 tid = IEEE80211_NUM_TIDS;
|
|
|
|
/* control port protocol needs a lot of special handling */
|
|
@@ -3757,6 +3764,8 @@ static bool ieee80211_xmit_fast(struct i
|
|
}
|
|
}
|
|
|
|
+ memcpy(ð, skb->data, ETH_HLEN - 2);
|
|
+
|
|
/* after this point (skb is modified) we cannot return false */
|
|
skb = ieee80211_tx_skb_fixup(skb, ieee80211_sdata_netdev_features(sdata));
|
|
if (!skb)
|
|
@@ -3764,7 +3773,8 @@ static bool ieee80211_xmit_fast(struct i
|
|
|
|
skb_list_walk_safe(skb, skb, next) {
|
|
skb_mark_not_on_list(skb);
|
|
- __ieee80211_xmit_fast(sdata, sta, fast_tx, skb, tid, tid_tx);
|
|
+ __ieee80211_xmit_fast(sdata, sta, fast_tx, skb, tid_tx,
|
|
+ eth.h_dest, eth.h_source);
|
|
}
|
|
|
|
return true;
|
|
@@ -4244,8 +4254,15 @@ void __ieee80211_subif_start_xmit(struct
|
|
return;
|
|
}
|
|
|
|
+ sk_pacing_shift_update(skb->sk, sdata->local->hw.tx_sk_pacing_shift);
|
|
+
|
|
rcu_read_lock();
|
|
|
|
+ if (ieee80211_vif_is_mesh(&sdata->vif) &&
|
|
+ ieee80211_hw_check(&local->hw, SUPPORT_FAST_XMIT) &&
|
|
+ ieee80211_mesh_xmit_fast(sdata, skb, ctrl_flags))
|
|
+ goto out;
|
|
+
|
|
if (ieee80211_lookup_ra_sta(sdata, skb, &sta))
|
|
goto out_free;
|
|
|
|
@@ -4255,8 +4272,6 @@ void __ieee80211_subif_start_xmit(struct
|
|
skb_set_queue_mapping(skb, ieee80211_select_queue(sdata, sta, skb));
|
|
ieee80211_aggr_check(sdata, sta, skb);
|
|
|
|
- sk_pacing_shift_update(skb->sk, sdata->local->hw.tx_sk_pacing_shift);
|
|
-
|
|
if (sta) {
|
|
struct ieee80211_fast_tx *fast_tx;
|
|
|