at v5.17-rc2 307 lines 8.1 kB view raw
1// SPDX-License-Identifier: GPL-2.0-only 2/* 3 * Queue between the tx operation and the bh workqueue. 4 * 5 * Copyright (c) 2017-2020, Silicon Laboratories, Inc. 6 * Copyright (c) 2010, ST-Ericsson 7 */ 8#include <linux/sched.h> 9#include <net/mac80211.h> 10 11#include "queue.h" 12#include "wfx.h" 13#include "sta.h" 14#include "data_tx.h" 15#include "traces.h" 16 17void wfx_tx_lock(struct wfx_dev *wdev) 18{ 19 atomic_inc(&wdev->tx_lock); 20} 21 22void wfx_tx_unlock(struct wfx_dev *wdev) 23{ 24 int tx_lock = atomic_dec_return(&wdev->tx_lock); 25 26 WARN(tx_lock < 0, "inconsistent tx_lock value"); 27 if (!tx_lock) 28 wfx_bh_request_tx(wdev); 29} 30 31void wfx_tx_flush(struct wfx_dev *wdev) 32{ 33 int ret; 34 35 /* Do not wait for any reply if chip is frozen */ 36 if (wdev->chip_frozen) 37 return; 38 39 wfx_tx_lock(wdev); 40 mutex_lock(&wdev->hif_cmd.lock); 41 ret = wait_event_timeout(wdev->hif.tx_buffers_empty, 42 !wdev->hif.tx_buffers_used, 43 msecs_to_jiffies(3000)); 44 if (!ret) { 45 dev_warn(wdev->dev, "cannot flush tx buffers (%d still busy)\n", 46 wdev->hif.tx_buffers_used); 47 wfx_pending_dump_old_frames(wdev, 3000); 48 /* FIXME: drop pending frames here */ 49 wdev->chip_frozen = true; 50 } 51 mutex_unlock(&wdev->hif_cmd.lock); 52 wfx_tx_unlock(wdev); 53} 54 55void wfx_tx_lock_flush(struct wfx_dev *wdev) 56{ 57 wfx_tx_lock(wdev); 58 wfx_tx_flush(wdev); 59} 60 61void wfx_tx_queues_init(struct wfx_vif *wvif) 62{ 63 /* The device is in charge to respect the details of the QoS parameters. 64 * The driver just ensure that it roughtly respect the priorities to 65 * avoid any shortage. 66 */ 67 const int priorities[IEEE80211_NUM_ACS] = { 1, 2, 64, 128 }; 68 int i; 69 70 for (i = 0; i < IEEE80211_NUM_ACS; ++i) { 71 skb_queue_head_init(&wvif->tx_queue[i].normal); 72 skb_queue_head_init(&wvif->tx_queue[i].cab); 73 wvif->tx_queue[i].priority = priorities[i]; 74 } 75} 76 77bool wfx_tx_queue_empty(struct wfx_vif *wvif, struct wfx_queue *queue) 78{ 79 return skb_queue_empty_lockless(&queue->normal) && 80 skb_queue_empty_lockless(&queue->cab); 81} 82 83void wfx_tx_queues_check_empty(struct wfx_vif *wvif) 84{ 85 int i; 86 87 for (i = 0; i < IEEE80211_NUM_ACS; ++i) { 88 WARN_ON(atomic_read(&wvif->tx_queue[i].pending_frames)); 89 WARN_ON(!wfx_tx_queue_empty(wvif, &wvif->tx_queue[i])); 90 } 91} 92 93static void __wfx_tx_queue_drop(struct wfx_vif *wvif, 94 struct sk_buff_head *skb_queue, 95 struct sk_buff_head *dropped) 96{ 97 struct sk_buff *skb, *tmp; 98 99 spin_lock_bh(&skb_queue->lock); 100 skb_queue_walk_safe(skb_queue, skb, tmp) { 101 __skb_unlink(skb, skb_queue); 102 skb_queue_head(dropped, skb); 103 } 104 spin_unlock_bh(&skb_queue->lock); 105} 106 107void wfx_tx_queue_drop(struct wfx_vif *wvif, struct wfx_queue *queue, 108 struct sk_buff_head *dropped) 109{ 110 __wfx_tx_queue_drop(wvif, &queue->cab, dropped); 111 __wfx_tx_queue_drop(wvif, &queue->normal, dropped); 112 wake_up(&wvif->wdev->tx_dequeue); 113} 114 115void wfx_tx_queues_put(struct wfx_vif *wvif, struct sk_buff *skb) 116{ 117 struct wfx_queue *queue = &wvif->tx_queue[skb_get_queue_mapping(skb)]; 118 struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb); 119 120 if (tx_info->flags & IEEE80211_TX_CTL_SEND_AFTER_DTIM) 121 skb_queue_tail(&queue->cab, skb); 122 else 123 skb_queue_tail(&queue->normal, skb); 124} 125 126void wfx_pending_drop(struct wfx_dev *wdev, struct sk_buff_head *dropped) 127{ 128 struct wfx_queue *queue; 129 struct wfx_vif *wvif; 130 struct hif_msg *hif; 131 struct sk_buff *skb; 132 133 WARN(!wdev->chip_frozen, "%s should only be used to recover a frozen device", 134 __func__); 135 while ((skb = skb_dequeue(&wdev->tx_pending)) != NULL) { 136 hif = (struct hif_msg *)skb->data; 137 wvif = wdev_to_wvif(wdev, hif->interface); 138 if (wvif) { 139 queue = &wvif->tx_queue[skb_get_queue_mapping(skb)]; 140 WARN_ON(skb_get_queue_mapping(skb) > 3); 141 WARN_ON(!atomic_read(&queue->pending_frames)); 142 atomic_dec(&queue->pending_frames); 143 } 144 skb_queue_head(dropped, skb); 145 } 146} 147 148struct sk_buff *wfx_pending_get(struct wfx_dev *wdev, u32 packet_id) 149{ 150 struct wfx_queue *queue; 151 struct hif_req_tx *req; 152 struct wfx_vif *wvif; 153 struct hif_msg *hif; 154 struct sk_buff *skb; 155 156 spin_lock_bh(&wdev->tx_pending.lock); 157 skb_queue_walk(&wdev->tx_pending, skb) { 158 hif = (struct hif_msg *)skb->data; 159 req = (struct hif_req_tx *)hif->body; 160 if (req->packet_id != packet_id) 161 continue; 162 spin_unlock_bh(&wdev->tx_pending.lock); 163 wvif = wdev_to_wvif(wdev, hif->interface); 164 if (wvif) { 165 queue = &wvif->tx_queue[skb_get_queue_mapping(skb)]; 166 WARN_ON(skb_get_queue_mapping(skb) > 3); 167 WARN_ON(!atomic_read(&queue->pending_frames)); 168 atomic_dec(&queue->pending_frames); 169 } 170 skb_unlink(skb, &wdev->tx_pending); 171 return skb; 172 } 173 spin_unlock_bh(&wdev->tx_pending.lock); 174 WARN(1, "cannot find packet in pending queue"); 175 return NULL; 176} 177 178void wfx_pending_dump_old_frames(struct wfx_dev *wdev, unsigned int limit_ms) 179{ 180 ktime_t now = ktime_get(); 181 struct wfx_tx_priv *tx_priv; 182 struct hif_req_tx *req; 183 struct sk_buff *skb; 184 bool first = true; 185 186 spin_lock_bh(&wdev->tx_pending.lock); 187 skb_queue_walk(&wdev->tx_pending, skb) { 188 tx_priv = wfx_skb_tx_priv(skb); 189 req = wfx_skb_txreq(skb); 190 if (ktime_after(now, ktime_add_ms(tx_priv->xmit_timestamp, 191 limit_ms))) { 192 if (first) { 193 dev_info(wdev->dev, "frames stuck in firmware since %dms or more:\n", 194 limit_ms); 195 first = false; 196 } 197 dev_info(wdev->dev, " id %08x sent %lldms ago\n", 198 req->packet_id, 199 ktime_ms_delta(now, tx_priv->xmit_timestamp)); 200 } 201 } 202 spin_unlock_bh(&wdev->tx_pending.lock); 203} 204 205unsigned int wfx_pending_get_pkt_us_delay(struct wfx_dev *wdev, 206 struct sk_buff *skb) 207{ 208 ktime_t now = ktime_get(); 209 struct wfx_tx_priv *tx_priv = wfx_skb_tx_priv(skb); 210 211 return ktime_us_delta(now, tx_priv->xmit_timestamp); 212} 213 214bool wfx_tx_queues_has_cab(struct wfx_vif *wvif) 215{ 216 int i; 217 218 if (wvif->vif->type != NL80211_IFTYPE_AP) 219 return false; 220 for (i = 0; i < IEEE80211_NUM_ACS; ++i) 221 /* Note: since only AP can have mcast frames in queue and only 222 * one vif can be AP, all queued frames has same interface id 223 */ 224 if (!skb_queue_empty_lockless(&wvif->tx_queue[i].cab)) 225 return true; 226 return false; 227} 228 229static int wfx_tx_queue_get_weight(struct wfx_queue *queue) 230{ 231 return atomic_read(&queue->pending_frames) * queue->priority; 232} 233 234static struct sk_buff *wfx_tx_queues_get_skb(struct wfx_dev *wdev) 235{ 236 struct wfx_queue *queues[IEEE80211_NUM_ACS * ARRAY_SIZE(wdev->vif)]; 237 int i, j, num_queues = 0; 238 struct wfx_vif *wvif; 239 struct hif_msg *hif; 240 struct sk_buff *skb; 241 242 /* sort the queues */ 243 wvif = NULL; 244 while ((wvif = wvif_iterate(wdev, wvif)) != NULL) { 245 for (i = 0; i < IEEE80211_NUM_ACS; i++) { 246 WARN_ON(num_queues >= ARRAY_SIZE(queues)); 247 queues[num_queues] = &wvif->tx_queue[i]; 248 for (j = num_queues; j > 0; j--) 249 if (wfx_tx_queue_get_weight(queues[j]) < 250 wfx_tx_queue_get_weight(queues[j - 1])) 251 swap(queues[j - 1], queues[j]); 252 num_queues++; 253 } 254 } 255 256 wvif = NULL; 257 while ((wvif = wvif_iterate(wdev, wvif)) != NULL) { 258 if (!wvif->after_dtim_tx_allowed) 259 continue; 260 for (i = 0; i < num_queues; i++) { 261 skb = skb_dequeue(&queues[i]->cab); 262 if (!skb) 263 continue; 264 /* Note: since only AP can have mcast frames in queue 265 * and only one vif can be AP, all queued frames has 266 * same interface id 267 */ 268 hif = (struct hif_msg *)skb->data; 269 WARN_ON(hif->interface != wvif->id); 270 WARN_ON(queues[i] != 271 &wvif->tx_queue[skb_get_queue_mapping(skb)]); 272 atomic_inc(&queues[i]->pending_frames); 273 trace_queues_stats(wdev, queues[i]); 274 return skb; 275 } 276 /* No more multicast to sent */ 277 wvif->after_dtim_tx_allowed = false; 278 schedule_work(&wvif->update_tim_work); 279 } 280 281 for (i = 0; i < num_queues; i++) { 282 skb = skb_dequeue(&queues[i]->normal); 283 if (skb) { 284 atomic_inc(&queues[i]->pending_frames); 285 trace_queues_stats(wdev, queues[i]); 286 return skb; 287 } 288 } 289 return NULL; 290} 291 292struct hif_msg *wfx_tx_queues_get(struct wfx_dev *wdev) 293{ 294 struct wfx_tx_priv *tx_priv; 295 struct sk_buff *skb; 296 297 if (atomic_read(&wdev->tx_lock)) 298 return NULL; 299 skb = wfx_tx_queues_get_skb(wdev); 300 if (!skb) 301 return NULL; 302 skb_queue_tail(&wdev->tx_pending, skb); 303 wake_up(&wdev->tx_dequeue); 304 tx_priv = wfx_skb_tx_priv(skb); 305 tx_priv->xmit_timestamp = ktime_get(); 306 return (struct hif_msg *)skb->data; 307}