1 /* SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause */
2 /*
3 * Copyright (C) 2012-2014, 2018-2022 Intel Corporation
4 * Copyright (C) 2013-2015 Intel Mobile Communications GmbH
5 * Copyright (C) 2016-2017 Intel Deutschland GmbH
6 */
7 #ifndef __IWL_MVM_H__
8 #define __IWL_MVM_H__
9
10 #include <linux/list.h>
11 #include <linux/spinlock.h>
12 #include <linux/leds.h>
13 #include <linux/in6.h>
14
15 #ifdef CONFIG_THERMAL
16 #include <linux/thermal.h>
17 #endif
18
19 #include <linux/ktime.h>
20
21 #include "iwl-op-mode.h"
22 #include "iwl-trans.h"
23 #include "fw/notif-wait.h"
24 #include "iwl-eeprom-parse.h"
25 #include "fw/file.h"
26 #include "iwl-config.h"
27 #include "sta.h"
28 #include "fw-api.h"
29 #include "constants.h"
30 #include "fw/runtime.h"
31 #include "fw/dbg.h"
32 #include "fw/acpi.h"
33 #include "mei/iwl-mei.h"
34 #include "iwl-nvm-parse.h"
35
36 #include <linux/average.h>
37
38 #define IWL_MVM_MAX_ADDRESSES 5
39 /* RSSI offset for WkP */
40 #define IWL_RSSI_OFFSET 50
41 #define IWL_MVM_MISSED_BEACONS_THRESHOLD 8
42 #define IWL_MVM_MISSED_BEACONS_THRESHOLD_LONG 16
43
44 /* A TimeUnit is 1024 microsecond */
45 #define MSEC_TO_TU(_msec) (_msec*1000/1024)
46
47 /* For GO, this value represents the number of TUs before CSA "beacon
48 * 0" TBTT when the CSA time-event needs to be scheduled to start. It
49 * must be big enough to ensure that we switch in time.
50 */
51 #define IWL_MVM_CHANNEL_SWITCH_TIME_GO 40
52
53 /* For client, this value represents the number of TUs before CSA
54 * "beacon 1" TBTT, instead. This is because we don't know when the
55 * GO/AP will be in the new channel, so we switch early enough.
56 */
57 #define IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT 10
58
59 /*
60 * This value (in TUs) is used to fine tune the CSA NoA end time which should
61 * be just before "beacon 0" TBTT.
62 */
63 #define IWL_MVM_CHANNEL_SWITCH_MARGIN 4
64
65 /*
66 * Number of beacons to transmit on a new channel until we unblock tx to
67 * the stations, even if we didn't identify them on a new channel
68 */
69 #define IWL_MVM_CS_UNBLOCK_TX_TIMEOUT 3
70
71 /* offchannel queue towards mac80211 */
72 #define IWL_MVM_OFFCHANNEL_QUEUE 0
73
74 extern const struct ieee80211_ops iwl_mvm_hw_ops;
75
76 /**
77 * struct iwl_mvm_mod_params - module parameters for iwlmvm
78 * @init_dbg: if true, then the NIC won't be stopped if the INIT fw asserted.
79 * We will register to mac80211 to have testmode working. The NIC must not
80 * be up'ed after the INIT fw asserted. This is useful to be able to use
81 * proprietary tools over testmode to debug the INIT fw.
82 * @power_scheme: one of enum iwl_power_scheme
83 */
84 struct iwl_mvm_mod_params {
85 bool init_dbg;
86 int power_scheme;
87 };
88 extern struct iwl_mvm_mod_params iwlmvm_mod_params;
89
90 struct iwl_mvm_phy_ctxt {
91 u16 id;
92 u16 color;
93 u32 ref;
94
95 enum nl80211_chan_width width;
96
97 struct ieee80211_channel *channel;
98
99 /* track for RLC config command */
100 u32 center_freq1;
101 };
102
103 struct iwl_mvm_time_event_data {
104 struct ieee80211_vif *vif;
105 struct list_head list;
106 unsigned long end_jiffies;
107 u32 duration;
108 bool running;
109 u32 uid;
110
111 /*
112 * The access to the 'id' field must be done when the
113 * mvm->time_event_lock is held, as it value is used to indicate
114 * if the te is in the time event list or not (when id == TE_MAX)
115 */
116 u32 id;
117 };
118
119 /* Power management */
120
121 /**
122 * enum iwl_power_scheme
123 * @IWL_POWER_LEVEL_CAM - Continuously Active Mode
124 * @IWL_POWER_LEVEL_BPS - Balanced Power Save (default)
125 * @IWL_POWER_LEVEL_LP - Low Power
126 */
127 enum iwl_power_scheme {
128 IWL_POWER_SCHEME_CAM = 1,
129 IWL_POWER_SCHEME_BPS,
130 IWL_POWER_SCHEME_LP
131 };
132
133 #define IWL_CONN_MAX_LISTEN_INTERVAL 10
134 #define IWL_UAPSD_MAX_SP IEEE80211_WMM_IE_STA_QOSINFO_SP_ALL
135
136 #ifdef CONFIG_IWLWIFI_DEBUGFS
137 enum iwl_dbgfs_pm_mask {
138 MVM_DEBUGFS_PM_KEEP_ALIVE = BIT(0),
139 MVM_DEBUGFS_PM_SKIP_OVER_DTIM = BIT(1),
140 MVM_DEBUGFS_PM_SKIP_DTIM_PERIODS = BIT(2),
141 MVM_DEBUGFS_PM_RX_DATA_TIMEOUT = BIT(3),
142 MVM_DEBUGFS_PM_TX_DATA_TIMEOUT = BIT(4),
143 MVM_DEBUGFS_PM_LPRX_ENA = BIT(6),
144 MVM_DEBUGFS_PM_LPRX_RSSI_THRESHOLD = BIT(7),
145 MVM_DEBUGFS_PM_SNOOZE_ENABLE = BIT(8),
146 MVM_DEBUGFS_PM_UAPSD_MISBEHAVING = BIT(9),
147 MVM_DEBUGFS_PM_USE_PS_POLL = BIT(10),
148 };
149
150 struct iwl_dbgfs_pm {
151 u16 keep_alive_seconds;
152 u32 rx_data_timeout;
153 u32 tx_data_timeout;
154 bool skip_over_dtim;
155 u8 skip_dtim_periods;
156 bool lprx_ena;
157 u32 lprx_rssi_threshold;
158 bool snooze_ena;
159 bool uapsd_misbehaving;
160 bool use_ps_poll;
161 int mask;
162 };
163
164 /* beacon filtering */
165
166 enum iwl_dbgfs_bf_mask {
167 MVM_DEBUGFS_BF_ENERGY_DELTA = BIT(0),
168 MVM_DEBUGFS_BF_ROAMING_ENERGY_DELTA = BIT(1),
169 MVM_DEBUGFS_BF_ROAMING_STATE = BIT(2),
170 MVM_DEBUGFS_BF_TEMP_THRESHOLD = BIT(3),
171 MVM_DEBUGFS_BF_TEMP_FAST_FILTER = BIT(4),
172 MVM_DEBUGFS_BF_TEMP_SLOW_FILTER = BIT(5),
173 MVM_DEBUGFS_BF_ENABLE_BEACON_FILTER = BIT(6),
174 MVM_DEBUGFS_BF_DEBUG_FLAG = BIT(7),
175 MVM_DEBUGFS_BF_ESCAPE_TIMER = BIT(8),
176 MVM_DEBUGFS_BA_ESCAPE_TIMER = BIT(9),
177 MVM_DEBUGFS_BA_ENABLE_BEACON_ABORT = BIT(10),
178 };
179
180 struct iwl_dbgfs_bf {
181 u32 bf_energy_delta;
182 u32 bf_roaming_energy_delta;
183 u32 bf_roaming_state;
184 u32 bf_temp_threshold;
185 u32 bf_temp_fast_filter;
186 u32 bf_temp_slow_filter;
187 u32 bf_enable_beacon_filter;
188 u32 bf_debug_flag;
189 u32 bf_escape_timer;
190 u32 ba_escape_timer;
191 u32 ba_enable_beacon_abort;
192 int mask;
193 };
194 #endif
195
196 enum iwl_mvm_smps_type_request {
197 IWL_MVM_SMPS_REQ_BT_COEX,
198 IWL_MVM_SMPS_REQ_TT,
199 IWL_MVM_SMPS_REQ_PROT,
200 IWL_MVM_SMPS_REQ_FW,
201 NUM_IWL_MVM_SMPS_REQ,
202 };
203
204 enum iwl_bt_force_ant_mode {
205 BT_FORCE_ANT_DIS = 0,
206 BT_FORCE_ANT_AUTO,
207 BT_FORCE_ANT_BT,
208 BT_FORCE_ANT_WIFI,
209
210 BT_FORCE_ANT_MAX,
211 };
212
213 /**
214 * struct iwl_mvm_low_latency_force - low latency force mode set by debugfs
215 * @LOW_LATENCY_FORCE_UNSET: unset force mode
216 * @LOW_LATENCY_FORCE_ON: for low latency on
217 * @LOW_LATENCY_FORCE_OFF: for low latency off
218 * @NUM_LOW_LATENCY_FORCE: max num of modes
219 */
220 enum iwl_mvm_low_latency_force {
221 LOW_LATENCY_FORCE_UNSET,
222 LOW_LATENCY_FORCE_ON,
223 LOW_LATENCY_FORCE_OFF,
224 NUM_LOW_LATENCY_FORCE
225 };
226
227 /**
228 * struct iwl_mvm_low_latency_cause - low latency set causes
229 * @LOW_LATENCY_TRAFFIC: indicates low latency traffic was detected
230 * @LOW_LATENCY_DEBUGFS: low latency mode set from debugfs
231 * @LOW_LATENCY_VCMD: low latency mode set from vendor command
232 * @LOW_LATENCY_VIF_TYPE: low latency mode set because of vif type (ap)
233 * @LOW_LATENCY_DEBUGFS_FORCE_ENABLE: indicate that force mode is enabled
234 * the actual set/unset is done with LOW_LATENCY_DEBUGFS_FORCE
235 * @LOW_LATENCY_DEBUGFS_FORCE: low latency force mode from debugfs
236 * set this with LOW_LATENCY_DEBUGFS_FORCE_ENABLE flag
237 * in low_latency.
238 */
239 enum iwl_mvm_low_latency_cause {
240 LOW_LATENCY_TRAFFIC = BIT(0),
241 LOW_LATENCY_DEBUGFS = BIT(1),
242 LOW_LATENCY_VCMD = BIT(2),
243 LOW_LATENCY_VIF_TYPE = BIT(3),
244 LOW_LATENCY_DEBUGFS_FORCE_ENABLE = BIT(4),
245 LOW_LATENCY_DEBUGFS_FORCE = BIT(5),
246 };
247
248 /**
249 * struct iwl_mvm_vif_bf_data - beacon filtering related data
250 * @bf_enabled: indicates if beacon filtering is enabled
251 * @ba_enabled: indicated if beacon abort is enabled
252 * @ave_beacon_signal: average beacon signal
253 * @last_cqm_event: rssi of the last cqm event
254 * @bt_coex_min_thold: minimum threshold for BT coex
255 * @bt_coex_max_thold: maximum threshold for BT coex
256 * @last_bt_coex_event: rssi of the last BT coex event
257 */
258 struct iwl_mvm_vif_bf_data {
259 bool bf_enabled;
260 bool ba_enabled;
261 int ave_beacon_signal;
262 int last_cqm_event;
263 int bt_coex_min_thold;
264 int bt_coex_max_thold;
265 int last_bt_coex_event;
266 };
267
268 /**
269 * struct iwl_probe_resp_data - data for NoA/CSA updates
270 * @rcu_head: used for freeing the data on update
271 * @notif: notification data
272 * @noa_len: length of NoA attribute, calculated from the notification
273 */
274 struct iwl_probe_resp_data {
275 struct rcu_head rcu_head;
276 struct iwl_probe_resp_data_notif notif;
277 int noa_len;
278 };
279
280 /**
281 * struct iwl_mvm_vif - data per Virtual Interface, it is a MAC context
282 * @id: between 0 and 3
283 * @color: to solve races upon MAC addition and removal
284 * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA
285 * @bssid: BSSID for this (client) interface
286 * @associated: indicates that we're currently associated, used only for
287 * managing the firmware state in iwl_mvm_bss_info_changed_station()
288 * @ap_assoc_sta_count: count of stations associated to us - valid only
289 * if VIF type is AP
290 * @uploaded: indicates the MAC context has been added to the device
291 * @ap_ibss_active: indicates that AP/IBSS is configured and that the interface
292 * should get quota etc.
293 * @pm_enabled - Indicate if MAC power management is allowed
294 * @monitor_active: indicates that monitor context is configured, and that the
295 * interface should get quota etc.
296 * @low_latency: bit flags for low latency
297 * see enum &iwl_mvm_low_latency_cause for causes.
298 * @low_latency_actual: boolean, indicates low latency is set,
299 * as a result from low_latency bit flags and takes force into account.
300 * @authorized: indicates the AP station was set to authorized
301 * @ps_disabled: indicates that this interface requires PS to be disabled
302 * @queue_params: QoS params for this MAC
303 * @bcast_sta: station used for broadcast packets. Used by the following
304 * vifs: P2P_DEVICE, GO and AP.
305 * @beacon_skb: the skb used to hold the AP/GO beacon template
306 * @smps_requests: the SMPS requests of different parts of the driver,
307 * combined on update to yield the overall request to mac80211.
308 * @beacon_stats: beacon statistics, containing the # of received beacons,
309 * # of received beacons accumulated over FW restart, and the current
310 * average signal of beacons retrieved from the firmware
311 * @csa_failed: CSA failed to schedule time event, report an error later
312 * @features: hw features active for this vif
313 * @probe_resp_data: data from FW notification to store NOA and CSA related
314 * data to be inserted into probe response.
315 */
316 struct iwl_mvm_vif {
317 struct iwl_mvm *mvm;
318 u16 id;
319 u16 color;
320 u8 ap_sta_id;
321
322 u8 bssid[ETH_ALEN];
323 bool associated;
324 u8 ap_assoc_sta_count;
325
326 u16 cab_queue;
327
328 bool uploaded;
329 bool ap_ibss_active;
330 bool pm_enabled;
331 bool monitor_active;
332 u8 low_latency: 6;
333 u8 low_latency_actual: 1;
334 u8 authorized:1;
335 bool ps_disabled;
336 struct iwl_mvm_vif_bf_data bf_data;
337
338 struct {
339 u32 num_beacons, accu_num_beacons;
340 u8 avg_signal;
341 } beacon_stats;
342
343 u32 ap_beacon_time;
344
345 enum iwl_tsf_id tsf_id;
346
347 /*
348 * QoS data from mac80211, need to store this here
349 * as mac80211 has a separate callback but we need
350 * to have the data for the MAC context
351 */
352 struct ieee80211_tx_queue_params queue_params[IEEE80211_NUM_ACS];
353 struct iwl_mvm_time_event_data time_event_data;
354 struct iwl_mvm_time_event_data hs_time_event_data;
355
356 struct iwl_mvm_int_sta bcast_sta;
357 struct iwl_mvm_int_sta mcast_sta;
358
359 /*
360 * Assigned while mac80211 has the interface in a channel context,
361 * or, for P2P Device, while it exists.
362 */
363 struct iwl_mvm_phy_ctxt *phy_ctxt;
364
365 #ifdef CONFIG_PM
366 /* WoWLAN GTK rekey data */
367 struct {
368 u8 kck[NL80211_KCK_EXT_LEN];
369 u8 kek[NL80211_KEK_EXT_LEN];
370 size_t kek_len;
371 size_t kck_len;
372 u32 akm;
373 __le64 replay_ctr;
374 bool valid;
375 } rekey_data;
376
377 int tx_key_idx;
378
379 bool seqno_valid;
380 u16 seqno;
381 #endif
382
383 #if IS_ENABLED(CONFIG_IPV6)
384 /* IPv6 addresses for WoWLAN */
385 struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX];
386 unsigned long tentative_addrs[BITS_TO_LONGS(IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX)];
387 int num_target_ipv6_addrs;
388 #endif
389
390 #ifdef CONFIG_IWLWIFI_DEBUGFS
391 struct dentry *dbgfs_dir;
392 struct dentry *dbgfs_slink;
393 struct iwl_dbgfs_pm dbgfs_pm;
394 struct iwl_dbgfs_bf dbgfs_bf;
395 struct iwl_mac_power_cmd mac_pwr_cmd;
396 int dbgfs_quota_min;
397 #endif
398
399 enum ieee80211_smps_mode smps_requests[NUM_IWL_MVM_SMPS_REQ];
400
401 /* FW identified misbehaving AP */
402 u8 uapsd_misbehaving_bssid[ETH_ALEN];
403
404 struct delayed_work uapsd_nonagg_detected_wk;
405
406 /* Indicates that CSA countdown may be started */
407 bool csa_countdown;
408 bool csa_failed;
409 u16 csa_target_freq;
410 u16 csa_count;
411 u16 csa_misbehave;
412 struct delayed_work csa_work;
413
414 /* Indicates that we are waiting for a beacon on a new channel */
415 bool csa_bcn_pending;
416
417 /* TCP Checksum Offload */
418 netdev_features_t features;
419
420 struct iwl_probe_resp_data __rcu *probe_resp_data;
421
422 /* we can only have 2 GTK + 2 IGTK active at a time */
423 struct ieee80211_key_conf *ap_early_keys[4];
424
425 /* 26-tone RU OFDMA transmissions should be blocked */
426 bool he_ru_2mhz_block;
427
428 struct {
429 struct ieee80211_key_conf __rcu *keys[2];
430 } bcn_prot;
431 };
432
433 static inline struct iwl_mvm_vif *
iwl_mvm_vif_from_mac80211(struct ieee80211_vif * vif)434 iwl_mvm_vif_from_mac80211(struct ieee80211_vif *vif)
435 {
436 return (void *)vif->drv_priv;
437 }
438
439 extern const u8 tid_to_mac80211_ac[];
440
441 #define IWL_MVM_SCAN_STOPPING_SHIFT 8
442
443 enum iwl_scan_status {
444 IWL_MVM_SCAN_REGULAR = BIT(0),
445 IWL_MVM_SCAN_SCHED = BIT(1),
446 IWL_MVM_SCAN_NETDETECT = BIT(2),
447
448 IWL_MVM_SCAN_STOPPING_REGULAR = BIT(8),
449 IWL_MVM_SCAN_STOPPING_SCHED = BIT(9),
450 IWL_MVM_SCAN_STOPPING_NETDETECT = BIT(10),
451
452 IWL_MVM_SCAN_REGULAR_MASK = IWL_MVM_SCAN_REGULAR |
453 IWL_MVM_SCAN_STOPPING_REGULAR,
454 IWL_MVM_SCAN_SCHED_MASK = IWL_MVM_SCAN_SCHED |
455 IWL_MVM_SCAN_STOPPING_SCHED,
456 IWL_MVM_SCAN_NETDETECT_MASK = IWL_MVM_SCAN_NETDETECT |
457 IWL_MVM_SCAN_STOPPING_NETDETECT,
458
459 IWL_MVM_SCAN_STOPPING_MASK = 0xff << IWL_MVM_SCAN_STOPPING_SHIFT,
460 IWL_MVM_SCAN_MASK = 0xff,
461 };
462
463 enum iwl_mvm_scan_type {
464 IWL_SCAN_TYPE_NOT_SET,
465 IWL_SCAN_TYPE_UNASSOC,
466 IWL_SCAN_TYPE_WILD,
467 IWL_SCAN_TYPE_MILD,
468 IWL_SCAN_TYPE_FRAGMENTED,
469 IWL_SCAN_TYPE_FAST_BALANCE,
470 };
471
472 enum iwl_mvm_sched_scan_pass_all_states {
473 SCHED_SCAN_PASS_ALL_DISABLED,
474 SCHED_SCAN_PASS_ALL_ENABLED,
475 SCHED_SCAN_PASS_ALL_FOUND,
476 };
477
478 /**
479 * struct iwl_mvm_tt_mgnt - Thermal Throttling Management structure
480 * @ct_kill_exit: worker to exit thermal kill
481 * @dynamic_smps: Is thermal throttling enabled dynamic_smps?
482 * @tx_backoff: The current thremal throttling tx backoff in uSec.
483 * @min_backoff: The minimal tx backoff due to power restrictions
484 * @params: Parameters to configure the thermal throttling algorithm.
485 * @throttle: Is thermal throttling is active?
486 */
487 struct iwl_mvm_tt_mgmt {
488 struct delayed_work ct_kill_exit;
489 bool dynamic_smps;
490 u32 tx_backoff;
491 u32 min_backoff;
492 struct iwl_tt_params params;
493 bool throttle;
494 };
495
496 #ifdef CONFIG_THERMAL
497 /**
498 *struct iwl_mvm_thermal_device - thermal zone related data
499 * @temp_trips: temperature thresholds for report
500 * @fw_trips_index: keep indexes to original array - temp_trips
501 * @tzone: thermal zone device data
502 */
503 struct iwl_mvm_thermal_device {
504 s16 temp_trips[IWL_MAX_DTS_TRIPS];
505 u8 fw_trips_index[IWL_MAX_DTS_TRIPS];
506 struct thermal_zone_device *tzone;
507 };
508
509 /*
510 * struct iwl_mvm_cooling_device
511 * @cur_state: current state
512 * @cdev: struct thermal cooling device
513 */
514 struct iwl_mvm_cooling_device {
515 u32 cur_state;
516 struct thermal_cooling_device *cdev;
517 };
518 #endif
519
520 #define IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES 8
521
522 struct iwl_mvm_frame_stats {
523 u32 legacy_frames;
524 u32 ht_frames;
525 u32 vht_frames;
526 u32 bw_20_frames;
527 u32 bw_40_frames;
528 u32 bw_80_frames;
529 u32 bw_160_frames;
530 u32 sgi_frames;
531 u32 ngi_frames;
532 u32 siso_frames;
533 u32 mimo2_frames;
534 u32 agg_frames;
535 u32 ampdu_count;
536 u32 success_frames;
537 u32 fail_frames;
538 u32 last_rates[IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES];
539 int last_frame_idx;
540 };
541
542 #define IWL_MVM_DEBUG_SET_TEMPERATURE_DISABLE 0xff
543 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MIN -100
544 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MAX 200
545
546 enum iwl_mvm_tdls_cs_state {
547 IWL_MVM_TDLS_SW_IDLE = 0,
548 IWL_MVM_TDLS_SW_REQ_SENT,
549 IWL_MVM_TDLS_SW_RESP_RCVD,
550 IWL_MVM_TDLS_SW_REQ_RCVD,
551 IWL_MVM_TDLS_SW_ACTIVE,
552 };
553
554 enum iwl_mvm_traffic_load {
555 IWL_MVM_TRAFFIC_LOW,
556 IWL_MVM_TRAFFIC_MEDIUM,
557 IWL_MVM_TRAFFIC_HIGH,
558 };
559
560 DECLARE_EWMA(rate, 16, 16)
561
562 struct iwl_mvm_tcm_mac {
563 struct {
564 u32 pkts[IEEE80211_NUM_ACS];
565 u32 airtime;
566 } tx;
567 struct {
568 u32 pkts[IEEE80211_NUM_ACS];
569 u32 airtime;
570 u32 last_ampdu_ref;
571 } rx;
572 struct {
573 /* track AP's transfer in client mode */
574 u64 rx_bytes;
575 struct ewma_rate rate;
576 bool detected;
577 } uapsd_nonagg_detect;
578 bool opened_rx_ba_sessions;
579 };
580
581 struct iwl_mvm_tcm {
582 struct delayed_work work;
583 spinlock_t lock; /* used when time elapsed */
584 unsigned long ts; /* timestamp when period ends */
585 unsigned long ll_ts;
586 unsigned long uapsd_nonagg_ts;
587 bool paused;
588 struct iwl_mvm_tcm_mac data[NUM_MAC_INDEX_DRIVER];
589 struct {
590 u32 elapsed; /* milliseconds for this TCM period */
591 u32 airtime[NUM_MAC_INDEX_DRIVER];
592 enum iwl_mvm_traffic_load load[NUM_MAC_INDEX_DRIVER];
593 enum iwl_mvm_traffic_load band_load[NUM_NL80211_BANDS];
594 enum iwl_mvm_traffic_load global_load;
595 bool low_latency[NUM_MAC_INDEX_DRIVER];
596 bool change[NUM_MAC_INDEX_DRIVER];
597 } result;
598 };
599
600 /**
601 * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer
602 * @head_sn: reorder window head sn
603 * @num_stored: number of mpdus stored in the buffer
604 * @buf_size: the reorder buffer size as set by the last addba request
605 * @queue: queue of this reorder buffer
606 * @last_amsdu: track last ASMDU SN for duplication detection
607 * @last_sub_index: track ASMDU sub frame index for duplication detection
608 * @reorder_timer: timer for frames are in the reorder buffer. For AMSDU
609 * it is the time of last received sub-frame
610 * @removed: prevent timer re-arming
611 * @valid: reordering is valid for this queue
612 * @lock: protect reorder buffer internal state
613 * @mvm: mvm pointer, needed for frame timer context
614 * @consec_oldsn_drops: consecutive drops due to old SN
615 * @consec_oldsn_ampdu_gp2: A-MPDU GP2 timestamp to track
616 * when to apply old SN consecutive drop workaround
617 * @consec_oldsn_prev_drop: track whether or not an MPDU
618 * that was single/part of the previous A-MPDU was
619 * dropped due to old SN
620 */
621 struct iwl_mvm_reorder_buffer {
622 u16 head_sn;
623 u16 num_stored;
624 u16 buf_size;
625 int queue;
626 u16 last_amsdu;
627 u8 last_sub_index;
628 struct timer_list reorder_timer;
629 bool removed;
630 bool valid;
631 spinlock_t lock;
632 struct iwl_mvm *mvm;
633 unsigned int consec_oldsn_drops;
634 u32 consec_oldsn_ampdu_gp2;
635 unsigned int consec_oldsn_prev_drop:1;
636 } ____cacheline_aligned_in_smp;
637
638 /**
639 * struct _iwl_mvm_reorder_buf_entry - reorder buffer entry per-queue/per-seqno
640 * @frames: list of skbs stored
641 * @reorder_time: time the packet was stored in the reorder buffer
642 */
643 struct _iwl_mvm_reorder_buf_entry {
644 struct sk_buff_head frames;
645 unsigned long reorder_time;
646 };
647
648 /* make this indirection to get the aligned thing */
649 struct iwl_mvm_reorder_buf_entry {
650 struct _iwl_mvm_reorder_buf_entry e;
651 }
652 #ifndef __CHECKER__
653 /* sparse doesn't like this construct: "bad integer constant expression" */
654 __aligned(roundup_pow_of_two(sizeof(struct _iwl_mvm_reorder_buf_entry)))
655 #endif
656 ;
657
658 /**
659 * struct iwl_mvm_baid_data - BA session data
660 * @sta_id: station id
661 * @tid: tid of the session
662 * @baid baid of the session
663 * @timeout: the timeout set in the addba request
664 * @entries_per_queue: # of buffers per queue, this actually gets
665 * aligned up to avoid cache line sharing between queues
666 * @last_rx: last rx jiffies, updated only if timeout passed from last update
667 * @session_timer: timer to check if BA session expired, runs at 2 * timeout
668 * @mvm: mvm pointer, needed for timer context
669 * @reorder_buf: reorder buffer, allocated per queue
670 * @reorder_buf_data: data
671 */
672 struct iwl_mvm_baid_data {
673 struct rcu_head rcu_head;
674 u8 sta_id;
675 u8 tid;
676 u8 baid;
677 u16 timeout;
678 u16 entries_per_queue;
679 unsigned long last_rx;
680 struct timer_list session_timer;
681 struct iwl_mvm_baid_data __rcu **rcu_ptr;
682 struct iwl_mvm *mvm;
683 struct iwl_mvm_reorder_buffer reorder_buf[IWL_MAX_RX_HW_QUEUES];
684 struct iwl_mvm_reorder_buf_entry entries[];
685 };
686
687 static inline struct iwl_mvm_baid_data *
iwl_mvm_baid_data_from_reorder_buf(struct iwl_mvm_reorder_buffer * buf)688 iwl_mvm_baid_data_from_reorder_buf(struct iwl_mvm_reorder_buffer *buf)
689 {
690 return (void *)((u8 *)buf -
691 offsetof(struct iwl_mvm_baid_data, reorder_buf) -
692 sizeof(*buf) * buf->queue);
693 }
694
695 /*
696 * enum iwl_mvm_queue_status - queue status
697 * @IWL_MVM_QUEUE_FREE: the queue is not allocated nor reserved
698 * Basically, this means that this queue can be used for any purpose
699 * @IWL_MVM_QUEUE_RESERVED: queue is reserved but not yet in use
700 * This is the state of a queue that has been dedicated for some RATID
701 * (agg'd or not), but that hasn't yet gone through the actual enablement
702 * of iwl_mvm_enable_txq(), and therefore no traffic can go through it yet.
703 * Note that in this state there is no requirement to already know what TID
704 * should be used with this queue, it is just marked as a queue that will
705 * be used, and shouldn't be allocated to anyone else.
706 * @IWL_MVM_QUEUE_READY: queue is ready to be used
707 * This is the state of a queue that has been fully configured (including
708 * SCD pointers, etc), has a specific RA/TID assigned to it, and can be
709 * used to send traffic.
710 * @IWL_MVM_QUEUE_SHARED: queue is shared, or in a process of becoming shared
711 * This is a state in which a single queue serves more than one TID, all of
712 * which are not aggregated. Note that the queue is only associated to one
713 * RA.
714 */
715 enum iwl_mvm_queue_status {
716 IWL_MVM_QUEUE_FREE,
717 IWL_MVM_QUEUE_RESERVED,
718 IWL_MVM_QUEUE_READY,
719 IWL_MVM_QUEUE_SHARED,
720 };
721
722 #define IWL_MVM_DQA_QUEUE_TIMEOUT (5 * HZ)
723 #define IWL_MVM_INVALID_QUEUE 0xFFFF
724
725 #define IWL_MVM_NUM_CIPHERS 10
726
727
728 struct iwl_mvm_txq {
729 struct list_head list;
730 u16 txq_id;
731 atomic_t tx_request;
732 #define IWL_MVM_TXQ_STATE_STOP_FULL 0
733 #define IWL_MVM_TXQ_STATE_STOP_REDIRECT 1
734 #define IWL_MVM_TXQ_STATE_READY 2
735 unsigned long state;
736 };
737
738 static inline struct iwl_mvm_txq *
iwl_mvm_txq_from_mac80211(struct ieee80211_txq * txq)739 iwl_mvm_txq_from_mac80211(struct ieee80211_txq *txq)
740 {
741 return (void *)txq->drv_priv;
742 }
743
744 static inline struct iwl_mvm_txq *
iwl_mvm_txq_from_tid(struct ieee80211_sta * sta,u8 tid)745 iwl_mvm_txq_from_tid(struct ieee80211_sta *sta, u8 tid)
746 {
747 if (tid == IWL_MAX_TID_COUNT)
748 tid = IEEE80211_NUM_TIDS;
749
750 return (void *)sta->txq[tid]->drv_priv;
751 }
752
753 /**
754 * struct iwl_mvm_tvqm_txq_info - maps TVQM hw queue to tid
755 *
756 * @sta_id: sta id
757 * @txq_tid: txq tid
758 */
759 struct iwl_mvm_tvqm_txq_info {
760 u8 sta_id;
761 u8 txq_tid;
762 };
763
764 struct iwl_mvm_dqa_txq_info {
765 u8 ra_sta_id; /* The RA this queue is mapped to, if exists */
766 bool reserved; /* Is this the TXQ reserved for a STA */
767 u8 mac80211_ac; /* The mac80211 AC this queue is mapped to */
768 u8 txq_tid; /* The TID "owner" of this queue*/
769 u16 tid_bitmap; /* Bitmap of the TIDs mapped to this queue */
770 /* Timestamp for inactivation per TID of this queue */
771 unsigned long last_frame_time[IWL_MAX_TID_COUNT + 1];
772 enum iwl_mvm_queue_status status;
773 };
774
775 struct iwl_mvm {
776 /* for logger access */
777 struct device *dev;
778
779 struct iwl_trans *trans;
780 const struct iwl_fw *fw;
781 const struct iwl_cfg *cfg;
782 struct iwl_phy_db *phy_db;
783 struct ieee80211_hw *hw;
784
785 /* for protecting access to iwl_mvm */
786 struct mutex mutex;
787 struct list_head async_handlers_list;
788 spinlock_t async_handlers_lock;
789 struct work_struct async_handlers_wk;
790
791 struct work_struct roc_done_wk;
792
793 unsigned long init_status;
794
795 unsigned long status;
796
797 u32 queue_sync_cookie;
798 unsigned long queue_sync_state;
799 /*
800 * for beacon filtering -
801 * currently only one interface can be supported
802 */
803 struct iwl_mvm_vif *bf_allowed_vif;
804
805 bool hw_registered;
806 bool rfkill_safe_init_done;
807
808 u8 cca_40mhz_workaround;
809
810 u32 ampdu_ref;
811 bool ampdu_toggle;
812
813 struct iwl_notif_wait_data notif_wait;
814
815 union {
816 struct mvm_statistics_rx_v3 rx_stats_v3;
817 struct mvm_statistics_rx rx_stats;
818 };
819
820 struct {
821 u64 rx_time;
822 u64 tx_time;
823 u64 on_time_rf;
824 u64 on_time_scan;
825 } radio_stats, accu_radio_stats;
826
827 struct list_head add_stream_txqs;
828 union {
829 struct iwl_mvm_dqa_txq_info queue_info[IWL_MAX_HW_QUEUES];
830 struct iwl_mvm_tvqm_txq_info tvqm_info[IWL_MAX_TVQM_QUEUES];
831 };
832 struct work_struct add_stream_wk; /* To add streams to queues */
833 spinlock_t add_stream_lock;
834
835 const char *nvm_file_name;
836 struct iwl_nvm_data *nvm_data;
837 struct iwl_mei_nvm *mei_nvm_data;
838 struct iwl_mvm_csme_conn_info __rcu *csme_conn_info;
839 bool mei_rfkill_blocked;
840 bool mei_registered;
841 struct work_struct sap_connected_wk;
842
843 /*
844 * NVM built based on the SAP data but that we can't free even after
845 * we get ownership because it contains the cfg80211's channel.
846 */
847 struct iwl_nvm_data *temp_nvm_data;
848
849 /* NVM sections */
850 struct iwl_nvm_section nvm_sections[NVM_MAX_NUM_SECTIONS];
851
852 struct iwl_fw_runtime fwrt;
853
854 /* EEPROM MAC addresses */
855 struct mac_address addresses[IWL_MVM_MAX_ADDRESSES];
856
857 /* data related to data path */
858 struct iwl_rx_phy_info last_phy_info;
859 struct ieee80211_sta __rcu *fw_id_to_mac_id[IWL_MVM_STATION_COUNT_MAX];
860 u8 rx_ba_sessions;
861
862 /* configured by mac80211 */
863 u32 rts_threshold;
864
865 /* Scan status, cmd (pre-allocated) and auxiliary station */
866 unsigned int scan_status;
867 size_t scan_cmd_size;
868 void *scan_cmd;
869 struct iwl_mcast_filter_cmd *mcast_filter_cmd;
870 /* For CDB this is low band scan type, for non-CDB - type. */
871 enum iwl_mvm_scan_type scan_type;
872 enum iwl_mvm_scan_type hb_scan_type;
873
874 enum iwl_mvm_sched_scan_pass_all_states sched_scan_pass_all;
875 struct delayed_work scan_timeout_dwork;
876
877 /* max number of simultaneous scans the FW supports */
878 unsigned int max_scans;
879
880 /* UMAC scan tracking */
881 u32 scan_uid_status[IWL_MVM_MAX_UMAC_SCANS];
882
883 /* start time of last scan in TSF of the mac that requested the scan */
884 u64 scan_start;
885
886 /* the vif that requested the current scan */
887 struct iwl_mvm_vif *scan_vif;
888
889 /* rx chain antennas set through debugfs for the scan command */
890 u8 scan_rx_ant;
891
892 /* Internal station */
893 struct iwl_mvm_int_sta aux_sta;
894 struct iwl_mvm_int_sta snif_sta;
895
896 bool last_ebs_successful;
897
898 u8 scan_last_antenna_idx; /* to toggle TX between antennas */
899 u8 mgmt_last_antenna_idx;
900
901 /* last smart fifo state that was successfully sent to firmware */
902 enum iwl_sf_state sf_state;
903
904 /*
905 * Leave this pointer outside the ifdef below so that it can be
906 * assigned without ifdef in the source code.
907 */
908 struct dentry *debugfs_dir;
909 #ifdef CONFIG_IWLWIFI_DEBUGFS
910 u32 dbgfs_sram_offset, dbgfs_sram_len;
911 u32 dbgfs_prph_reg_addr;
912 bool disable_power_off;
913 bool disable_power_off_d3;
914 bool beacon_inject_active;
915
916 bool scan_iter_notif_enabled;
917
918 struct debugfs_blob_wrapper nvm_hw_blob;
919 struct debugfs_blob_wrapper nvm_sw_blob;
920 struct debugfs_blob_wrapper nvm_calib_blob;
921 struct debugfs_blob_wrapper nvm_prod_blob;
922 struct debugfs_blob_wrapper nvm_phy_sku_blob;
923 struct debugfs_blob_wrapper nvm_reg_blob;
924
925 struct iwl_mvm_frame_stats drv_rx_stats;
926 spinlock_t drv_stats_lock;
927 u16 dbgfs_rx_phyinfo;
928 #endif
929
930 struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX];
931
932 struct list_head time_event_list;
933 spinlock_t time_event_lock;
934
935 /*
936 * A bitmap indicating the index of the key in use. The firmware
937 * can hold 16 keys at most. Reflect this fact.
938 */
939 unsigned long fw_key_table[BITS_TO_LONGS(STA_KEY_MAX_NUM)];
940 u8 fw_key_deleted[STA_KEY_MAX_NUM];
941
942 struct ieee80211_vif __rcu *vif_id_to_mac[NUM_MAC_INDEX_DRIVER];
943
944 /* -1 for always, 0 for never, >0 for that many times */
945 s8 fw_restart;
946 u8 *error_recovery_buf;
947
948 #ifdef CONFIG_IWLWIFI_LEDS
949 struct led_classdev led;
950 #endif
951
952 struct ieee80211_vif *p2p_device_vif;
953
954 #ifdef CONFIG_PM
955 struct wiphy_wowlan_support wowlan;
956 int gtk_ivlen, gtk_icvlen, ptk_ivlen, ptk_icvlen;
957
958 /* sched scan settings for net detect */
959 struct ieee80211_scan_ies nd_ies;
960 struct cfg80211_match_set *nd_match_sets;
961 int n_nd_match_sets;
962 struct ieee80211_channel **nd_channels;
963 int n_nd_channels;
964 bool net_detect;
965 u8 offload_tid;
966 #ifdef CONFIG_IWLWIFI_DEBUGFS
967 bool d3_wake_sysassert;
968 bool d3_test_active;
969 u32 d3_test_pme_ptr;
970 struct ieee80211_vif *keep_vif;
971 u32 last_netdetect_scans; /* no. of scans in the last net-detect wake */
972 #endif
973 #endif
974
975 wait_queue_head_t rx_sync_waitq;
976
977 /* BT-Coex */
978 struct iwl_bt_coex_profile_notif last_bt_notif;
979 struct iwl_bt_coex_ci_cmd last_bt_ci_cmd;
980
981 u8 bt_tx_prio;
982 enum iwl_bt_force_ant_mode bt_force_ant_mode;
983
984 /* Aux ROC */
985 struct list_head aux_roc_te_list;
986
987 /* Thermal Throttling and CTkill */
988 struct iwl_mvm_tt_mgmt thermal_throttle;
989 #ifdef CONFIG_THERMAL
990 struct iwl_mvm_thermal_device tz_device;
991 struct iwl_mvm_cooling_device cooling_dev;
992 #endif
993
994 s32 temperature; /* Celsius */
995 /*
996 * Debug option to set the NIC temperature. This option makes the
997 * driver think this is the actual NIC temperature, and ignore the
998 * real temperature that is received from the fw
999 */
1000 bool temperature_test; /* Debug test temperature is enabled */
1001
1002 bool fw_static_smps_request;
1003
1004 unsigned long bt_coex_last_tcm_ts;
1005 struct iwl_mvm_tcm tcm;
1006
1007 u8 uapsd_noagg_bssid_write_idx;
1008 struct mac_address uapsd_noagg_bssids[IWL_MVM_UAPSD_NOAGG_BSSIDS_NUM]
1009 __aligned(2);
1010
1011 struct iwl_time_quota_cmd last_quota_cmd;
1012
1013 #ifdef CONFIG_NL80211_TESTMODE
1014 u32 noa_duration;
1015 struct ieee80211_vif *noa_vif;
1016 #endif
1017
1018 /* Tx queues */
1019 u16 aux_queue;
1020 u16 snif_queue;
1021 u16 probe_queue;
1022 u16 p2p_dev_queue;
1023
1024 /* Indicate if device power save is allowed */
1025 u8 ps_disabled; /* u8 instead of bool to ease debugfs_create_* usage */
1026 /* Indicate if 32Khz external clock is valid */
1027 u32 ext_clock_valid;
1028
1029 /* This vif used by CSME to send / receive traffic */
1030 struct ieee80211_vif *csme_vif;
1031 struct ieee80211_vif __rcu *csa_vif;
1032 struct ieee80211_vif __rcu *csa_tx_blocked_vif;
1033 u8 csa_tx_block_bcn_timeout;
1034
1035 /* system time of last beacon (for AP/GO interface) */
1036 u32 ap_last_beacon_gp2;
1037
1038 /* indicates that we transmitted the last beacon */
1039 bool ibss_manager;
1040
1041 bool lar_regdom_set;
1042 enum iwl_mcc_source mcc_src;
1043
1044 /* TDLS channel switch data */
1045 struct {
1046 struct delayed_work dwork;
1047 enum iwl_mvm_tdls_cs_state state;
1048
1049 /*
1050 * Current cs sta - might be different from periodic cs peer
1051 * station. Value is meaningless when the cs-state is idle.
1052 */
1053 u8 cur_sta_id;
1054
1055 /* TDLS periodic channel-switch peer */
1056 struct {
1057 u8 sta_id;
1058 u8 op_class;
1059 bool initiator; /* are we the link initiator */
1060 struct cfg80211_chan_def chandef;
1061 struct sk_buff *skb; /* ch sw template */
1062 u32 ch_sw_tm_ie;
1063
1064 /* timestamp of last ch-sw request sent (GP2 time) */
1065 u32 sent_timestamp;
1066 } peer;
1067 } tdls_cs;
1068
1069
1070 u32 ciphers[IWL_MVM_NUM_CIPHERS];
1071
1072 struct cfg80211_ftm_responder_stats ftm_resp_stats;
1073 struct {
1074 struct cfg80211_pmsr_request *req;
1075 struct wireless_dev *req_wdev;
1076 struct list_head loc_list;
1077 int responses[IWL_MVM_TOF_MAX_APS];
1078 struct {
1079 struct list_head resp;
1080 } smooth;
1081 struct list_head pasn_list;
1082 } ftm_initiator;
1083
1084 struct list_head resp_pasn_list;
1085
1086 struct {
1087 u8 range_resp;
1088 } cmd_ver;
1089
1090 struct ieee80211_vif *nan_vif;
1091 struct iwl_mvm_baid_data __rcu *baid_map[IWL_MAX_BAID];
1092
1093 /*
1094 * Drop beacons from other APs in AP mode when there are no connected
1095 * clients.
1096 */
1097 bool drop_bcn_ap_mode;
1098
1099 struct delayed_work cs_tx_unblock_dwork;
1100
1101 /* does a monitor vif exist (only one can exist hence bool) */
1102 bool monitor_on;
1103
1104 /* sniffer data to include in radiotap */
1105 __le16 cur_aid;
1106 u8 cur_bssid[ETH_ALEN];
1107
1108 unsigned long last_6ghz_passive_scan_jiffies;
1109 unsigned long last_reset_or_resume_time_jiffies;
1110
1111 bool sta_remove_requires_queue_remove;
1112 };
1113
1114 /* Extract MVM priv from op_mode and _hw */
1115 #define IWL_OP_MODE_GET_MVM(_iwl_op_mode) \
1116 ((struct iwl_mvm *)(_iwl_op_mode)->op_mode_specific)
1117
1118 #define IWL_MAC80211_GET_MVM(_hw) \
1119 IWL_OP_MODE_GET_MVM((struct iwl_op_mode *)((_hw)->priv))
1120
1121 /**
1122 * enum iwl_mvm_status - MVM status bits
1123 * @IWL_MVM_STATUS_HW_RFKILL: HW RF-kill is asserted
1124 * @IWL_MVM_STATUS_HW_CTKILL: CT-kill is active
1125 * @IWL_MVM_STATUS_ROC_RUNNING: remain-on-channel is running
1126 * @IWL_MVM_STATUS_HW_RESTART_REQUESTED: HW restart was requested
1127 * @IWL_MVM_STATUS_IN_HW_RESTART: HW restart is active
1128 * @IWL_MVM_STATUS_ROC_AUX_RUNNING: AUX remain-on-channel is running
1129 * @IWL_MVM_STATUS_FIRMWARE_RUNNING: firmware is running
1130 * @IWL_MVM_STATUS_NEED_FLUSH_P2P: need to flush P2P bcast STA
1131 * @IWL_MVM_STATUS_IN_D3: in D3 (or at least about to go into it)
1132 * @IWL_MVM_STATUS_SUPPRESS_ERROR_LOG_ONCE: suppress one error log
1133 * if this is set, when intentionally triggered
1134 * @IWL_MVM_STATUS_STARTING: starting mac,
1135 * used to disable restart flow while in STARTING state
1136 */
1137 enum iwl_mvm_status {
1138 IWL_MVM_STATUS_HW_RFKILL,
1139 IWL_MVM_STATUS_HW_CTKILL,
1140 IWL_MVM_STATUS_ROC_RUNNING,
1141 IWL_MVM_STATUS_HW_RESTART_REQUESTED,
1142 IWL_MVM_STATUS_IN_HW_RESTART,
1143 IWL_MVM_STATUS_ROC_AUX_RUNNING,
1144 IWL_MVM_STATUS_FIRMWARE_RUNNING,
1145 IWL_MVM_STATUS_NEED_FLUSH_P2P,
1146 IWL_MVM_STATUS_IN_D3,
1147 IWL_MVM_STATUS_SUPPRESS_ERROR_LOG_ONCE,
1148 IWL_MVM_STATUS_STARTING,
1149 };
1150
1151 struct iwl_mvm_csme_conn_info {
1152 struct rcu_head rcu_head;
1153 struct iwl_mei_conn_info conn_info;
1154 };
1155
1156 /* Keep track of completed init configuration */
1157 enum iwl_mvm_init_status {
1158 IWL_MVM_INIT_STATUS_THERMAL_INIT_COMPLETE = BIT(0),
1159 IWL_MVM_INIT_STATUS_LEDS_INIT_COMPLETE = BIT(1),
1160 };
1161
iwl_mvm_is_radio_killed(struct iwl_mvm * mvm)1162 static inline bool iwl_mvm_is_radio_killed(struct iwl_mvm *mvm)
1163 {
1164 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status) ||
1165 test_bit(IWL_MVM_STATUS_HW_CTKILL, &mvm->status);
1166 }
1167
iwl_mvm_is_radio_hw_killed(struct iwl_mvm * mvm)1168 static inline bool iwl_mvm_is_radio_hw_killed(struct iwl_mvm *mvm)
1169 {
1170 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status);
1171 }
1172
iwl_mvm_firmware_running(struct iwl_mvm * mvm)1173 static inline bool iwl_mvm_firmware_running(struct iwl_mvm *mvm)
1174 {
1175 return test_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1176 }
1177
1178 /* Must be called with rcu_read_lock() held and it can only be
1179 * released when mvmsta is not needed anymore.
1180 */
1181 static inline struct iwl_mvm_sta *
iwl_mvm_sta_from_staid_rcu(struct iwl_mvm * mvm,u8 sta_id)1182 iwl_mvm_sta_from_staid_rcu(struct iwl_mvm *mvm, u8 sta_id)
1183 {
1184 struct ieee80211_sta *sta;
1185
1186 if (sta_id >= mvm->fw->ucode_capa.num_stations)
1187 return NULL;
1188
1189 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]);
1190
1191 /* This can happen if the station has been removed right now */
1192 if (IS_ERR_OR_NULL(sta))
1193 return NULL;
1194
1195 return iwl_mvm_sta_from_mac80211(sta);
1196 }
1197
1198 static inline struct iwl_mvm_sta *
iwl_mvm_sta_from_staid_protected(struct iwl_mvm * mvm,u8 sta_id)1199 iwl_mvm_sta_from_staid_protected(struct iwl_mvm *mvm, u8 sta_id)
1200 {
1201 struct ieee80211_sta *sta;
1202
1203 if (sta_id >= mvm->fw->ucode_capa.num_stations)
1204 return NULL;
1205
1206 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id],
1207 lockdep_is_held(&mvm->mutex));
1208
1209 /* This can happen if the station has been removed right now */
1210 if (IS_ERR_OR_NULL(sta))
1211 return NULL;
1212
1213 return iwl_mvm_sta_from_mac80211(sta);
1214 }
1215
1216 static inline struct ieee80211_vif *
iwl_mvm_rcu_dereference_vif_id(struct iwl_mvm * mvm,u8 vif_id,bool rcu)1217 iwl_mvm_rcu_dereference_vif_id(struct iwl_mvm *mvm, u8 vif_id, bool rcu)
1218 {
1219 if (WARN_ON(vif_id >= ARRAY_SIZE(mvm->vif_id_to_mac)))
1220 return NULL;
1221
1222 if (rcu)
1223 return rcu_dereference(mvm->vif_id_to_mac[vif_id]);
1224
1225 return rcu_dereference_protected(mvm->vif_id_to_mac[vif_id],
1226 lockdep_is_held(&mvm->mutex));
1227 }
1228
iwl_mvm_is_adaptive_dwell_supported(struct iwl_mvm * mvm)1229 static inline bool iwl_mvm_is_adaptive_dwell_supported(struct iwl_mvm *mvm)
1230 {
1231 return fw_has_api(&mvm->fw->ucode_capa,
1232 IWL_UCODE_TLV_API_ADAPTIVE_DWELL);
1233 }
1234
iwl_mvm_is_adaptive_dwell_v2_supported(struct iwl_mvm * mvm)1235 static inline bool iwl_mvm_is_adaptive_dwell_v2_supported(struct iwl_mvm *mvm)
1236 {
1237 return fw_has_api(&mvm->fw->ucode_capa,
1238 IWL_UCODE_TLV_API_ADAPTIVE_DWELL_V2);
1239 }
1240
iwl_mvm_is_adwell_hb_ap_num_supported(struct iwl_mvm * mvm)1241 static inline bool iwl_mvm_is_adwell_hb_ap_num_supported(struct iwl_mvm *mvm)
1242 {
1243 return fw_has_api(&mvm->fw->ucode_capa,
1244 IWL_UCODE_TLV_API_ADWELL_HB_DEF_N_AP);
1245 }
1246
iwl_mvm_is_oce_supported(struct iwl_mvm * mvm)1247 static inline bool iwl_mvm_is_oce_supported(struct iwl_mvm *mvm)
1248 {
1249 /* OCE should never be enabled for LMAC scan FWs */
1250 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_OCE);
1251 }
1252
iwl_mvm_is_frag_ebs_supported(struct iwl_mvm * mvm)1253 static inline bool iwl_mvm_is_frag_ebs_supported(struct iwl_mvm *mvm)
1254 {
1255 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_FRAG_EBS);
1256 }
1257
iwl_mvm_is_short_beacon_notif_supported(struct iwl_mvm * mvm)1258 static inline bool iwl_mvm_is_short_beacon_notif_supported(struct iwl_mvm *mvm)
1259 {
1260 return fw_has_api(&mvm->fw->ucode_capa,
1261 IWL_UCODE_TLV_API_SHORT_BEACON_NOTIF);
1262 }
1263
iwl_mvm_is_dqa_data_queue(struct iwl_mvm * mvm,u8 queue)1264 static inline bool iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue)
1265 {
1266 return (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE) &&
1267 (queue <= IWL_MVM_DQA_MAX_DATA_QUEUE);
1268 }
1269
iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm * mvm,u8 queue)1270 static inline bool iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue)
1271 {
1272 return (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE) &&
1273 (queue <= IWL_MVM_DQA_MAX_MGMT_QUEUE);
1274 }
1275
iwl_mvm_is_lar_supported(struct iwl_mvm * mvm)1276 static inline bool iwl_mvm_is_lar_supported(struct iwl_mvm *mvm)
1277 {
1278 bool nvm_lar = mvm->nvm_data->lar_enabled;
1279 bool tlv_lar = fw_has_capa(&mvm->fw->ucode_capa,
1280 IWL_UCODE_TLV_CAPA_LAR_SUPPORT);
1281
1282 /*
1283 * Enable LAR only if it is supported by the FW (TLV) &&
1284 * enabled in the NVM
1285 */
1286 if (mvm->cfg->nvm_type == IWL_NVM_EXT)
1287 return nvm_lar && tlv_lar;
1288 else
1289 return tlv_lar;
1290 }
1291
iwl_mvm_is_wifi_mcc_supported(struct iwl_mvm * mvm)1292 static inline bool iwl_mvm_is_wifi_mcc_supported(struct iwl_mvm *mvm)
1293 {
1294 return fw_has_api(&mvm->fw->ucode_capa,
1295 IWL_UCODE_TLV_API_WIFI_MCC_UPDATE) ||
1296 fw_has_capa(&mvm->fw->ucode_capa,
1297 IWL_UCODE_TLV_CAPA_LAR_MULTI_MCC);
1298 }
1299
iwl_mvm_bt_is_rrc_supported(struct iwl_mvm * mvm)1300 static inline bool iwl_mvm_bt_is_rrc_supported(struct iwl_mvm *mvm)
1301 {
1302 return fw_has_capa(&mvm->fw->ucode_capa,
1303 IWL_UCODE_TLV_CAPA_BT_COEX_RRC) &&
1304 IWL_MVM_BT_COEX_RRC;
1305 }
1306
iwl_mvm_is_csum_supported(struct iwl_mvm * mvm)1307 static inline bool iwl_mvm_is_csum_supported(struct iwl_mvm *mvm)
1308 {
1309 return fw_has_capa(&mvm->fw->ucode_capa,
1310 IWL_UCODE_TLV_CAPA_CSUM_SUPPORT) &&
1311 !IWL_MVM_HW_CSUM_DISABLE;
1312 }
1313
iwl_mvm_is_mplut_supported(struct iwl_mvm * mvm)1314 static inline bool iwl_mvm_is_mplut_supported(struct iwl_mvm *mvm)
1315 {
1316 return fw_has_capa(&mvm->fw->ucode_capa,
1317 IWL_UCODE_TLV_CAPA_BT_MPLUT_SUPPORT) &&
1318 IWL_MVM_BT_COEX_MPLUT;
1319 }
1320
1321 static inline
iwl_mvm_is_p2p_scm_uapsd_supported(struct iwl_mvm * mvm)1322 bool iwl_mvm_is_p2p_scm_uapsd_supported(struct iwl_mvm *mvm)
1323 {
1324 return fw_has_capa(&mvm->fw->ucode_capa,
1325 IWL_UCODE_TLV_CAPA_P2P_SCM_UAPSD) &&
1326 !(iwlwifi_mod_params.uapsd_disable &
1327 IWL_DISABLE_UAPSD_P2P_CLIENT);
1328 }
1329
iwl_mvm_has_new_rx_api(struct iwl_mvm * mvm)1330 static inline bool iwl_mvm_has_new_rx_api(struct iwl_mvm *mvm)
1331 {
1332 return fw_has_capa(&mvm->fw->ucode_capa,
1333 IWL_UCODE_TLV_CAPA_MULTI_QUEUE_RX_SUPPORT);
1334 }
1335
iwl_mvm_has_new_tx_api(struct iwl_mvm * mvm)1336 static inline bool iwl_mvm_has_new_tx_api(struct iwl_mvm *mvm)
1337 {
1338 /* TODO - replace with TLV once defined */
1339 return mvm->trans->trans_cfg->use_tfh;
1340 }
1341
iwl_mvm_has_unified_ucode(struct iwl_mvm * mvm)1342 static inline bool iwl_mvm_has_unified_ucode(struct iwl_mvm *mvm)
1343 {
1344 /* TODO - better define this */
1345 return mvm->trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_22000;
1346 }
1347
iwl_mvm_is_cdb_supported(struct iwl_mvm * mvm)1348 static inline bool iwl_mvm_is_cdb_supported(struct iwl_mvm *mvm)
1349 {
1350 /*
1351 * TODO:
1352 * The issue of how to determine CDB APIs and usage is still not fully
1353 * defined.
1354 * There is a compilation for CDB and non-CDB FW, but there may
1355 * be also runtime check.
1356 * For now there is a TLV for checking compilation mode, but a
1357 * runtime check will also have to be here - once defined.
1358 */
1359 return fw_has_capa(&mvm->fw->ucode_capa,
1360 IWL_UCODE_TLV_CAPA_CDB_SUPPORT);
1361 }
1362
iwl_mvm_cdb_scan_api(struct iwl_mvm * mvm)1363 static inline bool iwl_mvm_cdb_scan_api(struct iwl_mvm *mvm)
1364 {
1365 /*
1366 * TODO: should this be the same as iwl_mvm_is_cdb_supported()?
1367 * but then there's a little bit of code in scan that won't make
1368 * any sense...
1369 */
1370 return mvm->trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_22000;
1371 }
1372
iwl_mvm_is_scan_ext_chan_supported(struct iwl_mvm * mvm)1373 static inline bool iwl_mvm_is_scan_ext_chan_supported(struct iwl_mvm *mvm)
1374 {
1375 return fw_has_api(&mvm->fw->ucode_capa,
1376 IWL_UCODE_TLV_API_SCAN_EXT_CHAN_VER);
1377 }
1378
1379
iwl_mvm_is_reduced_config_scan_supported(struct iwl_mvm * mvm)1380 static inline bool iwl_mvm_is_reduced_config_scan_supported(struct iwl_mvm *mvm)
1381 {
1382 return fw_has_api(&mvm->fw->ucode_capa,
1383 IWL_UCODE_TLV_API_REDUCED_SCAN_CONFIG);
1384 }
1385
iwl_mvm_is_band_in_rx_supported(struct iwl_mvm * mvm)1386 static inline bool iwl_mvm_is_band_in_rx_supported(struct iwl_mvm *mvm)
1387 {
1388 return fw_has_api(&mvm->fw->ucode_capa,
1389 IWL_UCODE_TLV_API_BAND_IN_RX_DATA);
1390 }
1391
iwl_mvm_has_new_rx_stats_api(struct iwl_mvm * mvm)1392 static inline bool iwl_mvm_has_new_rx_stats_api(struct iwl_mvm *mvm)
1393 {
1394 return fw_has_api(&mvm->fw->ucode_capa,
1395 IWL_UCODE_TLV_API_NEW_RX_STATS);
1396 }
1397
iwl_mvm_has_quota_low_latency(struct iwl_mvm * mvm)1398 static inline bool iwl_mvm_has_quota_low_latency(struct iwl_mvm *mvm)
1399 {
1400 return fw_has_api(&mvm->fw->ucode_capa,
1401 IWL_UCODE_TLV_API_QUOTA_LOW_LATENCY);
1402 }
1403
iwl_mvm_has_tlc_offload(const struct iwl_mvm * mvm)1404 static inline bool iwl_mvm_has_tlc_offload(const struct iwl_mvm *mvm)
1405 {
1406 return fw_has_capa(&mvm->fw->ucode_capa,
1407 IWL_UCODE_TLV_CAPA_TLC_OFFLOAD);
1408 }
1409
1410 static inline struct agg_tx_status *
iwl_mvm_get_agg_status(struct iwl_mvm * mvm,void * tx_resp)1411 iwl_mvm_get_agg_status(struct iwl_mvm *mvm, void *tx_resp)
1412 {
1413 if (iwl_mvm_has_new_tx_api(mvm))
1414 return &((struct iwl_mvm_tx_resp *)tx_resp)->status;
1415 else
1416 return ((struct iwl_mvm_tx_resp_v3 *)tx_resp)->status;
1417 }
1418
iwl_mvm_is_tt_in_fw(struct iwl_mvm * mvm)1419 static inline bool iwl_mvm_is_tt_in_fw(struct iwl_mvm *mvm)
1420 {
1421 /* these two TLV are redundant since the responsibility to CT-kill by
1422 * FW happens only after we send at least one command of
1423 * temperature THs report.
1424 */
1425 return fw_has_capa(&mvm->fw->ucode_capa,
1426 IWL_UCODE_TLV_CAPA_CT_KILL_BY_FW) &&
1427 fw_has_capa(&mvm->fw->ucode_capa,
1428 IWL_UCODE_TLV_CAPA_TEMP_THS_REPORT_SUPPORT);
1429 }
1430
iwl_mvm_is_ctdp_supported(struct iwl_mvm * mvm)1431 static inline bool iwl_mvm_is_ctdp_supported(struct iwl_mvm *mvm)
1432 {
1433 return fw_has_capa(&mvm->fw->ucode_capa,
1434 IWL_UCODE_TLV_CAPA_CTDP_SUPPORT);
1435 }
1436
1437 extern const u8 iwl_mvm_ac_to_tx_fifo[];
1438 extern const u8 iwl_mvm_ac_to_gen2_tx_fifo[];
1439
iwl_mvm_mac_ac_to_tx_fifo(struct iwl_mvm * mvm,enum ieee80211_ac_numbers ac)1440 static inline u8 iwl_mvm_mac_ac_to_tx_fifo(struct iwl_mvm *mvm,
1441 enum ieee80211_ac_numbers ac)
1442 {
1443 return iwl_mvm_has_new_tx_api(mvm) ?
1444 iwl_mvm_ac_to_gen2_tx_fifo[ac] : iwl_mvm_ac_to_tx_fifo[ac];
1445 }
1446
1447 struct iwl_rate_info {
1448 u8 plcp; /* uCode API: IWL_RATE_6M_PLCP, etc. */
1449 u8 plcp_siso; /* uCode API: IWL_RATE_SISO_6M_PLCP, etc. */
1450 u8 plcp_mimo2; /* uCode API: IWL_RATE_MIMO2_6M_PLCP, etc. */
1451 u8 plcp_mimo3; /* uCode API: IWL_RATE_MIMO3_6M_PLCP, etc. */
1452 u8 ieee; /* MAC header: IWL_RATE_6M_IEEE, etc. */
1453 };
1454
1455 void __iwl_mvm_mac_stop(struct iwl_mvm *mvm);
1456 int __iwl_mvm_mac_start(struct iwl_mvm *mvm);
1457
1458 /******************
1459 * MVM Methods
1460 ******************/
1461 /* uCode */
1462 int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm);
1463
1464 /* Utils */
1465 int iwl_mvm_legacy_hw_idx_to_mac80211_idx(u32 rate_n_flags,
1466 enum nl80211_band band);
1467 int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags,
1468 enum nl80211_band band);
1469 void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags,
1470 enum nl80211_band band,
1471 struct ieee80211_tx_rate *r);
1472 void iwl_mvm_hwrate_to_tx_rate_v1(u32 rate_n_flags,
1473 enum nl80211_band band,
1474 struct ieee80211_tx_rate *r);
1475 u8 iwl_mvm_mac80211_idx_to_hwrate(const struct iwl_fw *fw, int rate_idx);
1476 u8 iwl_mvm_mac80211_ac_to_ucode_ac(enum ieee80211_ac_numbers ac);
1477
iwl_mvm_dump_nic_error_log(struct iwl_mvm * mvm)1478 static inline void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm)
1479 {
1480 iwl_fwrt_dump_error_logs(&mvm->fwrt);
1481 }
1482
1483 u8 first_antenna(u8 mask);
1484 u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx);
1485 void iwl_mvm_get_sync_time(struct iwl_mvm *mvm, int clock_type, u32 *gp2,
1486 u64 *boottime, ktime_t *realtime);
1487 u32 iwl_mvm_get_systime(struct iwl_mvm *mvm);
1488
1489 /* Tx / Host Commands */
1490 int __must_check iwl_mvm_send_cmd(struct iwl_mvm *mvm,
1491 struct iwl_host_cmd *cmd);
1492 int __must_check iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u32 id,
1493 u32 flags, u16 len, const void *data);
1494 int __must_check iwl_mvm_send_cmd_status(struct iwl_mvm *mvm,
1495 struct iwl_host_cmd *cmd,
1496 u32 *status);
1497 int __must_check iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u32 id,
1498 u16 len, const void *data,
1499 u32 *status);
1500 int iwl_mvm_tx_skb_sta(struct iwl_mvm *mvm, struct sk_buff *skb,
1501 struct ieee80211_sta *sta);
1502 int iwl_mvm_tx_skb_non_sta(struct iwl_mvm *mvm, struct sk_buff *skb);
1503 void iwl_mvm_set_tx_cmd(struct iwl_mvm *mvm, struct sk_buff *skb,
1504 struct iwl_tx_cmd *tx_cmd,
1505 struct ieee80211_tx_info *info, u8 sta_id);
1506 void iwl_mvm_set_tx_cmd_rate(struct iwl_mvm *mvm, struct iwl_tx_cmd *tx_cmd,
1507 struct ieee80211_tx_info *info,
1508 struct ieee80211_sta *sta, __le16 fc);
1509 void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq);
1510 unsigned int iwl_mvm_max_amsdu_size(struct iwl_mvm *mvm,
1511 struct ieee80211_sta *sta,
1512 unsigned int tid);
1513 u32 iwl_mvm_tx_csum_bz(struct iwl_mvm *mvm, struct sk_buff *skb, bool amsdu);
1514
1515 #ifdef CONFIG_IWLWIFI_DEBUG
1516 const char *iwl_mvm_get_tx_fail_reason(u32 status);
1517 #else
iwl_mvm_get_tx_fail_reason(u32 status)1518 static inline const char *iwl_mvm_get_tx_fail_reason(u32 status) { return ""; }
1519 #endif
1520 int iwl_mvm_flush_tx_path(struct iwl_mvm *mvm, u32 tfd_msk);
1521 int iwl_mvm_flush_sta(struct iwl_mvm *mvm, void *sta, bool internal);
1522 int iwl_mvm_flush_sta_tids(struct iwl_mvm *mvm, u32 sta_id, u16 tids);
1523
1524 void iwl_mvm_async_handlers_purge(struct iwl_mvm *mvm);
1525
iwl_mvm_set_tx_cmd_ccmp(struct ieee80211_tx_info * info,struct iwl_tx_cmd * tx_cmd)1526 static inline void iwl_mvm_set_tx_cmd_ccmp(struct ieee80211_tx_info *info,
1527 struct iwl_tx_cmd *tx_cmd)
1528 {
1529 struct ieee80211_key_conf *keyconf = info->control.hw_key;
1530
1531 tx_cmd->sec_ctl = TX_CMD_SEC_CCM;
1532 memcpy(tx_cmd->key, keyconf->key, keyconf->keylen);
1533 }
1534
iwl_mvm_wait_for_async_handlers(struct iwl_mvm * mvm)1535 static inline void iwl_mvm_wait_for_async_handlers(struct iwl_mvm *mvm)
1536 {
1537 flush_work(&mvm->async_handlers_wk);
1538 }
1539
1540 /* Statistics */
1541 void iwl_mvm_handle_rx_statistics(struct iwl_mvm *mvm,
1542 struct iwl_rx_packet *pkt);
1543 void iwl_mvm_rx_statistics(struct iwl_mvm *mvm,
1544 struct iwl_rx_cmd_buffer *rxb);
1545 int iwl_mvm_request_statistics(struct iwl_mvm *mvm, bool clear);
1546 void iwl_mvm_accu_radio_stats(struct iwl_mvm *mvm);
1547
1548 /* NVM */
1549 int iwl_nvm_init(struct iwl_mvm *mvm);
1550 int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm);
1551
iwl_mvm_get_valid_tx_ant(struct iwl_mvm * mvm)1552 static inline u8 iwl_mvm_get_valid_tx_ant(struct iwl_mvm *mvm)
1553 {
1554 return mvm->nvm_data && mvm->nvm_data->valid_tx_ant ?
1555 mvm->fw->valid_tx_ant & mvm->nvm_data->valid_tx_ant :
1556 mvm->fw->valid_tx_ant;
1557 }
1558
iwl_mvm_get_valid_rx_ant(struct iwl_mvm * mvm)1559 static inline u8 iwl_mvm_get_valid_rx_ant(struct iwl_mvm *mvm)
1560 {
1561 return mvm->nvm_data && mvm->nvm_data->valid_rx_ant ?
1562 mvm->fw->valid_rx_ant & mvm->nvm_data->valid_rx_ant :
1563 mvm->fw->valid_rx_ant;
1564 }
1565
iwl_mvm_toggle_tx_ant(struct iwl_mvm * mvm,u8 * ant)1566 static inline void iwl_mvm_toggle_tx_ant(struct iwl_mvm *mvm, u8 *ant)
1567 {
1568 *ant = iwl_mvm_next_antenna(mvm, iwl_mvm_get_valid_tx_ant(mvm), *ant);
1569 }
1570
iwl_mvm_get_phy_config(struct iwl_mvm * mvm)1571 static inline u32 iwl_mvm_get_phy_config(struct iwl_mvm *mvm)
1572 {
1573 u32 phy_config = ~(FW_PHY_CFG_TX_CHAIN |
1574 FW_PHY_CFG_RX_CHAIN);
1575 u32 valid_rx_ant = iwl_mvm_get_valid_rx_ant(mvm);
1576 u32 valid_tx_ant = iwl_mvm_get_valid_tx_ant(mvm);
1577
1578 phy_config |= valid_tx_ant << FW_PHY_CFG_TX_CHAIN_POS |
1579 valid_rx_ant << FW_PHY_CFG_RX_CHAIN_POS;
1580
1581 return mvm->fw->phy_config & phy_config;
1582 }
1583
1584 int iwl_mvm_up(struct iwl_mvm *mvm);
1585 int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm);
1586
1587 int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm);
1588
1589 /*
1590 * FW notifications / CMD responses handlers
1591 * Convention: iwl_mvm_rx_<NAME OF THE CMD>
1592 */
1593 void iwl_mvm_rx_mq(struct iwl_op_mode *op_mode,
1594 struct napi_struct *napi,
1595 struct iwl_rx_cmd_buffer *rxb);
1596 void iwl_mvm_rx_rx_phy_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1597 void iwl_mvm_rx_rx_mpdu(struct iwl_mvm *mvm, struct napi_struct *napi,
1598 struct iwl_rx_cmd_buffer *rxb);
1599 void iwl_mvm_rx_mpdu_mq(struct iwl_mvm *mvm, struct napi_struct *napi,
1600 struct iwl_rx_cmd_buffer *rxb, int queue);
1601 void iwl_mvm_rx_monitor_no_data(struct iwl_mvm *mvm, struct napi_struct *napi,
1602 struct iwl_rx_cmd_buffer *rxb, int queue);
1603 void iwl_mvm_rx_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi,
1604 struct iwl_rx_cmd_buffer *rxb, int queue);
1605 void iwl_mvm_rx_bar_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi,
1606 struct iwl_rx_cmd_buffer *rxb, int queue);
1607 void iwl_mvm_rx_queue_notif(struct iwl_mvm *mvm, struct napi_struct *napi,
1608 struct iwl_rx_cmd_buffer *rxb, int queue);
1609 void iwl_mvm_rx_tx_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1610 void iwl_mvm_mfu_assert_dump_notif(struct iwl_mvm *mvm,
1611 struct iwl_rx_cmd_buffer *rxb);
1612 void iwl_mvm_send_recovery_cmd(struct iwl_mvm *mvm, u32 flags);
1613 void iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1614 void iwl_mvm_rx_ant_coupling_notif(struct iwl_mvm *mvm,
1615 struct iwl_rx_cmd_buffer *rxb);
1616 void iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1617 void iwl_mvm_rx_mfuart_notif(struct iwl_mvm *mvm,
1618 struct iwl_rx_cmd_buffer *rxb);
1619 void iwl_mvm_rx_shared_mem_cfg_notif(struct iwl_mvm *mvm,
1620 struct iwl_rx_cmd_buffer *rxb);
1621
1622 /* MVM PHY */
1623 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1624 struct cfg80211_chan_def *chandef,
1625 u8 chains_static, u8 chains_dynamic);
1626 int iwl_mvm_phy_ctxt_changed(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1627 struct cfg80211_chan_def *chandef,
1628 u8 chains_static, u8 chains_dynamic);
1629 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm,
1630 struct iwl_mvm_phy_ctxt *ctxt);
1631 void iwl_mvm_phy_ctxt_unref(struct iwl_mvm *mvm,
1632 struct iwl_mvm_phy_ctxt *ctxt);
1633 int iwl_mvm_phy_ctx_count(struct iwl_mvm *mvm);
1634 u8 iwl_mvm_get_channel_width(struct cfg80211_chan_def *chandef);
1635 u8 iwl_mvm_get_ctrl_pos(struct cfg80211_chan_def *chandef);
1636
1637 /* MAC (virtual interface) programming */
1638 int iwl_mvm_mac_ctxt_init(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1639 int iwl_mvm_mac_ctxt_add(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1640 int iwl_mvm_mac_ctxt_changed(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1641 bool force_assoc_off, const u8 *bssid_override);
1642 int iwl_mvm_mac_ctxt_remove(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1643 int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm,
1644 struct ieee80211_vif *vif);
1645 int iwl_mvm_mac_ctxt_send_beacon(struct iwl_mvm *mvm,
1646 struct ieee80211_vif *vif,
1647 struct sk_buff *beacon);
1648 int iwl_mvm_mac_ctxt_send_beacon_cmd(struct iwl_mvm *mvm,
1649 struct sk_buff *beacon,
1650 void *data, int len);
1651 u8 iwl_mvm_mac_ctxt_get_lowest_rate(struct ieee80211_tx_info *info,
1652 struct ieee80211_vif *vif);
1653 u16 iwl_mvm_mac_ctxt_get_beacon_flags(const struct iwl_fw *fw,
1654 u8 rate_idx);
1655 void iwl_mvm_mac_ctxt_set_tim(struct iwl_mvm *mvm,
1656 __le32 *tim_index, __le32 *tim_size,
1657 u8 *beacon, u32 frame_size);
1658 void iwl_mvm_rx_beacon_notif(struct iwl_mvm *mvm,
1659 struct iwl_rx_cmd_buffer *rxb);
1660 void iwl_mvm_rx_missed_beacons_notif(struct iwl_mvm *mvm,
1661 struct iwl_rx_cmd_buffer *rxb);
1662 void iwl_mvm_rx_stored_beacon_notif(struct iwl_mvm *mvm,
1663 struct iwl_rx_cmd_buffer *rxb);
1664 void iwl_mvm_mu_mimo_grp_notif(struct iwl_mvm *mvm,
1665 struct iwl_rx_cmd_buffer *rxb);
1666 void iwl_mvm_sta_pm_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1667 void iwl_mvm_window_status_notif(struct iwl_mvm *mvm,
1668 struct iwl_rx_cmd_buffer *rxb);
1669 void iwl_mvm_mac_ctxt_recalc_tsf_id(struct iwl_mvm *mvm,
1670 struct ieee80211_vif *vif);
1671 void iwl_mvm_probe_resp_data_notif(struct iwl_mvm *mvm,
1672 struct iwl_rx_cmd_buffer *rxb);
1673 void iwl_mvm_rx_missed_vap_notif(struct iwl_mvm *mvm,
1674 struct iwl_rx_cmd_buffer *rxb);
1675 void iwl_mvm_channel_switch_start_notif(struct iwl_mvm *mvm,
1676 struct iwl_rx_cmd_buffer *rxb);
1677 void iwl_mvm_channel_switch_error_notif(struct iwl_mvm *mvm,
1678 struct iwl_rx_cmd_buffer *rxb);
1679 /* Bindings */
1680 int iwl_mvm_binding_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1681 int iwl_mvm_binding_remove_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1682
1683 /* Quota management */
iwl_mvm_quota_cmd_size(struct iwl_mvm * mvm)1684 static inline size_t iwl_mvm_quota_cmd_size(struct iwl_mvm *mvm)
1685 {
1686 return iwl_mvm_has_quota_low_latency(mvm) ?
1687 sizeof(struct iwl_time_quota_cmd) :
1688 sizeof(struct iwl_time_quota_cmd_v1);
1689 }
1690
1691 static inline struct iwl_time_quota_data
iwl_mvm_quota_cmd_get_quota(struct iwl_mvm * mvm,struct iwl_time_quota_cmd * cmd,int i)1692 *iwl_mvm_quota_cmd_get_quota(struct iwl_mvm *mvm,
1693 struct iwl_time_quota_cmd *cmd,
1694 int i)
1695 {
1696 struct iwl_time_quota_data_v1 *quotas;
1697
1698 if (iwl_mvm_has_quota_low_latency(mvm))
1699 return &cmd->quotas[i];
1700
1701 quotas = (struct iwl_time_quota_data_v1 *)cmd->quotas;
1702 return (struct iwl_time_quota_data *)"as[i];
1703 }
1704
1705 int iwl_mvm_update_quotas(struct iwl_mvm *mvm, bool force_upload,
1706 struct ieee80211_vif *disabled_vif);
1707
1708 /* Scanning */
1709 int iwl_mvm_reg_scan_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1710 struct cfg80211_scan_request *req,
1711 struct ieee80211_scan_ies *ies);
1712 size_t iwl_mvm_scan_size(struct iwl_mvm *mvm);
1713 int iwl_mvm_scan_stop(struct iwl_mvm *mvm, int type, bool notify);
1714 int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm);
1715 void iwl_mvm_report_scan_aborted(struct iwl_mvm *mvm);
1716 void iwl_mvm_scan_timeout_wk(struct work_struct *work);
1717
1718 /* Scheduled scan */
1719 void iwl_mvm_rx_lmac_scan_complete_notif(struct iwl_mvm *mvm,
1720 struct iwl_rx_cmd_buffer *rxb);
1721 void iwl_mvm_rx_lmac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1722 struct iwl_rx_cmd_buffer *rxb);
1723 int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm,
1724 struct ieee80211_vif *vif,
1725 struct cfg80211_sched_scan_request *req,
1726 struct ieee80211_scan_ies *ies,
1727 int type);
1728 void iwl_mvm_rx_scan_match_found(struct iwl_mvm *mvm,
1729 struct iwl_rx_cmd_buffer *rxb);
1730
1731 /* UMAC scan */
1732 int iwl_mvm_config_scan(struct iwl_mvm *mvm);
1733 void iwl_mvm_rx_umac_scan_complete_notif(struct iwl_mvm *mvm,
1734 struct iwl_rx_cmd_buffer *rxb);
1735 void iwl_mvm_rx_umac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1736 struct iwl_rx_cmd_buffer *rxb);
1737
1738 /* MVM debugfs */
1739 #ifdef CONFIG_IWLWIFI_DEBUGFS
1740 void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm);
1741 void iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1742 void iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1743 #else
iwl_mvm_dbgfs_register(struct iwl_mvm * mvm)1744 static inline void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm)
1745 {
1746 }
1747 static inline void
iwl_mvm_vif_dbgfs_register(struct iwl_mvm * mvm,struct ieee80211_vif * vif)1748 iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1749 {
1750 }
1751 static inline void
iwl_mvm_vif_dbgfs_clean(struct iwl_mvm * mvm,struct ieee80211_vif * vif)1752 iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1753 {
1754 }
1755 #endif /* CONFIG_IWLWIFI_DEBUGFS */
1756
1757 /* rate scaling */
1758 int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq);
1759 void iwl_mvm_update_frame_stats(struct iwl_mvm *mvm, u32 rate, bool agg);
1760 int rs_pretty_print_rate_v1(char *buf, int bufsz, const u32 rate);
1761 void rs_update_last_rssi(struct iwl_mvm *mvm,
1762 struct iwl_mvm_sta *mvmsta,
1763 struct ieee80211_rx_status *rx_status);
1764
1765 /* power management */
1766 int iwl_mvm_power_update_device(struct iwl_mvm *mvm);
1767 int iwl_mvm_power_update_mac(struct iwl_mvm *mvm);
1768 int iwl_mvm_power_update_ps(struct iwl_mvm *mvm);
1769 int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1770 char *buf, int bufsz);
1771
1772 void iwl_mvm_power_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1773 void iwl_mvm_power_uapsd_misbehaving_ap_notif(struct iwl_mvm *mvm,
1774 struct iwl_rx_cmd_buffer *rxb);
1775
1776 #ifdef CONFIG_IWLWIFI_LEDS
1777 int iwl_mvm_leds_init(struct iwl_mvm *mvm);
1778 void iwl_mvm_leds_exit(struct iwl_mvm *mvm);
1779 void iwl_mvm_leds_sync(struct iwl_mvm *mvm);
1780 #else
iwl_mvm_leds_init(struct iwl_mvm * mvm)1781 static inline int iwl_mvm_leds_init(struct iwl_mvm *mvm)
1782 {
1783 return 0;
1784 }
iwl_mvm_leds_exit(struct iwl_mvm * mvm)1785 static inline void iwl_mvm_leds_exit(struct iwl_mvm *mvm)
1786 {
1787 }
iwl_mvm_leds_sync(struct iwl_mvm * mvm)1788 static inline void iwl_mvm_leds_sync(struct iwl_mvm *mvm)
1789 {
1790 }
1791 #endif
1792
1793 /* D3 (WoWLAN, NetDetect) */
1794 int iwl_mvm_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wowlan);
1795 int iwl_mvm_resume(struct ieee80211_hw *hw);
1796 void iwl_mvm_set_wakeup(struct ieee80211_hw *hw, bool enabled);
1797 void iwl_mvm_set_rekey_data(struct ieee80211_hw *hw,
1798 struct ieee80211_vif *vif,
1799 struct cfg80211_gtk_rekey_data *data);
1800 void iwl_mvm_ipv6_addr_change(struct ieee80211_hw *hw,
1801 struct ieee80211_vif *vif,
1802 struct inet6_dev *idev);
1803 void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw,
1804 struct ieee80211_vif *vif, int idx);
1805 extern const struct file_operations iwl_dbgfs_d3_test_ops;
1806 #ifdef CONFIG_PM
1807 void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm,
1808 struct ieee80211_vif *vif);
1809 #else
1810 static inline void
iwl_mvm_set_last_nonqos_seq(struct iwl_mvm * mvm,struct ieee80211_vif * vif)1811 iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1812 {
1813 }
1814 #endif
1815 void iwl_mvm_set_wowlan_qos_seq(struct iwl_mvm_sta *mvm_ap_sta,
1816 struct iwl_wowlan_config_cmd *cmd);
1817 int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm,
1818 struct ieee80211_vif *vif,
1819 bool disable_offloading,
1820 bool offload_ns,
1821 u32 cmd_flags);
1822
1823 /* BT Coex */
1824 int iwl_mvm_send_bt_init_conf(struct iwl_mvm *mvm);
1825 void iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm,
1826 struct iwl_rx_cmd_buffer *rxb);
1827 void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1828 enum ieee80211_rssi_event_data);
1829 void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm);
1830 u16 iwl_mvm_coex_agg_time_limit(struct iwl_mvm *mvm,
1831 struct ieee80211_sta *sta);
1832 bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm,
1833 struct ieee80211_sta *sta);
1834 bool iwl_mvm_bt_coex_is_ant_avail(struct iwl_mvm *mvm, u8 ant);
1835 bool iwl_mvm_bt_coex_is_shared_ant_avail(struct iwl_mvm *mvm);
1836 bool iwl_mvm_bt_coex_is_tpc_allowed(struct iwl_mvm *mvm,
1837 enum nl80211_band band);
1838 u8 iwl_mvm_bt_coex_get_single_ant_msk(struct iwl_mvm *mvm, u8 enabled_ants);
1839 u8 iwl_mvm_bt_coex_tx_prio(struct iwl_mvm *mvm, struct ieee80211_hdr *hdr,
1840 struct ieee80211_tx_info *info, u8 ac);
1841
1842 /* beacon filtering */
1843 #ifdef CONFIG_IWLWIFI_DEBUGFS
1844 void
1845 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1846 struct iwl_beacon_filter_cmd *cmd);
1847 #else
1848 static inline void
iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif * vif,struct iwl_beacon_filter_cmd * cmd)1849 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1850 struct iwl_beacon_filter_cmd *cmd)
1851 {}
1852 #endif
1853 int iwl_mvm_enable_beacon_filter(struct iwl_mvm *mvm,
1854 struct ieee80211_vif *vif,
1855 u32 flags);
1856 int iwl_mvm_disable_beacon_filter(struct iwl_mvm *mvm,
1857 struct ieee80211_vif *vif,
1858 u32 flags);
1859 /* SMPS */
1860 void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1861 enum iwl_mvm_smps_type_request req_type,
1862 enum ieee80211_smps_mode smps_request);
1863 bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm,
1864 struct iwl_mvm_phy_ctxt *ctxt);
1865 void iwl_mvm_apply_fw_smps_request(struct ieee80211_vif *vif);
1866
1867 /* Low latency */
1868 int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1869 bool low_latency,
1870 enum iwl_mvm_low_latency_cause cause);
1871 /* get SystemLowLatencyMode - only needed for beacon threshold? */
1872 bool iwl_mvm_low_latency(struct iwl_mvm *mvm);
1873 bool iwl_mvm_low_latency_band(struct iwl_mvm *mvm, enum nl80211_band band);
1874 void iwl_mvm_send_low_latency_cmd(struct iwl_mvm *mvm, bool low_latency,
1875 u16 mac_id);
1876
1877 /* get VMACLowLatencyMode */
iwl_mvm_vif_low_latency(struct iwl_mvm_vif * mvmvif)1878 static inline bool iwl_mvm_vif_low_latency(struct iwl_mvm_vif *mvmvif)
1879 {
1880 /*
1881 * should this consider associated/active/... state?
1882 *
1883 * Normally low-latency should only be active on interfaces
1884 * that are active, but at least with debugfs it can also be
1885 * enabled on interfaces that aren't active. However, when
1886 * interface aren't active then they aren't added into the
1887 * binding, so this has no real impact. For now, just return
1888 * the current desired low-latency state.
1889 */
1890 return mvmvif->low_latency_actual;
1891 }
1892
1893 static inline
iwl_mvm_vif_set_low_latency(struct iwl_mvm_vif * mvmvif,bool set,enum iwl_mvm_low_latency_cause cause)1894 void iwl_mvm_vif_set_low_latency(struct iwl_mvm_vif *mvmvif, bool set,
1895 enum iwl_mvm_low_latency_cause cause)
1896 {
1897 u8 new_state;
1898
1899 if (set)
1900 mvmvif->low_latency |= cause;
1901 else
1902 mvmvif->low_latency &= ~cause;
1903
1904 /*
1905 * if LOW_LATENCY_DEBUGFS_FORCE_ENABLE is enabled no changes are
1906 * allowed to actual mode.
1907 */
1908 if (mvmvif->low_latency & LOW_LATENCY_DEBUGFS_FORCE_ENABLE &&
1909 cause != LOW_LATENCY_DEBUGFS_FORCE_ENABLE)
1910 return;
1911
1912 if (cause == LOW_LATENCY_DEBUGFS_FORCE_ENABLE && set)
1913 /*
1914 * We enter force state
1915 */
1916 new_state = !!(mvmvif->low_latency &
1917 LOW_LATENCY_DEBUGFS_FORCE);
1918 else
1919 /*
1920 * Check if any other one set low latency
1921 */
1922 new_state = !!(mvmvif->low_latency &
1923 ~(LOW_LATENCY_DEBUGFS_FORCE_ENABLE |
1924 LOW_LATENCY_DEBUGFS_FORCE));
1925
1926 mvmvif->low_latency_actual = new_state;
1927 }
1928
1929 /* Return a bitmask with all the hw supported queues, except for the
1930 * command queue, which can't be flushed.
1931 */
iwl_mvm_flushable_queues(struct iwl_mvm * mvm)1932 static inline u32 iwl_mvm_flushable_queues(struct iwl_mvm *mvm)
1933 {
1934 return ((BIT(mvm->trans->trans_cfg->base_params->num_of_queues) - 1) &
1935 ~BIT(IWL_MVM_DQA_CMD_QUEUE));
1936 }
1937
1938 void iwl_mvm_stop_device(struct iwl_mvm *mvm);
1939
1940 /* Thermal management and CT-kill */
1941 void iwl_mvm_tt_tx_backoff(struct iwl_mvm *mvm, u32 backoff);
1942 void iwl_mvm_temp_notif(struct iwl_mvm *mvm,
1943 struct iwl_rx_cmd_buffer *rxb);
1944 void iwl_mvm_tt_handler(struct iwl_mvm *mvm);
1945 void iwl_mvm_thermal_initialize(struct iwl_mvm *mvm, u32 min_backoff);
1946 void iwl_mvm_thermal_exit(struct iwl_mvm *mvm);
1947 void iwl_mvm_set_hw_ctkill_state(struct iwl_mvm *mvm, bool state);
1948 int iwl_mvm_get_temp(struct iwl_mvm *mvm, s32 *temp);
1949 void iwl_mvm_ct_kill_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1950 void iwl_mvm_enter_ctkill(struct iwl_mvm *mvm);
1951 int iwl_mvm_send_temp_report_ths_cmd(struct iwl_mvm *mvm);
1952 int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 budget);
1953
1954 #if IS_ENABLED(CONFIG_IWLMEI)
1955
1956 /* vendor commands */
1957 void iwl_mvm_vendor_cmds_register(struct iwl_mvm *mvm);
1958
1959 #else
1960
iwl_mvm_vendor_cmds_register(struct iwl_mvm * mvm)1961 static inline void iwl_mvm_vendor_cmds_register(struct iwl_mvm *mvm) {}
1962
1963 #endif
1964
1965 /* Location Aware Regulatory */
1966 struct iwl_mcc_update_resp *
1967 iwl_mvm_update_mcc(struct iwl_mvm *mvm, const char *alpha2,
1968 enum iwl_mcc_source src_id);
1969 int iwl_mvm_init_mcc(struct iwl_mvm *mvm);
1970 void iwl_mvm_rx_chub_update_mcc(struct iwl_mvm *mvm,
1971 struct iwl_rx_cmd_buffer *rxb);
1972 struct ieee80211_regdomain *iwl_mvm_get_regdomain(struct wiphy *wiphy,
1973 const char *alpha2,
1974 enum iwl_mcc_source src_id,
1975 bool *changed);
1976 struct ieee80211_regdomain *iwl_mvm_get_current_regdomain(struct iwl_mvm *mvm,
1977 bool *changed);
1978 int iwl_mvm_init_fw_regd(struct iwl_mvm *mvm);
1979 void iwl_mvm_update_changed_regdom(struct iwl_mvm *mvm);
1980
1981 /* smart fifo */
1982 int iwl_mvm_sf_update(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1983 bool added_vif);
1984
1985 /* FTM responder */
1986 int iwl_mvm_ftm_start_responder(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1987 void iwl_mvm_ftm_restart_responder(struct iwl_mvm *mvm,
1988 struct ieee80211_vif *vif);
1989 void iwl_mvm_ftm_responder_stats(struct iwl_mvm *mvm,
1990 struct iwl_rx_cmd_buffer *rxb);
1991 int iwl_mvm_ftm_resp_remove_pasn_sta(struct iwl_mvm *mvm,
1992 struct ieee80211_vif *vif, u8 *addr);
1993 int iwl_mvm_ftm_respoder_add_pasn_sta(struct iwl_mvm *mvm,
1994 struct ieee80211_vif *vif,
1995 u8 *addr, u32 cipher, u8 *tk, u32 tk_len,
1996 u8 *hltk, u32 hltk_len);
1997 void iwl_mvm_ftm_responder_clear(struct iwl_mvm *mvm,
1998 struct ieee80211_vif *vif);
1999
2000 /* FTM initiator */
2001 void iwl_mvm_ftm_restart(struct iwl_mvm *mvm);
2002 void iwl_mvm_ftm_range_resp(struct iwl_mvm *mvm,
2003 struct iwl_rx_cmd_buffer *rxb);
2004 void iwl_mvm_ftm_lc_notif(struct iwl_mvm *mvm,
2005 struct iwl_rx_cmd_buffer *rxb);
2006 int iwl_mvm_ftm_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2007 struct cfg80211_pmsr_request *request);
2008 void iwl_mvm_ftm_abort(struct iwl_mvm *mvm, struct cfg80211_pmsr_request *req);
2009 void iwl_mvm_ftm_initiator_smooth_config(struct iwl_mvm *mvm);
2010 void iwl_mvm_ftm_initiator_smooth_stop(struct iwl_mvm *mvm);
2011 int iwl_mvm_ftm_add_pasn_sta(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2012 u8 *addr, u32 cipher, u8 *tk, u32 tk_len,
2013 u8 *hltk, u32 hltk_len);
2014 void iwl_mvm_ftm_remove_pasn_sta(struct iwl_mvm *mvm, u8 *addr);
2015
2016 /* TDLS */
2017
2018 /*
2019 * We use TID 4 (VI) as a FW-used-only TID when TDLS connections are present.
2020 * This TID is marked as used vs the AP and all connected TDLS peers.
2021 */
2022 #define IWL_MVM_TDLS_FW_TID 4
2023
2024 int iwl_mvm_tdls_sta_count(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
2025 void iwl_mvm_teardown_tdls_peers(struct iwl_mvm *mvm);
2026 void iwl_mvm_recalc_tdls_state(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2027 bool sta_added);
2028 void iwl_mvm_mac_mgd_protect_tdls_discover(struct ieee80211_hw *hw,
2029 struct ieee80211_vif *vif);
2030 int iwl_mvm_tdls_channel_switch(struct ieee80211_hw *hw,
2031 struct ieee80211_vif *vif,
2032 struct ieee80211_sta *sta, u8 oper_class,
2033 struct cfg80211_chan_def *chandef,
2034 struct sk_buff *tmpl_skb, u32 ch_sw_tm_ie);
2035 void iwl_mvm_tdls_recv_channel_switch(struct ieee80211_hw *hw,
2036 struct ieee80211_vif *vif,
2037 struct ieee80211_tdls_ch_sw_params *params);
2038 void iwl_mvm_tdls_cancel_channel_switch(struct ieee80211_hw *hw,
2039 struct ieee80211_vif *vif,
2040 struct ieee80211_sta *sta);
2041 void iwl_mvm_rx_tdls_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
2042 void iwl_mvm_tdls_ch_switch_work(struct work_struct *work);
2043
2044 void iwl_mvm_sync_rx_queues_internal(struct iwl_mvm *mvm,
2045 enum iwl_mvm_rxq_notif_type type,
2046 bool sync,
2047 const void *data, u32 size);
2048 void iwl_mvm_reorder_timer_expired(struct timer_list *t);
2049 struct ieee80211_vif *iwl_mvm_get_bss_vif(struct iwl_mvm *mvm);
2050 struct ieee80211_vif *iwl_mvm_get_vif_by_macid(struct iwl_mvm *mvm, u32 macid);
2051 bool iwl_mvm_is_vif_assoc(struct iwl_mvm *mvm);
2052
2053 #define MVM_TCM_PERIOD_MSEC 500
2054 #define MVM_TCM_PERIOD (HZ * MVM_TCM_PERIOD_MSEC / 1000)
2055 #define MVM_LL_PERIOD (10 * HZ)
2056 void iwl_mvm_tcm_work(struct work_struct *work);
2057 void iwl_mvm_recalc_tcm(struct iwl_mvm *mvm);
2058 void iwl_mvm_pause_tcm(struct iwl_mvm *mvm, bool with_cancel);
2059 void iwl_mvm_resume_tcm(struct iwl_mvm *mvm);
2060 void iwl_mvm_tcm_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
2061 void iwl_mvm_tcm_rm_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
2062 u8 iwl_mvm_tcm_load_percentage(u32 airtime, u32 elapsed);
2063
2064 void iwl_mvm_nic_restart(struct iwl_mvm *mvm, bool fw_error);
2065 unsigned int iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm,
2066 struct ieee80211_vif *vif,
2067 bool tdls, bool cmd_q);
2068 void iwl_mvm_connection_loss(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2069 const char *errmsg);
2070 void iwl_mvm_event_frame_timeout_callback(struct iwl_mvm *mvm,
2071 struct ieee80211_vif *vif,
2072 const struct ieee80211_sta *sta,
2073 u16 tid);
2074
2075 int iwl_mvm_sar_select_profile(struct iwl_mvm *mvm, int prof_a, int prof_b);
2076 int iwl_mvm_get_sar_geo_profile(struct iwl_mvm *mvm);
2077 int iwl_mvm_ppag_send_cmd(struct iwl_mvm *mvm);
2078 void iwl_mvm_get_acpi_tables(struct iwl_mvm *mvm);
2079 #ifdef CONFIG_IWLWIFI_DEBUGFS
2080 void iwl_mvm_sta_add_debugfs(struct ieee80211_hw *hw,
2081 struct ieee80211_vif *vif,
2082 struct ieee80211_sta *sta,
2083 struct dentry *dir);
2084 #endif
2085
2086 int iwl_rfi_send_config_cmd(struct iwl_mvm *mvm,
2087 struct iwl_rfi_lut_entry *rfi_table);
2088 struct iwl_rfi_freq_table_resp_cmd *iwl_rfi_get_freq_table(struct iwl_mvm *mvm);
2089 void iwl_rfi_deactivate_notif_handler(struct iwl_mvm *mvm,
2090 struct iwl_rx_cmd_buffer *rxb);
2091
iwl_mvm_phy_band_from_nl80211(enum nl80211_band band)2092 static inline u8 iwl_mvm_phy_band_from_nl80211(enum nl80211_band band)
2093 {
2094 switch (band) {
2095 case NL80211_BAND_2GHZ:
2096 return PHY_BAND_24;
2097 case NL80211_BAND_5GHZ:
2098 return PHY_BAND_5;
2099 case NL80211_BAND_6GHZ:
2100 return PHY_BAND_6;
2101 default:
2102 WARN_ONCE(1, "Unsupported band (%u)\n", band);
2103 return PHY_BAND_5;
2104 }
2105 }
2106
2107 /* Channel info utils */
iwl_mvm_has_ultra_hb_channel(struct iwl_mvm * mvm)2108 static inline bool iwl_mvm_has_ultra_hb_channel(struct iwl_mvm *mvm)
2109 {
2110 return fw_has_capa(&mvm->fw->ucode_capa,
2111 IWL_UCODE_TLV_CAPA_ULTRA_HB_CHANNELS);
2112 }
2113
iwl_mvm_chan_info_cmd_tail(struct iwl_mvm * mvm,struct iwl_fw_channel_info * ci)2114 static inline void *iwl_mvm_chan_info_cmd_tail(struct iwl_mvm *mvm,
2115 struct iwl_fw_channel_info *ci)
2116 {
2117 return (u8 *)ci + (iwl_mvm_has_ultra_hb_channel(mvm) ?
2118 sizeof(struct iwl_fw_channel_info) :
2119 sizeof(struct iwl_fw_channel_info_v1));
2120 }
2121
iwl_mvm_chan_info_padding(struct iwl_mvm * mvm)2122 static inline size_t iwl_mvm_chan_info_padding(struct iwl_mvm *mvm)
2123 {
2124 return iwl_mvm_has_ultra_hb_channel(mvm) ? 0 :
2125 sizeof(struct iwl_fw_channel_info) -
2126 sizeof(struct iwl_fw_channel_info_v1);
2127 }
2128
iwl_mvm_set_chan_info(struct iwl_mvm * mvm,struct iwl_fw_channel_info * ci,u32 chan,u8 band,u8 width,u8 ctrl_pos)2129 static inline void iwl_mvm_set_chan_info(struct iwl_mvm *mvm,
2130 struct iwl_fw_channel_info *ci,
2131 u32 chan, u8 band, u8 width,
2132 u8 ctrl_pos)
2133 {
2134 if (iwl_mvm_has_ultra_hb_channel(mvm)) {
2135 ci->channel = cpu_to_le32(chan);
2136 ci->band = band;
2137 ci->width = width;
2138 ci->ctrl_pos = ctrl_pos;
2139 } else {
2140 struct iwl_fw_channel_info_v1 *ci_v1 =
2141 (struct iwl_fw_channel_info_v1 *)ci;
2142
2143 ci_v1->channel = chan;
2144 ci_v1->band = band;
2145 ci_v1->width = width;
2146 ci_v1->ctrl_pos = ctrl_pos;
2147 }
2148 }
2149
2150 static inline void
iwl_mvm_set_chan_info_chandef(struct iwl_mvm * mvm,struct iwl_fw_channel_info * ci,struct cfg80211_chan_def * chandef)2151 iwl_mvm_set_chan_info_chandef(struct iwl_mvm *mvm,
2152 struct iwl_fw_channel_info *ci,
2153 struct cfg80211_chan_def *chandef)
2154 {
2155 enum nl80211_band band = chandef->chan->band;
2156
2157 iwl_mvm_set_chan_info(mvm, ci, chandef->chan->hw_value,
2158 iwl_mvm_phy_band_from_nl80211(band),
2159 iwl_mvm_get_channel_width(chandef),
2160 iwl_mvm_get_ctrl_pos(chandef));
2161 }
2162
iwl_umac_scan_get_max_profiles(const struct iwl_fw * fw)2163 static inline int iwl_umac_scan_get_max_profiles(const struct iwl_fw *fw)
2164 {
2165 u8 ver = iwl_fw_lookup_cmd_ver(fw, SCAN_OFFLOAD_UPDATE_PROFILES_CMD,
2166 IWL_FW_CMD_VER_UNKNOWN);
2167 return (ver == IWL_FW_CMD_VER_UNKNOWN || ver < 3) ?
2168 IWL_SCAN_MAX_PROFILES : IWL_SCAN_MAX_PROFILES_V2;
2169 }
2170
2171 static inline
iwl_mvm_cipher_to_location_cipher(u32 cipher)2172 enum iwl_location_cipher iwl_mvm_cipher_to_location_cipher(u32 cipher)
2173 {
2174 switch (cipher) {
2175 case WLAN_CIPHER_SUITE_CCMP:
2176 return IWL_LOCATION_CIPHER_CCMP_128;
2177 case WLAN_CIPHER_SUITE_GCMP:
2178 return IWL_LOCATION_CIPHER_GCMP_128;
2179 case WLAN_CIPHER_SUITE_GCMP_256:
2180 return IWL_LOCATION_CIPHER_GCMP_256;
2181 default:
2182 return IWL_LOCATION_CIPHER_INVALID;
2183 }
2184 }
2185
2186 struct iwl_mvm_csme_conn_info *iwl_mvm_get_csme_conn_info(struct iwl_mvm *mvm);
iwl_mvm_mei_get_ownership(struct iwl_mvm * mvm)2187 static inline int iwl_mvm_mei_get_ownership(struct iwl_mvm *mvm)
2188 {
2189 if (mvm->mei_registered)
2190 return iwl_mei_get_ownership();
2191 return 0;
2192 }
2193
iwl_mvm_mei_tx_copy_to_csme(struct iwl_mvm * mvm,struct sk_buff * skb,unsigned int ivlen)2194 static inline void iwl_mvm_mei_tx_copy_to_csme(struct iwl_mvm *mvm,
2195 struct sk_buff *skb,
2196 unsigned int ivlen)
2197 {
2198 if (mvm->mei_registered)
2199 iwl_mei_tx_copy_to_csme(skb, ivlen);
2200 }
2201
iwl_mvm_mei_host_disassociated(struct iwl_mvm * mvm)2202 static inline void iwl_mvm_mei_host_disassociated(struct iwl_mvm *mvm)
2203 {
2204 if (mvm->mei_registered)
2205 iwl_mei_host_disassociated();
2206 }
2207
iwl_mvm_mei_device_state(struct iwl_mvm * mvm,bool up)2208 static inline void iwl_mvm_mei_device_state(struct iwl_mvm *mvm, bool up)
2209 {
2210 if (mvm->mei_registered)
2211 iwl_mei_device_state(up);
2212 }
2213
iwl_mvm_mei_set_sw_rfkill_state(struct iwl_mvm * mvm)2214 static inline void iwl_mvm_mei_set_sw_rfkill_state(struct iwl_mvm *mvm)
2215 {
2216 bool sw_rfkill =
2217 mvm->hw_registered ? rfkill_soft_blocked(mvm->hw->wiphy->rfkill) : false;
2218
2219 if (mvm->mei_registered)
2220 iwl_mei_set_rfkill_state(iwl_mvm_is_radio_killed(mvm),
2221 sw_rfkill);
2222 }
2223
2224 void iwl_mvm_send_roaming_forbidden_event(struct iwl_mvm *mvm,
2225 struct ieee80211_vif *vif,
2226 bool forbidden);
2227
2228 #endif /* __IWL_MVM_H__ */
2229