1 /******************************************************************************
2 *
3 * This file is provided under a dual BSD/GPLv2 license. When using or
4 * redistributing this file, you may do so under either license.
5 *
6 * GPL LICENSE SUMMARY
7 *
8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
9 * Copyright(c) 2013 - 2014 Intel Mobile Communications GmbH
10 *
11 * This program is free software; you can redistribute it and/or modify
12 * it under the terms of version 2 of the GNU General Public License as
13 * published by the Free Software Foundation.
14 *
15 * This program is distributed in the hope that it will be useful, but
16 * WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
18 * General Public License for more details.
19 *
20 * You should have received a copy of the GNU General Public License
21 * along with this program; if not, write to the Free Software
22 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
23 * USA
24 *
25 * The full GNU General Public License is included in this distribution
26 * in the file called COPYING.
27 *
28 * Contact Information:
29 * Intel Linux Wireless <ilw@linux.intel.com>
30 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
31 *
32 * BSD LICENSE
33 *
34 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
35 * Copyright(c) 2013 - 2014 Intel Mobile Communications GmbH
36 * All rights reserved.
37 *
38 * Redistribution and use in source and binary forms, with or without
39 * modification, are permitted provided that the following conditions
40 * are met:
41 *
42 * * Redistributions of source code must retain the above copyright
43 * notice, this list of conditions and the following disclaimer.
44 * * Redistributions in binary form must reproduce the above copyright
45 * notice, this list of conditions and the following disclaimer in
46 * the documentation and/or other materials provided with the
47 * distribution.
48 * * Neither the name Intel Corporation nor the names of its
49 * contributors may be used to endorse or promote products derived
50 * from this software without specific prior written permission.
51 *
52 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
53 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
54 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
55 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
56 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
57 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
58 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
59 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
60 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
61 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
62 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
63 *
64 *****************************************************************************/
65 #include <net/mac80211.h>
66
67 #include "iwl-debug.h"
68 #include "iwl-io.h"
69 #include "iwl-prph.h"
70
71 #include "mvm.h"
72 #include "fw-api-rs.h"
73
74 /*
75 * Will return 0 even if the cmd failed when RFKILL is asserted unless
76 * CMD_WANT_SKB is set in cmd->flags.
77 */
iwl_mvm_send_cmd(struct iwl_mvm * mvm,struct iwl_host_cmd * cmd)78 int iwl_mvm_send_cmd(struct iwl_mvm *mvm, struct iwl_host_cmd *cmd)
79 {
80 int ret;
81
82 #if defined(CONFIG_IWLWIFI_DEBUGFS) && defined(CONFIG_PM_SLEEP)
83 if (WARN_ON(mvm->d3_test_active))
84 return -EIO;
85 #endif
86
87 /*
88 * Synchronous commands from this op-mode must hold
89 * the mutex, this ensures we don't try to send two
90 * (or more) synchronous commands at a time.
91 */
92 if (!(cmd->flags & CMD_ASYNC))
93 lockdep_assert_held(&mvm->mutex);
94
95 ret = iwl_trans_send_cmd(mvm->trans, cmd);
96
97 /*
98 * If the caller wants the SKB, then don't hide any problems, the
99 * caller might access the response buffer which will be NULL if
100 * the command failed.
101 */
102 if (cmd->flags & CMD_WANT_SKB)
103 return ret;
104
105 /* Silently ignore failures if RFKILL is asserted */
106 if (!ret || ret == -ERFKILL)
107 return 0;
108 return ret;
109 }
110
iwl_mvm_send_cmd_pdu(struct iwl_mvm * mvm,u8 id,u32 flags,u16 len,const void * data)111 int iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u8 id,
112 u32 flags, u16 len, const void *data)
113 {
114 struct iwl_host_cmd cmd = {
115 .id = id,
116 .len = { len, },
117 .data = { data, },
118 .flags = flags,
119 };
120
121 return iwl_mvm_send_cmd(mvm, &cmd);
122 }
123
124 /*
125 * We assume that the caller set the status to the sucess value
126 */
iwl_mvm_send_cmd_status(struct iwl_mvm * mvm,struct iwl_host_cmd * cmd,u32 * status)127 int iwl_mvm_send_cmd_status(struct iwl_mvm *mvm, struct iwl_host_cmd *cmd,
128 u32 *status)
129 {
130 struct iwl_rx_packet *pkt;
131 struct iwl_cmd_response *resp;
132 int ret, resp_len;
133
134 lockdep_assert_held(&mvm->mutex);
135
136 #if defined(CONFIG_IWLWIFI_DEBUGFS) && defined(CONFIG_PM_SLEEP)
137 if (WARN_ON(mvm->d3_test_active))
138 return -EIO;
139 #endif
140
141 /*
142 * Only synchronous commands can wait for status,
143 * we use WANT_SKB so the caller can't.
144 */
145 if (WARN_ONCE(cmd->flags & (CMD_ASYNC | CMD_WANT_SKB),
146 "cmd flags %x", cmd->flags))
147 return -EINVAL;
148
149 cmd->flags |= CMD_WANT_SKB;
150
151 ret = iwl_trans_send_cmd(mvm->trans, cmd);
152 if (ret == -ERFKILL) {
153 /*
154 * The command failed because of RFKILL, don't update
155 * the status, leave it as success and return 0.
156 */
157 return 0;
158 } else if (ret) {
159 return ret;
160 }
161
162 pkt = cmd->resp_pkt;
163 /* Can happen if RFKILL is asserted */
164 if (!pkt) {
165 ret = 0;
166 goto out_free_resp;
167 }
168
169 if (pkt->hdr.flags & IWL_CMD_FAILED_MSK) {
170 ret = -EIO;
171 goto out_free_resp;
172 }
173
174 resp_len = iwl_rx_packet_payload_len(pkt);
175 if (WARN_ON_ONCE(resp_len != sizeof(*resp))) {
176 ret = -EIO;
177 goto out_free_resp;
178 }
179
180 resp = (void *)pkt->data;
181 *status = le32_to_cpu(resp->status);
182 out_free_resp:
183 iwl_free_resp(cmd);
184 return ret;
185 }
186
187 /*
188 * We assume that the caller set the status to the sucess value
189 */
iwl_mvm_send_cmd_pdu_status(struct iwl_mvm * mvm,u8 id,u16 len,const void * data,u32 * status)190 int iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u8 id, u16 len,
191 const void *data, u32 *status)
192 {
193 struct iwl_host_cmd cmd = {
194 .id = id,
195 .len = { len, },
196 .data = { data, },
197 };
198
199 return iwl_mvm_send_cmd_status(mvm, &cmd, status);
200 }
201
202 #define IWL_DECLARE_RATE_INFO(r) \
203 [IWL_RATE_##r##M_INDEX] = IWL_RATE_##r##M_PLCP
204
205 /*
206 * Translate from fw_rate_index (IWL_RATE_XXM_INDEX) to PLCP
207 */
208 static const u8 fw_rate_idx_to_plcp[IWL_RATE_COUNT] = {
209 IWL_DECLARE_RATE_INFO(1),
210 IWL_DECLARE_RATE_INFO(2),
211 IWL_DECLARE_RATE_INFO(5),
212 IWL_DECLARE_RATE_INFO(11),
213 IWL_DECLARE_RATE_INFO(6),
214 IWL_DECLARE_RATE_INFO(9),
215 IWL_DECLARE_RATE_INFO(12),
216 IWL_DECLARE_RATE_INFO(18),
217 IWL_DECLARE_RATE_INFO(24),
218 IWL_DECLARE_RATE_INFO(36),
219 IWL_DECLARE_RATE_INFO(48),
220 IWL_DECLARE_RATE_INFO(54),
221 };
222
iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags,enum ieee80211_band band)223 int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags,
224 enum ieee80211_band band)
225 {
226 int rate = rate_n_flags & RATE_LEGACY_RATE_MSK;
227 int idx;
228 int band_offset = 0;
229
230 /* Legacy rate format, search for match in table */
231 if (band == IEEE80211_BAND_5GHZ)
232 band_offset = IWL_FIRST_OFDM_RATE;
233 for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++)
234 if (fw_rate_idx_to_plcp[idx] == rate)
235 return idx - band_offset;
236
237 return -1;
238 }
239
iwl_mvm_mac80211_idx_to_hwrate(int rate_idx)240 u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx)
241 {
242 /* Get PLCP rate for tx_cmd->rate_n_flags */
243 return fw_rate_idx_to_plcp[rate_idx];
244 }
245
iwl_mvm_rx_fw_error(struct iwl_mvm * mvm,struct iwl_rx_cmd_buffer * rxb,struct iwl_device_cmd * cmd)246 int iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb,
247 struct iwl_device_cmd *cmd)
248 {
249 struct iwl_rx_packet *pkt = rxb_addr(rxb);
250 struct iwl_error_resp *err_resp = (void *)pkt->data;
251
252 IWL_ERR(mvm, "FW Error notification: type 0x%08X cmd_id 0x%02X\n",
253 le32_to_cpu(err_resp->error_type), err_resp->cmd_id);
254 IWL_ERR(mvm, "FW Error notification: seq 0x%04X service 0x%08X\n",
255 le16_to_cpu(err_resp->bad_cmd_seq_num),
256 le32_to_cpu(err_resp->error_service));
257 IWL_ERR(mvm, "FW Error notification: timestamp 0x%16llX\n",
258 le64_to_cpu(err_resp->timestamp));
259 return 0;
260 }
261
262 /*
263 * Returns the first antenna as ANT_[ABC], as defined in iwl-config.h.
264 * The parameter should also be a combination of ANT_[ABC].
265 */
first_antenna(u8 mask)266 u8 first_antenna(u8 mask)
267 {
268 BUILD_BUG_ON(ANT_A != BIT(0)); /* using ffs is wrong if not */
269 if (WARN_ON_ONCE(!mask)) /* ffs will return 0 if mask is zeroed */
270 return BIT(0);
271 return BIT(ffs(mask) - 1);
272 }
273
274 /*
275 * Toggles between TX antennas to send the probe request on.
276 * Receives the bitmask of valid TX antennas and the *index* used
277 * for the last TX, and returns the next valid *index* to use.
278 * In order to set it in the tx_cmd, must do BIT(idx).
279 */
iwl_mvm_next_antenna(struct iwl_mvm * mvm,u8 valid,u8 last_idx)280 u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx)
281 {
282 u8 ind = last_idx;
283 int i;
284
285 for (i = 0; i < RATE_MCS_ANT_NUM; i++) {
286 ind = (ind + 1) % RATE_MCS_ANT_NUM;
287 if (valid & BIT(ind))
288 return ind;
289 }
290
291 WARN_ONCE(1, "Failed to toggle between antennas 0x%x", valid);
292 return last_idx;
293 }
294
295 static const struct {
296 const char *name;
297 u8 num;
298 } advanced_lookup[] = {
299 { "NMI_INTERRUPT_WDG", 0x34 },
300 { "SYSASSERT", 0x35 },
301 { "UCODE_VERSION_MISMATCH", 0x37 },
302 { "BAD_COMMAND", 0x38 },
303 { "NMI_INTERRUPT_DATA_ACTION_PT", 0x3C },
304 { "FATAL_ERROR", 0x3D },
305 { "NMI_TRM_HW_ERR", 0x46 },
306 { "NMI_INTERRUPT_TRM", 0x4C },
307 { "NMI_INTERRUPT_BREAK_POINT", 0x54 },
308 { "NMI_INTERRUPT_WDG_RXF_FULL", 0x5C },
309 { "NMI_INTERRUPT_WDG_NO_RBD_RXF_FULL", 0x64 },
310 { "NMI_INTERRUPT_HOST", 0x66 },
311 { "NMI_INTERRUPT_ACTION_PT", 0x7C },
312 { "NMI_INTERRUPT_UNKNOWN", 0x84 },
313 { "NMI_INTERRUPT_INST_ACTION_PT", 0x86 },
314 { "ADVANCED_SYSASSERT", 0 },
315 };
316
desc_lookup(u32 num)317 static const char *desc_lookup(u32 num)
318 {
319 int i;
320
321 for (i = 0; i < ARRAY_SIZE(advanced_lookup) - 1; i++)
322 if (advanced_lookup[i].num == num)
323 return advanced_lookup[i].name;
324
325 /* No entry matches 'num', so it is the last: ADVANCED_SYSASSERT */
326 return advanced_lookup[i].name;
327 }
328
329 /*
330 * Note: This structure is read from the device with IO accesses,
331 * and the reading already does the endian conversion. As it is
332 * read with u32-sized accesses, any members with a different size
333 * need to be ordered correctly though!
334 */
335 struct iwl_error_event_table {
336 u32 valid; /* (nonzero) valid, (0) log is empty */
337 u32 error_id; /* type of error */
338 u32 pc; /* program counter */
339 u32 blink1; /* branch link */
340 u32 blink2; /* branch link */
341 u32 ilink1; /* interrupt link */
342 u32 ilink2; /* interrupt link */
343 u32 data1; /* error-specific data */
344 u32 data2; /* error-specific data */
345 u32 data3; /* error-specific data */
346 u32 bcon_time; /* beacon timer */
347 u32 tsf_low; /* network timestamp function timer */
348 u32 tsf_hi; /* network timestamp function timer */
349 u32 gp1; /* GP1 timer register */
350 u32 gp2; /* GP2 timer register */
351 u32 gp3; /* GP3 timer register */
352 u32 ucode_ver; /* uCode version */
353 u32 hw_ver; /* HW Silicon version */
354 u32 brd_ver; /* HW board version */
355 u32 log_pc; /* log program counter */
356 u32 frame_ptr; /* frame pointer */
357 u32 stack_ptr; /* stack pointer */
358 u32 hcmd; /* last host command header */
359 u32 isr0; /* isr status register LMPM_NIC_ISR0:
360 * rxtx_flag */
361 u32 isr1; /* isr status register LMPM_NIC_ISR1:
362 * host_flag */
363 u32 isr2; /* isr status register LMPM_NIC_ISR2:
364 * enc_flag */
365 u32 isr3; /* isr status register LMPM_NIC_ISR3:
366 * time_flag */
367 u32 isr4; /* isr status register LMPM_NIC_ISR4:
368 * wico interrupt */
369 u32 isr_pref; /* isr status register LMPM_NIC_PREF_STAT */
370 u32 wait_event; /* wait event() caller address */
371 u32 l2p_control; /* L2pControlField */
372 u32 l2p_duration; /* L2pDurationField */
373 u32 l2p_mhvalid; /* L2pMhValidBits */
374 u32 l2p_addr_match; /* L2pAddrMatchStat */
375 u32 lmpm_pmg_sel; /* indicate which clocks are turned on
376 * (LMPM_PMG_SEL) */
377 u32 u_timestamp; /* indicate when the date and time of the
378 * compilation */
379 u32 flow_handler; /* FH read/write pointers, RX credit */
380 } __packed;
381
382 /*
383 * UMAC error struct - relevant starting from family 8000 chip.
384 * Note: This structure is read from the device with IO accesses,
385 * and the reading already does the endian conversion. As it is
386 * read with u32-sized accesses, any members with a different size
387 * need to be ordered correctly though!
388 */
389 struct iwl_umac_error_event_table {
390 u32 valid; /* (nonzero) valid, (0) log is empty */
391 u32 error_id; /* type of error */
392 u32 blink1; /* branch link */
393 u32 blink2; /* branch link */
394 u32 ilink1; /* interrupt link */
395 u32 ilink2; /* interrupt link */
396 u32 data1; /* error-specific data */
397 u32 data2; /* error-specific data */
398 u32 data3; /* error-specific data */
399 u32 umac_fw_ver; /* UMAC version */
400 u32 umac_fw_api_ver; /* UMAC FW API ver */
401 u32 frame_pointer; /* core register 27*/
402 u32 stack_pointer; /* core register 28 */
403 u32 cmd_header; /* latest host cmd sent to UMAC */
404 u32 nic_isr_pref; /* ISR status register */
405 } __packed;
406
407 #define ERROR_START_OFFSET (1 * sizeof(u32))
408 #define ERROR_ELEM_SIZE (7 * sizeof(u32))
409
iwl_mvm_dump_umac_error_log(struct iwl_mvm * mvm)410 static void iwl_mvm_dump_umac_error_log(struct iwl_mvm *mvm)
411 {
412 struct iwl_trans *trans = mvm->trans;
413 struct iwl_umac_error_event_table table;
414 u32 base;
415
416 base = mvm->umac_error_event_table;
417
418 if (base < 0x800000) {
419 IWL_ERR(mvm,
420 "Not valid error log pointer 0x%08X for %s uCode\n",
421 base,
422 (mvm->cur_ucode == IWL_UCODE_INIT)
423 ? "Init" : "RT");
424 return;
425 }
426
427 iwl_trans_read_mem_bytes(trans, base, &table, sizeof(table));
428
429 if (ERROR_START_OFFSET <= table.valid * ERROR_ELEM_SIZE) {
430 IWL_ERR(trans, "Start IWL Error Log Dump:\n");
431 IWL_ERR(trans, "Status: 0x%08lX, count: %d\n",
432 mvm->status, table.valid);
433 }
434
435 IWL_ERR(mvm, "0x%08X | %-28s\n", table.error_id,
436 desc_lookup(table.error_id));
437 IWL_ERR(mvm, "0x%08X | umac branchlink1\n", table.blink1);
438 IWL_ERR(mvm, "0x%08X | umac branchlink2\n", table.blink2);
439 IWL_ERR(mvm, "0x%08X | umac interruptlink1\n", table.ilink1);
440 IWL_ERR(mvm, "0x%08X | umac interruptlink2\n", table.ilink2);
441 IWL_ERR(mvm, "0x%08X | umac data1\n", table.data1);
442 IWL_ERR(mvm, "0x%08X | umac data2\n", table.data2);
443 IWL_ERR(mvm, "0x%08X | umac data3\n", table.data3);
444 IWL_ERR(mvm, "0x%08X | umac version\n", table.umac_fw_ver);
445 IWL_ERR(mvm, "0x%08X | umac api version\n", table.umac_fw_api_ver);
446 IWL_ERR(mvm, "0x%08X | frame pointer\n", table.frame_pointer);
447 IWL_ERR(mvm, "0x%08X | stack pointer\n", table.stack_pointer);
448 IWL_ERR(mvm, "0x%08X | last host cmd\n", table.cmd_header);
449 IWL_ERR(mvm, "0x%08X | isr status reg\n", table.nic_isr_pref);
450 }
451
iwl_mvm_dump_nic_error_log(struct iwl_mvm * mvm)452 void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm)
453 {
454 struct iwl_trans *trans = mvm->trans;
455 struct iwl_error_event_table table;
456 u32 base;
457
458 base = mvm->error_event_table;
459 if (mvm->cur_ucode == IWL_UCODE_INIT) {
460 if (!base)
461 base = mvm->fw->init_errlog_ptr;
462 } else {
463 if (!base)
464 base = mvm->fw->inst_errlog_ptr;
465 }
466
467 if (base < 0x800000) {
468 IWL_ERR(mvm,
469 "Not valid error log pointer 0x%08X for %s uCode\n",
470 base,
471 (mvm->cur_ucode == IWL_UCODE_INIT)
472 ? "Init" : "RT");
473 return;
474 }
475
476 iwl_trans_read_mem_bytes(trans, base, &table, sizeof(table));
477
478 if (ERROR_START_OFFSET <= table.valid * ERROR_ELEM_SIZE) {
479 IWL_ERR(trans, "Start IWL Error Log Dump:\n");
480 IWL_ERR(trans, "Status: 0x%08lX, count: %d\n",
481 mvm->status, table.valid);
482 }
483
484 /* Do not change this output - scripts rely on it */
485
486 IWL_ERR(mvm, "Loaded firmware version: %s\n", mvm->fw->fw_version);
487
488 trace_iwlwifi_dev_ucode_error(trans->dev, table.error_id, table.tsf_low,
489 table.data1, table.data2, table.data3,
490 table.blink1, table.blink2, table.ilink1,
491 table.ilink2, table.bcon_time, table.gp1,
492 table.gp2, table.gp3, table.ucode_ver,
493 table.hw_ver, table.brd_ver);
494 IWL_ERR(mvm, "0x%08X | %-28s\n", table.error_id,
495 desc_lookup(table.error_id));
496 IWL_ERR(mvm, "0x%08X | uPc\n", table.pc);
497 IWL_ERR(mvm, "0x%08X | branchlink1\n", table.blink1);
498 IWL_ERR(mvm, "0x%08X | branchlink2\n", table.blink2);
499 IWL_ERR(mvm, "0x%08X | interruptlink1\n", table.ilink1);
500 IWL_ERR(mvm, "0x%08X | interruptlink2\n", table.ilink2);
501 IWL_ERR(mvm, "0x%08X | data1\n", table.data1);
502 IWL_ERR(mvm, "0x%08X | data2\n", table.data2);
503 IWL_ERR(mvm, "0x%08X | data3\n", table.data3);
504 IWL_ERR(mvm, "0x%08X | beacon time\n", table.bcon_time);
505 IWL_ERR(mvm, "0x%08X | tsf low\n", table.tsf_low);
506 IWL_ERR(mvm, "0x%08X | tsf hi\n", table.tsf_hi);
507 IWL_ERR(mvm, "0x%08X | time gp1\n", table.gp1);
508 IWL_ERR(mvm, "0x%08X | time gp2\n", table.gp2);
509 IWL_ERR(mvm, "0x%08X | time gp3\n", table.gp3);
510 IWL_ERR(mvm, "0x%08X | uCode version\n", table.ucode_ver);
511 IWL_ERR(mvm, "0x%08X | hw version\n", table.hw_ver);
512 IWL_ERR(mvm, "0x%08X | board version\n", table.brd_ver);
513 IWL_ERR(mvm, "0x%08X | hcmd\n", table.hcmd);
514 IWL_ERR(mvm, "0x%08X | isr0\n", table.isr0);
515 IWL_ERR(mvm, "0x%08X | isr1\n", table.isr1);
516 IWL_ERR(mvm, "0x%08X | isr2\n", table.isr2);
517 IWL_ERR(mvm, "0x%08X | isr3\n", table.isr3);
518 IWL_ERR(mvm, "0x%08X | isr4\n", table.isr4);
519 IWL_ERR(mvm, "0x%08X | isr_pref\n", table.isr_pref);
520 IWL_ERR(mvm, "0x%08X | wait_event\n", table.wait_event);
521 IWL_ERR(mvm, "0x%08X | l2p_control\n", table.l2p_control);
522 IWL_ERR(mvm, "0x%08X | l2p_duration\n", table.l2p_duration);
523 IWL_ERR(mvm, "0x%08X | l2p_mhvalid\n", table.l2p_mhvalid);
524 IWL_ERR(mvm, "0x%08X | l2p_addr_match\n", table.l2p_addr_match);
525 IWL_ERR(mvm, "0x%08X | lmpm_pmg_sel\n", table.lmpm_pmg_sel);
526 IWL_ERR(mvm, "0x%08X | timestamp\n", table.u_timestamp);
527 IWL_ERR(mvm, "0x%08X | flow_handler\n", table.flow_handler);
528
529 if (mvm->support_umac_log)
530 iwl_mvm_dump_umac_error_log(mvm);
531 }
532
iwl_mvm_enable_txq(struct iwl_mvm * mvm,int queue,u16 ssn,const struct iwl_trans_txq_scd_cfg * cfg)533 void iwl_mvm_enable_txq(struct iwl_mvm *mvm, int queue, u16 ssn,
534 const struct iwl_trans_txq_scd_cfg *cfg)
535 {
536 if (iwl_mvm_is_dqa_supported(mvm)) {
537 struct iwl_scd_txq_cfg_cmd cmd = {
538 .scd_queue = queue,
539 .enable = 1,
540 .window = cfg->frame_limit,
541 .sta_id = cfg->sta_id,
542 .ssn = cpu_to_le16(ssn),
543 .tx_fifo = cfg->fifo,
544 .aggregate = cfg->aggregate,
545 .flags = IWL_SCD_FLAGS_DQA_ENABLED,
546 .tid = cfg->tid,
547 .control = IWL_SCD_CONTROL_SET_SSN,
548 };
549 int ret = iwl_mvm_send_cmd_pdu(mvm, SCD_QUEUE_CFG, 0,
550 sizeof(cmd), &cmd);
551 if (ret)
552 IWL_ERR(mvm,
553 "Failed to configure queue %d on FIFO %d\n",
554 queue, cfg->fifo);
555 }
556
557 iwl_trans_txq_enable_cfg(mvm->trans, queue, ssn,
558 iwl_mvm_is_dqa_supported(mvm) ? NULL : cfg);
559 }
560
iwl_mvm_disable_txq(struct iwl_mvm * mvm,int queue)561 void iwl_mvm_disable_txq(struct iwl_mvm *mvm, int queue)
562 {
563 iwl_trans_txq_disable(mvm->trans, queue,
564 !iwl_mvm_is_dqa_supported(mvm));
565
566 if (iwl_mvm_is_dqa_supported(mvm)) {
567 struct iwl_scd_txq_cfg_cmd cmd = {
568 .scd_queue = queue,
569 .enable = 0,
570 };
571 int ret = iwl_mvm_send_cmd_pdu(mvm, SCD_QUEUE_CFG, CMD_ASYNC,
572 sizeof(cmd), &cmd);
573 if (ret)
574 IWL_ERR(mvm, "Failed to disable queue %d (ret=%d)\n",
575 queue, ret);
576 }
577 }
578
579 /**
580 * iwl_mvm_send_lq_cmd() - Send link quality command
581 * @init: This command is sent as part of station initialization right
582 * after station has been added.
583 *
584 * The link quality command is sent as the last step of station creation.
585 * This is the special case in which init is set and we call a callback in
586 * this case to clear the state indicating that station creation is in
587 * progress.
588 */
iwl_mvm_send_lq_cmd(struct iwl_mvm * mvm,struct iwl_lq_cmd * lq,bool init)589 int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq, bool init)
590 {
591 struct iwl_host_cmd cmd = {
592 .id = LQ_CMD,
593 .len = { sizeof(struct iwl_lq_cmd), },
594 .flags = init ? 0 : CMD_ASYNC,
595 .data = { lq, },
596 };
597
598 if (WARN_ON(lq->sta_id == IWL_MVM_STATION_COUNT))
599 return -EINVAL;
600
601 return iwl_mvm_send_cmd(mvm, &cmd);
602 }
603
604 /**
605 * iwl_mvm_update_smps - Get a requst to change the SMPS mode
606 * @req_type: The part of the driver who call for a change.
607 * @smps_requests: The request to change the SMPS mode.
608 *
609 * Get a requst to change the SMPS mode,
610 * and change it according to all other requests in the driver.
611 */
iwl_mvm_update_smps(struct iwl_mvm * mvm,struct ieee80211_vif * vif,enum iwl_mvm_smps_type_request req_type,enum ieee80211_smps_mode smps_request)612 void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
613 enum iwl_mvm_smps_type_request req_type,
614 enum ieee80211_smps_mode smps_request)
615 {
616 struct iwl_mvm_vif *mvmvif;
617 enum ieee80211_smps_mode smps_mode;
618 int i;
619
620 lockdep_assert_held(&mvm->mutex);
621
622 /* SMPS is irrelevant for NICs that don't have at least 2 RX antenna */
623 if (num_of_ant(mvm->fw->valid_rx_ant) == 1)
624 return;
625
626 if (vif->type == NL80211_IFTYPE_AP)
627 smps_mode = IEEE80211_SMPS_OFF;
628 else
629 smps_mode = IEEE80211_SMPS_AUTOMATIC;
630
631 mvmvif = iwl_mvm_vif_from_mac80211(vif);
632 mvmvif->smps_requests[req_type] = smps_request;
633 for (i = 0; i < NUM_IWL_MVM_SMPS_REQ; i++) {
634 if (mvmvif->smps_requests[i] == IEEE80211_SMPS_STATIC) {
635 smps_mode = IEEE80211_SMPS_STATIC;
636 break;
637 }
638 if (mvmvif->smps_requests[i] == IEEE80211_SMPS_DYNAMIC)
639 smps_mode = IEEE80211_SMPS_DYNAMIC;
640 }
641
642 ieee80211_request_smps(vif, smps_mode);
643 }
644
iwl_mvm_diversity_iter(void * _data,u8 * mac,struct ieee80211_vif * vif)645 static void iwl_mvm_diversity_iter(void *_data, u8 *mac,
646 struct ieee80211_vif *vif)
647 {
648 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
649 bool *result = _data;
650 int i;
651
652 for (i = 0; i < NUM_IWL_MVM_SMPS_REQ; i++) {
653 if (mvmvif->smps_requests[i] == IEEE80211_SMPS_STATIC ||
654 mvmvif->smps_requests[i] == IEEE80211_SMPS_DYNAMIC)
655 *result = false;
656 }
657 }
658
iwl_mvm_rx_diversity_allowed(struct iwl_mvm * mvm)659 bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm)
660 {
661 bool result = true;
662
663 lockdep_assert_held(&mvm->mutex);
664
665 if (num_of_ant(mvm->fw->valid_rx_ant) == 1)
666 return false;
667
668 if (mvm->cfg->rx_with_siso_diversity)
669 return false;
670
671 ieee80211_iterate_active_interfaces_atomic(
672 mvm->hw, IEEE80211_IFACE_ITER_NORMAL,
673 iwl_mvm_diversity_iter, &result);
674
675 return result;
676 }
677
iwl_mvm_update_low_latency(struct iwl_mvm * mvm,struct ieee80211_vif * vif,bool value)678 int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
679 bool value)
680 {
681 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
682 int res;
683
684 lockdep_assert_held(&mvm->mutex);
685
686 if (mvmvif->low_latency == value)
687 return 0;
688
689 mvmvif->low_latency = value;
690
691 res = iwl_mvm_update_quotas(mvm, NULL);
692 if (res)
693 return res;
694
695 iwl_mvm_bt_coex_vif_change(mvm);
696
697 return iwl_mvm_power_update_mac(mvm);
698 }
699
iwl_mvm_ll_iter(void * _data,u8 * mac,struct ieee80211_vif * vif)700 static void iwl_mvm_ll_iter(void *_data, u8 *mac, struct ieee80211_vif *vif)
701 {
702 bool *result = _data;
703
704 if (iwl_mvm_vif_low_latency(iwl_mvm_vif_from_mac80211(vif)))
705 *result = true;
706 }
707
iwl_mvm_low_latency(struct iwl_mvm * mvm)708 bool iwl_mvm_low_latency(struct iwl_mvm *mvm)
709 {
710 bool result = false;
711
712 ieee80211_iterate_active_interfaces_atomic(
713 mvm->hw, IEEE80211_IFACE_ITER_NORMAL,
714 iwl_mvm_ll_iter, &result);
715
716 return result;
717 }
718
iwl_mvm_idle_iter(void * _data,u8 * mac,struct ieee80211_vif * vif)719 static void iwl_mvm_idle_iter(void *_data, u8 *mac, struct ieee80211_vif *vif)
720 {
721 bool *idle = _data;
722
723 if (!vif->bss_conf.idle)
724 *idle = false;
725 }
726
iwl_mvm_is_idle(struct iwl_mvm * mvm)727 bool iwl_mvm_is_idle(struct iwl_mvm *mvm)
728 {
729 bool idle = true;
730
731 ieee80211_iterate_active_interfaces_atomic(
732 mvm->hw, IEEE80211_IFACE_ITER_NORMAL,
733 iwl_mvm_idle_iter, &idle);
734
735 return idle;
736 }
737