• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Broadcom Dongle Host Driver (DHD), common DHD core.
3  *
4  * Copyright (C) 2020, Broadcom.
5  *
6  *      Unless you and Broadcom execute a separate written software license
7  * agreement governing use of this software, this software is licensed to you
8  * under the terms of the GNU General Public License version 2 (the "GPL"),
9  * available at http://www.broadcom.com/licenses/GPLv2.php, with the
10  * following added to such license:
11  *
12  *      As a special exception, the copyright holders of this software give you
13  * permission to link this software with independent modules, and to copy and
14  * distribute the resulting executable under terms of your choice, provided that
15  * you also meet, for each linked independent module, the terms and conditions of
16  * the license of that module.  An independent module is a module which is not
17  * derived from this software.  The special exception does not apply to any
18  * modifications of the software.
19  *
20  *
21  * <<Broadcom-WL-IPTag/Open:>>
22  *
23  * $Id$
24  */
25 #include <typedefs.h>
26 #include <osl.h>
27 
28 #include <epivers.h>
29 #include <bcmutils.h>
30 #include <bcmstdlib_s.h>
31 
32 #include <bcmendian.h>
33 #include <dngl_stats.h>
34 #include <dhd.h>
35 #include <dhd_ip.h>
36 #include <bcmevent.h>
37 #include <dhdioctl.h>
38 #ifdef DHD_SDTC_ETB_DUMP
39 #include <bcmiov.h>
40 #endif /* DHD_SDTC_ETB_DUMP */
41 
42 #ifdef BCMDBG
43 #include <dhd_macdbg.h>
44 #endif /* BCMDBG */
45 
46 #ifdef PCIE_FULL_DONGLE
47 #include <bcmmsgbuf.h>
48 #endif /* PCIE_FULL_DONGLE */
49 
50 #ifdef SHOW_LOGTRACE
51 #include <event_log.h>
52 #endif /* SHOW_LOGTRACE */
53 
54 #ifdef BCMPCIE
55 #include <dhd_flowring.h>
56 #endif
57 
58 #include <dhd_bus.h>
59 #include <dhd_proto.h>
60 #include <bcmsdbus.h>
61 #include <dhd_dbg.h>
62 #include <802.1d.h>
63 #include <dhd_debug.h>
64 #include <dhd_dbg_ring.h>
65 #include <dhd_mschdbg.h>
66 #include <msgtrace.h>
67 #include <dhd_config.h>
68 #include <wl_android.h>
69 
70 #ifdef WL_CFG80211
71 #include <wl_cfg80211.h>
72 #include <wl_cfgvif.h>
73 #endif
74 #if defined(OEM_ANDROID) && defined(PNO_SUPPORT)
75 #include <dhd_pno.h>
76 #endif /* (OEM_ANDROID) && (PNO_SUPPORT) */
77 #ifdef RTT_SUPPORT
78 #include <dhd_rtt.h>
79 #endif
80 
81 #ifdef DNGL_EVENT_SUPPORT
82 #include <dnglevent.h>
83 #endif
84 
85 #ifdef IL_BIGENDIAN
86 #include <bcmendian.h>
87 #define htod32(i) (bcmswap32(i))
88 #define htod16(i) (bcmswap16(i))
89 #define dtoh32(i) (bcmswap32(i))
90 #define dtoh16(i) (bcmswap16(i))
91 #define htodchanspec(i) htod16(i)
92 #define dtohchanspec(i) dtoh16(i)
93 #else
94 #define htod32(i) (i)
95 #define htod16(i) (i)
96 #define dtoh32(i) (i)
97 #define dtoh16(i) (i)
98 #define htodchanspec(i) (i)
99 #define dtohchanspec(i) (i)
100 #endif /* IL_BIGENDINA */
101 
102 #ifdef PROP_TXSTATUS
103 #include <wlfc_proto.h>
104 #include <dhd_wlfc.h>
105 #endif
106 
107 #if defined(__linux__)
108 #include <dhd_linux.h>
109 #endif /* __linux__ */
110 
111 #ifdef DHD_WMF
112 #include <dhd_wmf_linux.h>
113 #endif /* DHD_WMF */
114 
115 #ifdef DHD_L2_FILTER
116 #include <dhd_l2_filter.h>
117 #endif /* DHD_L2_FILTER */
118 
119 #ifdef DHD_PSTA
120 #include <dhd_psta.h>
121 #endif /* DHD_PSTA */
122 #ifdef DHD_TIMESYNC
123 #include <dhd_timesync.h>
124 #endif /* DHD_TIMESYNC */
125 
126 #ifdef DHD_WET
127 #include <dhd_wet.h>
128 #endif /* DHD_WET */
129 #if defined(NDIS)
130 #include <siutils.h>
131 #endif
132 
133 #ifdef DHD_LOG_DUMP
134 #include <dhd_dbg.h>
135 #ifdef DHD_PKT_LOGGING
136 #include <dhd_pktlog.h>
137 #endif
138 #endif /* DHD_LOG_DUMP */
139 
140 #ifdef DHD_LOG_PRINT_RATE_LIMIT
141 int log_print_threshold = 0;
142 #endif /* DHD_LOG_PRINT_RATE_LIMIT */
143 
144 #ifdef DHD_DEBUGABILITY_LOG_DUMP_RING
145 int dbgring_msg_level = DHD_ERROR_VAL | DHD_FWLOG_VAL | DHD_INFO_VAL
146 		| DHD_EVENT_VAL | DHD_PKT_MON_VAL | DHD_IOVAR_MEM_VAL;
147 int dhd_msg_level = DHD_ERROR_VAL;
148 #else
149 int dbgring_msg_level = 0;
150 /* For CUSTOMER_HW4/Hikey do not enable DHD_ERROR_MEM_VAL by default */
151 int dhd_msg_level = DHD_ERROR_VAL | DHD_FWLOG_VAL;
152 #endif /* DHD_DEBUGABILITY_LOG_DUMP_RING */
153 
154 #ifdef NDIS
155 extern uint wl_msg_level;
156 #endif
157 
158 #if defined(WL_WLC_SHIM)
159 #include <wl_shim.h>
160 #else
161 #if defined(NDIS)
162 #include <wl_port_if.h>
163 #endif
164 #endif /* WL_WLC_SHIM */
165 
166 #ifdef DHD_DEBUG
167 #include <sdiovar.h>
168 #endif /* DHD_DEBUG */
169 
170 #ifdef DHD_PCIE_NATIVE_RUNTIMEPM
171 #include <linux/pm_runtime.h>
172 #endif /* DHD_PCIE_NATIVE_RUNTIMEPM */
173 
174 #ifdef CSI_SUPPORT
175 #include <dhd_csi.h>
176 #endif /* CSI_SUPPORT */
177 
178 #if defined(BTLOG) && !defined(BCMPCIE)
179 #error "BT logging supported only with PCIe"
180 #endif  /* defined(BTLOG) && !defined(BCMPCIE) */
181 
182 #ifdef SOFTAP
183 char fw_path2[MOD_PARAM_PATHLEN];
184 extern bool softap_enabled;
185 #endif
186 
187 #ifdef REPORT_FATAL_TIMEOUTS
188 #ifdef BCMINTERNAL
189 /*
190  * Internal Builds are used by DVT.
191  * The timeouts are not required for DVT builds, since they use IOVARs like
192  * SROM programming etc, that takes long time. So make the timeout values
193  * as 0. If DVT needs to use this feature they can enable them using IOVAR
194  *
195  * SVT any way uses external builds
196  */
197 #define SCAN_TIMEOUT_DEFAULT	0
198 #define JOIN_TIMEOUT_DEFAULT	0
199 #define BUS_TIMEOUT_DEFAULT     0
200 #define CMD_TIMEOUT_DEFAULT     0
201 #else
202 /* Default timeout value in ms */
203 #ifdef DHD_EFI
204 #define BUS_TIMEOUT_DEFAULT     800  /* 800ms */
205 #define CMD_TIMEOUT_DEFAULT     1500 /* 1.5s */
206 #define SCAN_TIMEOUT_DEFAULT    0
207 #define JOIN_TIMEOUT_DEFAULT    0
208 #else
209 #define BUS_TIMEOUT_DEFAULT     800
210 #define CMD_TIMEOUT_DEFAULT     1200
211 #define SCAN_TIMEOUT_DEFAULT    17000
212 #define JOIN_TIMEOUT_DEFAULT    7500
213 #endif /* DHD_EFI */
214 #endif /* BCMINTERNAL */
215 #endif /* REPORT_FATAL_TIMEOUTS */
216 
217 #ifdef SHOW_LOGTRACE
218 #define BYTES_AHEAD_NUM		10	/* address in map file is before these many bytes */
219 #define READ_NUM_BYTES		1000 /* read map file each time this No. of bytes */
220 #define GO_BACK_FILE_POS_NUM_BYTES	100 /* set file pos back to cur pos */
221 static char *ramstart_str = " text_start"; /* string in mapfile has addr ramstart */
222 static char *rodata_start_str = " rodata_start"; /* string in mapfile has addr rodata start */
223 static char *rodata_end_str = " rodata_end"; /* string in mapfile has addr rodata end */
224 #define RAMSTART_BIT	0x01
225 #define RDSTART_BIT		0x02
226 #define RDEND_BIT		0x04
227 #define ALL_MAP_VAL		(RAMSTART_BIT | RDSTART_BIT | RDEND_BIT)
228 #endif /* SHOW_LOGTRACE */
229 
230 #ifdef SHOW_LOGTRACE
231 #if defined(LINUX) || defined(linux)
232 /* the fw file path is taken from either the module parameter at
233  * insmod time or is defined as a constant of different values
234  * for different platforms
235  */
236 extern char *st_str_file_path;
237 #else
238 static char *st_str_file_path = "rtecdc.bin";
239 #endif /* LINUX */
240 #endif /* SHOW_LOGTRACE */
241 
242 #ifdef EWP_EDL
243 typedef struct msg_hdr_edl {
244 	uint32 infobuf_ver;
245 	info_buf_payload_hdr_t pyld_hdr;
246 	msgtrace_hdr_t trace_hdr;
247 } msg_hdr_edl_t;
248 #endif /* EWP_EDL */
249 
250 #define DHD_TPUT_MAX_TX_PKTS_BATCH	1000
251 
252 /* Last connection success/failure status */
253 uint32 dhd_conn_event;
254 uint32 dhd_conn_status;
255 uint32 dhd_conn_reason;
256 
257 extern int dhd_iscan_request(void * dhdp, uint16 action);
258 extern void dhd_ind_scan_confirm(void *h, bool status);
259 extern int dhd_iscan_in_progress(void *h);
260 void dhd_iscan_lock(void);
261 void dhd_iscan_unlock(void);
262 extern int dhd_change_mtu(dhd_pub_t *dhd, int new_mtu, int ifidx);
263 #if defined(OEM_ANDROID) && !defined(AP) && defined(WLP2P)
264 extern int dhd_get_concurrent_capabilites(dhd_pub_t *dhd);
265 #endif
266 
267 extern int dhd_socram_dump(struct dhd_bus *bus);
268 extern void dhd_set_packet_filter(dhd_pub_t *dhd);
269 
270 #ifdef DNGL_EVENT_SUPPORT
271 static void dngl_host_event_process(dhd_pub_t *dhdp, bcm_dngl_event_t *event,
272 	bcm_dngl_event_msg_t *dngl_event, size_t pktlen);
273 static int dngl_host_event(dhd_pub_t *dhdp, void *pktdata, bcm_dngl_event_msg_t *dngl_event,
274 	size_t pktlen);
275 #endif /* DNGL_EVENT_SUPPORT */
276 
277 #ifdef WL_CFGVENDOR_SEND_HANG_EVENT
278 static void copy_hang_info_ioctl_timeout(dhd_pub_t *dhd, int ifidx, wl_ioctl_t *ioc);
279 #endif /* WL_CFGVENDOR_SEND_HANG_EVENT */
280 
281 #ifdef REPORT_FATAL_TIMEOUTS
282 static void dhd_set_join_error(dhd_pub_t *pub, uint32 mask);
283 #endif /* REPORT_FATAL_TIMEOUTS */
284 
285 #ifdef DHD_SEND_HANG_IOCTL_SUSPEND_ERROR
286 #define MAX_IOCTL_SUSPEND_ERROR	10
287 static int ioctl_suspend_error = 0;
288 #endif /* DHD_SEND_HANG_IOCTL_SUSPEND_ERROR */
289 
290 /* Should ideally read this from target(taken from wlu) */
291 #define MAX_CHUNK_LEN 1408 /* 8 * 8 * 22 */
292 
293 #if defined(OEM_ANDROID)
294 /* note these variables will be used with wext */
295 bool ap_cfg_running = FALSE;
296 bool ap_fw_loaded = FALSE;
297 #endif /* defined(OEM_ANDROID) && defined(SOFTAP) */
298 
299 #ifdef WLEASYMESH
300 extern int dhd_set_1905_almac(dhd_pub_t *dhdp, uint8 ifidx, uint8* ea, bool mcast);
301 extern int dhd_get_1905_almac(dhd_pub_t *dhdp, uint8 ifidx, uint8* ea, bool mcast);
302 #endif /* WLEASYMESH */
303 
304 #define CHIPID_MISMATCH	8
305 
306 #define DHD_VERSION "Dongle Host Driver, version " EPI_VERSION_STR "\n"
307 
308 #if defined(DHD_DEBUG) && defined(DHD_COMPILED)
309 const char dhd_version[] = DHD_VERSION DHD_COMPILED " compiled on "
310 			__DATE__ " at " __TIME__ "\n\0<TIMESTAMP>";
311 #else
312 const char dhd_version[] = DHD_VERSION;
313 #endif /* DHD_DEBUG && DHD_COMPILED */
314 
315 char fw_version[FW_VER_STR_LEN] = "\0";
316 char clm_version[CLM_VER_STR_LEN] = "\0";
317 
318 char bus_api_revision[BUS_API_REV_STR_LEN] = "\0";
319 
320 void dhd_set_timer(void *bus, uint wdtick);
321 
322 #if defined(BCM_ROUTER_DHD)
323 static int traffic_mgmt_add_dwm_filter(dhd_pub_t *dhd,
324 	trf_mgmt_filter_list_t * trf_mgmt_filter_list, int len);
325 #endif
326 
327 static char* ioctl2str(uint32 ioctl);
328 
329 /* IOVar table */
330 enum {
331 	IOV_VERSION = 1,
332 	IOV_WLMSGLEVEL,
333 	IOV_MSGLEVEL,
334 	IOV_BCMERRORSTR,
335 	IOV_BCMERROR,
336 	IOV_WDTICK,
337 	IOV_DUMP,
338 	IOV_CLEARCOUNTS,
339 	IOV_LOGDUMP,
340 	IOV_LOGCAL,
341 	IOV_LOGSTAMP,
342 	IOV_GPIOOB,
343 	IOV_IOCTLTIMEOUT,
344 	IOV_CONS,
345 	IOV_DCONSOLE_POLL,
346 #if defined(DHD_DEBUG)
347 	IOV_DHD_JOIN_TIMEOUT_DBG,
348 	IOV_SCAN_TIMEOUT,
349 	IOV_MEM_DEBUG,
350 #ifdef BCMPCIE
351 	IOV_FLOW_RING_DEBUG,
352 #endif /* BCMPCIE */
353 #endif /* defined(DHD_DEBUG) */
354 #ifdef PROP_TXSTATUS
355 	IOV_PROPTXSTATUS_ENABLE,
356 	IOV_PROPTXSTATUS_MODE,
357 	IOV_PROPTXSTATUS_OPT,
358 #ifdef QMONITOR
359 	IOV_QMON_TIME_THRES,
360 	IOV_QMON_TIME_PERCENT,
361 #endif /* QMONITOR */
362 	IOV_PROPTXSTATUS_MODULE_IGNORE,
363 	IOV_PROPTXSTATUS_CREDIT_IGNORE,
364 	IOV_PROPTXSTATUS_TXSTATUS_IGNORE,
365 	IOV_PROPTXSTATUS_RXPKT_CHK,
366 #endif /* PROP_TXSTATUS */
367 	IOV_BUS_TYPE,
368 	IOV_CHANGEMTU,
369 	IOV_HOSTREORDER_FLOWS,
370 #ifdef DHDTCPACK_SUPPRESS
371 	IOV_TCPACK_SUPPRESS,
372 #endif /* DHDTCPACK_SUPPRESS */
373 #ifdef DHD_WMF
374 	IOV_WMF_BSS_ENAB,
375 	IOV_WMF_UCAST_IGMP,
376 	IOV_WMF_MCAST_DATA_SENDUP,
377 #ifdef WL_IGMP_UCQUERY
378 	IOV_WMF_UCAST_IGMP_QUERY,
379 #endif /* WL_IGMP_UCQUERY */
380 #ifdef DHD_UCAST_UPNP
381 	IOV_WMF_UCAST_UPNP,
382 #endif /* DHD_UCAST_UPNP */
383 	IOV_WMF_PSTA_DISABLE,
384 #endif /* DHD_WMF */
385 #if defined(BCM_ROUTER_DHD)
386 	IOV_TRAFFIC_MGMT_DWM,
387 #endif /* BCM_ROUTER_DHD */
388 	IOV_AP_ISOLATE,
389 #ifdef DHD_L2_FILTER
390 	IOV_DHCP_UNICAST,
391 	IOV_BLOCK_PING,
392 	IOV_PROXY_ARP,
393 	IOV_GRAT_ARP,
394 	IOV_BLOCK_TDLS,
395 #endif /* DHD_L2_FILTER */
396 	IOV_DHD_IE,
397 #ifdef DHD_PSTA
398 	IOV_PSTA,
399 #endif /* DHD_PSTA */
400 #ifdef DHD_WET
401 	IOV_WET,
402 	IOV_WET_HOST_IPV4,
403 	IOV_WET_HOST_MAC,
404 #endif /* DHD_WET */
405 	IOV_CFG80211_OPMODE,
406 	IOV_ASSERT_TYPE,
407 #if defined(NDIS)
408 	IOV_WAKEIND,
409 #endif /* NDIS */
410 #if !defined(NDIS) && !defined(BCM_ROUTER_DHD)
411 	IOV_LMTEST,
412 #endif
413 #ifdef DHD_MCAST_REGEN
414 	IOV_MCAST_REGEN_BSS_ENABLE,
415 #endif
416 #ifdef BCMDBG
417 	IOV_MACDBG_PD11REGS,
418 	IOV_MACDBG_REGLIST,
419 	IOV_MACDBG_PSVMPMEMS,
420 #endif /* BCMDBG */
421 #ifdef SHOW_LOGTRACE
422 	IOV_DUMP_TRACE_LOG,
423 #endif /* SHOW_LOGTRACE */
424 #ifdef REPORT_FATAL_TIMEOUTS
425 	IOV_SCAN_TO,
426 	IOV_JOIN_TO,
427 	IOV_CMD_TO,
428 	IOV_OQS_TO,
429 #endif /* REPORT_FATAL_TIMEOUTS */
430 	IOV_DONGLE_TRAP_TYPE,
431 	IOV_DONGLE_TRAP_INFO,
432 	IOV_BPADDR,
433 	IOV_DUMP_DONGLE, /**< dumps core registers and d11 memories */
434 #if defined(DHD_LOG_DUMP)
435 #if defined(DHD_EFI)
436 	IOV_LOG_CAPTURE_ENABLE,
437 #endif
438 	IOV_LOG_DUMP,
439 #endif /* DHD_LOG_DUMP */
440 #ifdef BTLOG
441 	IOV_DUMP_BT_LOG,
442 	IOV_BTLOG,
443 #endif	/* BTLOG */
444 #ifdef SNAPSHOT_UPLOAD
445 	IOV_BT_MEM_DUMP,
446 	IOV_BT_UPLOAD,
447 #endif	/* SNAPSHOT_UPLOAD */
448 	IOV_TPUT_TEST,
449 #if defined(DHD_AWDL) && defined(AWDL_SLOT_STATS)
450 	IOV_PKT_LATENCY,
451 #endif /* defined(DHD_AWDL) && defined(AWDL_SLOT_STATS) */
452 	IOV_DEBUG_BUF_DEST_STAT,
453 #ifdef DHD_PKTTS
454 	IOV_PKTTS_ENAB,
455 	IOV_PKTTS_FLOW,
456 #endif /* DHD_PKTTS */
457 #ifdef DHD_DEBUG
458 	IOV_INDUCE_ERROR,
459 #endif /* DHD_DEBUG */
460 #if defined(DHD_EFI)
461 	IOV_INTR_POLL,
462 #endif
463 	IOV_FIS_TRIGGER,
464 #ifdef WL_IFACE_MGMT_CONF
465 #ifdef WL_CFG80211
466 #ifdef WL_NANP2P
467 	IOV_CONC_DISC,
468 #endif /* WL_NANP2P */
469 #ifdef WL_IFACE_MGMT
470 	IOV_IFACE_POLICY,
471 #endif /* WL_IFACE_MGMT */
472 #endif /* WL_CFG80211 */
473 #endif /* WL_IFACE_MGMT_CONF */
474 #ifdef RTT_GEOFENCE_CONT
475 #if defined (RTT_SUPPORT) && defined (WL_NAN)
476 	IOV_RTT_GEOFENCE_TYPE_OVRD,
477 #endif /* RTT_SUPPORT && WL_NAN */
478 #endif /* RTT_GEOFENCE_CONT */
479 	IOV_FW_VBS,
480 #ifdef DHD_TX_PROFILE
481 	IOV_TX_PROFILE_TAG,
482 	IOV_TX_PROFILE_ENABLE,
483 	IOV_TX_PROFILE_DUMP,
484 #endif /* defined(DHD_TX_PROFILE) */
485 	IOV_CHECK_TRAP_ROT,
486 #if defined(DHD_AWDL)
487 	IOV_AWDL_LLC_ENABLE,
488 #endif
489 #ifdef WLEASYMESH
490 	IOV_1905_AL_UCAST,
491 	IOV_1905_AL_MCAST,
492 #endif /* WLEASYMESH */
493 	IOV_LAST
494 };
495 
496 const bcm_iovar_t dhd_iovars[] = {
497 	/* name         varid                   flags   flags2 type     minlen */
498 	{"version",	IOV_VERSION,		0,	0, IOVT_BUFFER,	0},
499 	{"wlmsglevel",	IOV_WLMSGLEVEL,	0,	0,	IOVT_UINT32,	0 },
500 #ifdef DHD_DEBUG
501 	{"msglevel",	IOV_MSGLEVEL,		0,	0, IOVT_UINT32,	0},
502 	{"mem_debug",   IOV_MEM_DEBUG,  0,      0,      IOVT_BUFFER,    0 },
503 #ifdef BCMPCIE
504 	{"flow_ring_debug", IOV_FLOW_RING_DEBUG, 0, 0, IOVT_BUFFER, 0 },
505 #endif /* BCMPCIE */
506 #ifdef NDIS
507 	{"wlmsglevel",	IOV_WLMSGLEVEL,		0,	0, IOVT_UINT32,	0},
508 #endif /* NDIS */
509 #endif /* DHD_DEBUG */
510 	{"bcmerrorstr", IOV_BCMERRORSTR,	0,	0, IOVT_BUFFER,	BCME_STRLEN},
511 	{"bcmerror",	IOV_BCMERROR,		0,	0, IOVT_INT8,	0},
512 	{"wdtick",	IOV_WDTICK,		0,	0, IOVT_UINT32,	0},
513 	{"dump",	IOV_DUMP,		0,	0, IOVT_BUFFER,	DHD_IOCTL_MAXLEN_32K},
514 	{"cons",	IOV_CONS,		0,	0, IOVT_BUFFER,	0},
515 	{"dconpoll",	IOV_DCONSOLE_POLL,	0,	0, IOVT_UINT32,	0},
516 	{"clearcounts", IOV_CLEARCOUNTS,	0,	0, IOVT_VOID,	0},
517 #ifdef BCMPERFSTATS
518 	{"logdump", IOV_LOGDUMP,		0,	0, IOVT_BUFFER,	DHD_IOCTL_MAXLEN},
519 	{"logcal",	IOV_LOGCAL,		0,	0, IOVT_UINT32,	0},
520 	{"logstamp",	IOV_LOGSTAMP,		0,	0, IOVT_BUFFER,	0},
521 #endif
522 	{"gpioob",	IOV_GPIOOB,		0,	0, IOVT_UINT32,	0},
523 	{"ioctl_timeout", IOV_IOCTLTIMEOUT,	0,	0, IOVT_UINT32,	0},
524 #ifdef PROP_TXSTATUS
525 	{"proptx",	IOV_PROPTXSTATUS_ENABLE,	0,	0, IOVT_BOOL,	0 },
526 	/*
527 	set the proptxtstatus operation mode:
528 	0 - Do not do any proptxtstatus flow control
529 	1 - Use implied credit from a packet status
530 	2 - Use explicit credit
531 	*/
532 	{"ptxmode",	IOV_PROPTXSTATUS_MODE,	0,	0, IOVT_UINT32,	0 },
533 	{"proptx_opt", IOV_PROPTXSTATUS_OPT,	0,	0, IOVT_UINT32,	0 },
534 #ifdef QMONITOR
535 	{"qtime_thres",	IOV_QMON_TIME_THRES,	0,	0, IOVT_UINT32,	0 },
536 	{"qtime_percent", IOV_QMON_TIME_PERCENT, 0,	0, IOVT_UINT32,	0 },
537 #endif /* QMONITOR */
538 	{"pmodule_ignore", IOV_PROPTXSTATUS_MODULE_IGNORE, 0, 0, IOVT_BOOL, 0 },
539 	{"pcredit_ignore", IOV_PROPTXSTATUS_CREDIT_IGNORE, 0, 0, IOVT_BOOL, 0 },
540 	{"ptxstatus_ignore", IOV_PROPTXSTATUS_TXSTATUS_IGNORE, 0, 0,  IOVT_BOOL, 0 },
541 	{"rxpkt_chk", IOV_PROPTXSTATUS_RXPKT_CHK, 0, 0, IOVT_BOOL, 0 },
542 #endif /* PROP_TXSTATUS */
543 	{"bustype", IOV_BUS_TYPE, 0, 0, IOVT_UINT32, 0},
544 	{"changemtu", IOV_CHANGEMTU, 0, 0, IOVT_UINT32, 0 },
545 	{"host_reorder_flows", IOV_HOSTREORDER_FLOWS, 0, 0, IOVT_BUFFER,
546 	(WLHOST_REORDERDATA_MAXFLOWS + 1) },
547 #ifdef DHDTCPACK_SUPPRESS
548 	{"tcpack_suppress",	IOV_TCPACK_SUPPRESS,	0,	0, IOVT_UINT8,	0 },
549 #endif /* DHDTCPACK_SUPPRESS */
550 #ifdef DHD_WMF
551 	{"wmf_bss_enable", IOV_WMF_BSS_ENAB,	0,	0, IOVT_BOOL,	0 },
552 	{"wmf_ucast_igmp", IOV_WMF_UCAST_IGMP,	0,	0, IOVT_BOOL,	0 },
553 	{"wmf_mcast_data_sendup", IOV_WMF_MCAST_DATA_SENDUP,	0,	0, IOVT_BOOL,	0 },
554 #ifdef WL_IGMP_UCQUERY
555 	{"wmf_ucast_igmp_query", IOV_WMF_UCAST_IGMP_QUERY, (0), 0, IOVT_BOOL, 0 },
556 #endif /* WL_IGMP_UCQUERY */
557 #ifdef DHD_UCAST_UPNP
558 	{"wmf_ucast_upnp", IOV_WMF_UCAST_UPNP, (0), 0, IOVT_BOOL, 0 },
559 #endif /* DHD_UCAST_UPNP */
560 	{"wmf_psta_disable", IOV_WMF_PSTA_DISABLE, (0), 0, IOVT_BOOL, 0 },
561 #endif /* DHD_WMF */
562 #if defined(BCM_ROUTER_DHD)
563 	{"trf_mgmt_filters_add", IOV_TRAFFIC_MGMT_DWM, (0), 0, IOVT_BUFFER, 0},
564 #endif /* BCM_ROUTER_DHD */
565 #ifdef DHD_L2_FILTER
566 	{"dhcp_unicast", IOV_DHCP_UNICAST, (0), 0, IOVT_BOOL, 0 },
567 #endif /* DHD_L2_FILTER */
568 	{"ap_isolate", IOV_AP_ISOLATE, (0), 0, IOVT_BOOL, 0},
569 #ifdef DHD_L2_FILTER
570 	{"block_ping", IOV_BLOCK_PING, (0), 0, IOVT_BOOL, 0},
571 	{"proxy_arp", IOV_PROXY_ARP, (0), 0, IOVT_BOOL, 0},
572 	{"grat_arp", IOV_GRAT_ARP, (0), 0, IOVT_BOOL, 0},
573 	{"block_tdls", IOV_BLOCK_TDLS, (0), IOVT_BOOL, 0},
574 #endif /* DHD_L2_FILTER */
575 	{"dhd_ie", IOV_DHD_IE, (0), 0, IOVT_BUFFER, 0},
576 #ifdef DHD_PSTA
577 	/* PSTA/PSR Mode configuration. 0: DIABLED 1: PSTA 2: PSR */
578 	{"psta", IOV_PSTA, 0, 0, IOVT_UINT32, 0},
579 #endif /* DHD PSTA */
580 #ifdef DHD_WET
581 	/* WET Mode configuration. 0: DIABLED 1: WET */
582 	{"wet", IOV_WET, 0, 0, IOVT_UINT32, 0},
583 	{"wet_host_ipv4", IOV_WET_HOST_IPV4, 0, 0, IOVT_UINT32, 0},
584 	{"wet_host_mac", IOV_WET_HOST_MAC, 0, 0, IOVT_BUFFER, 0},
585 #endif /* DHD WET */
586 	{"op_mode",	IOV_CFG80211_OPMODE,	0,	0, IOVT_UINT32,	0 },
587 	{"assert_type", IOV_ASSERT_TYPE, (0), 0, IOVT_UINT32, 0},
588 #if defined(NDIS)
589 	{ "wowl_wakeind", IOV_WAKEIND, 0, 0, IOVT_UINT32, 0 },
590 #endif /* NDIS */
591 #if !defined(NDIS) && !defined(BCM_ROUTER_DHD)
592 	{"lmtest", IOV_LMTEST,	0,	0, IOVT_UINT32,	0 },
593 #endif
594 #ifdef DHD_MCAST_REGEN
595 	{"mcast_regen_bss_enable", IOV_MCAST_REGEN_BSS_ENABLE, 0, 0, IOVT_BOOL, 0},
596 #endif
597 #ifdef BCMDBG
598 	{"pd11regs", IOV_MACDBG_PD11REGS, 0, 0, IOVT_BUFFER, 0},
599 	{"mreglist", IOV_MACDBG_REGLIST, 0, 0, IOVT_BUFFER, 0},
600 	{"psvmpmems", IOV_MACDBG_PSVMPMEMS, 0, 0, IOVT_BUFFER, 0},
601 #endif /* BCMDBG */
602 #ifdef SHOW_LOGTRACE
603 	{"dump_trace_buf", IOV_DUMP_TRACE_LOG,	0, 0, IOVT_BUFFER,	sizeof(trace_buf_info_t) },
604 #endif /* SHOW_LOGTRACE */
605 #ifdef REPORT_FATAL_TIMEOUTS
606 	{"scan_timeout", IOV_SCAN_TO, 0, 0, IOVT_UINT32, 0 },
607 	{"join_timeout", IOV_JOIN_TO, 0, 0, IOVT_UINT32, 0 },
608 	{"cmd_timeout", IOV_CMD_TO, 0, 0, IOVT_UINT32, 0 },
609 	{"oqs_timeout", IOV_OQS_TO, 0, 0, IOVT_UINT32, 0 },
610 #endif /* REPORT_FATAL_TIMEOUTS */
611 	{"trap_type", IOV_DONGLE_TRAP_TYPE, 0, 0, IOVT_UINT32, 0 },
612 	{"trap_info", IOV_DONGLE_TRAP_INFO, 0, 0, IOVT_BUFFER, sizeof(trap_t) },
613 #ifdef DHD_DEBUG
614 	{"bpaddr", IOV_BPADDR,	0, 0, IOVT_BUFFER,	sizeof(sdreg_t) },
615 #endif /* DHD_DEBUG */
616 	{"dump_dongle", IOV_DUMP_DONGLE, 0, 0, IOVT_BUFFER,
617 	MAX(sizeof(dump_dongle_in_t), sizeof(dump_dongle_out_t)) },
618 #if defined(DHD_LOG_DUMP)
619 #if defined(DHD_EFI)
620 	{"log_capture_enable", IOV_LOG_CAPTURE_ENABLE, 0, 0, IOVT_UINT8, 0},
621 #endif
622 	{"log_dump", IOV_LOG_DUMP,	0, 0, IOVT_UINT8, 0},
623 #endif /* DHD_LOG_DUMP */
624 #ifdef BTLOG
625 	{"dump_bt_log", IOV_DUMP_BT_LOG, 0, 0, IOVT_BUFFER, sizeof(bt_log_buf_info_t) },
626 	{"btlog", IOV_BTLOG, 0, 0, IOVT_UINT32, 0 },
627 #endif	/* BTLOG */
628 #ifdef SNAPSHOT_UPLOAD
629 	{"bt_mem_dump", IOV_BT_MEM_DUMP, 0, 0, IOVT_UINT32, 0},
630 	{"bt_upload", IOV_BT_UPLOAD, 0, 0, IOVT_BUFFER, sizeof(bt_log_buf_info_t) },
631 #endif	/* SNAPSHOT_UPLOAD */
632 	{"tput_test", IOV_TPUT_TEST, 0, 0, IOVT_BUFFER, sizeof(tput_test_t)},
633 	{"debug_buf_dest_stat", IOV_DEBUG_BUF_DEST_STAT, 0, 0, IOVT_UINT32, 0 },
634 #ifdef DHD_PKTTS
635 	{"pktts_enab", IOV_PKTTS_ENAB, (0), 0, IOVT_BOOL, 0 },
636 	{"pktts_flow", IOV_PKTTS_FLOW, (0), 0, IOVT_BUFFER, sizeof(tput_test_t) },
637 #endif /* DHD_PKTTS */
638 #if defined(DHD_EFI)
639 	{"intr_poll", IOV_INTR_POLL, 0, 0, IOVT_BUFFER, sizeof(intr_poll_t)},
640 #endif
641 #if defined(DHD_AWDL) && defined(AWDL_SLOT_STATS)
642 	{"pkt_latency",	IOV_PKT_LATENCY,	0,	0,	IOVT_UINT32,	0 },
643 #endif /* defined(DHD_AWDL) && defined(AWDL_SLOT_STATS) */
644 #if defined(DHD_SSSR_DUMP)
645 	{"fis_trigger", IOV_FIS_TRIGGER, 0, 0, IOVT_UINT32, 0},
646 #endif
647 #ifdef DHD_DEBUG
648 	{"induce_error", IOV_INDUCE_ERROR, (0), 0, IOVT_UINT16, 0 },
649 #endif /* DHD_DEBUG */
650 #ifdef WL_IFACE_MGMT_CONF
651 #ifdef WL_CFG80211
652 #ifdef WL_NANP2P
653 	{"conc_disc", IOV_CONC_DISC, (0), 0, IOVT_UINT16, 0 },
654 #endif /* WL_NANP2P */
655 #ifdef WL_IFACE_MGMT
656 	{"if_policy", IOV_IFACE_POLICY, (0), 0, IOVT_BUFFER, sizeof(iface_mgmt_data_t)},
657 #endif /* WL_IFACE_MGMT */
658 #endif /* WL_CFG80211 */
659 #endif /* WL_IFACE_MGMT_CONF */
660 #ifdef RTT_GEOFENCE_CONT
661 #if defined (RTT_SUPPORT) && defined (WL_NAN)
662 	{"rtt_geofence_type_ovrd", IOV_RTT_GEOFENCE_TYPE_OVRD, (0), 0, IOVT_BOOL, 0},
663 #endif /* RTT_SUPPORT && WL_NAN */
664 #endif /* RTT_GEOFENCE_CONT */
665 	{"fw_verbose", IOV_FW_VBS, 0, 0, IOVT_UINT32, 0},
666 #ifdef DHD_TX_PROFILE
667 	{"tx_profile_tag", IOV_TX_PROFILE_TAG, 0, 0, IOVT_BUFFER,
668 	sizeof(dhd_tx_profile_protocol_t)},
669 	{"tx_profile_enable",	IOV_TX_PROFILE_ENABLE,	0,	0,	IOVT_BOOL,	0},
670 	{"tx_profile_dump",	IOV_TX_PROFILE_DUMP,	0,	0,	IOVT_UINT32,	0},
671 #endif /* defined(DHD_TX_PROFILE) */
672 	{"check_trap_rot", IOV_CHECK_TRAP_ROT, (0), 0, IOVT_BOOL, 0},
673 #if defined(DHD_AWDL)
674 	{"awdl_llc_enable", IOV_AWDL_LLC_ENABLE, 0, 0, IOVT_BOOL, 0},
675 #endif
676 	/* --- add new iovars *ABOVE* this line --- */
677 #ifdef WLEASYMESH
678 	{"1905_al_ucast", IOV_1905_AL_UCAST, 0, 0, IOVT_BUFFER, ETHER_ADDR_LEN},
679 	{"1905_al_mcast", IOV_1905_AL_MCAST, 0, 0, IOVT_BUFFER, ETHER_ADDR_LEN},
680 #endif /* WLEASYMESH */
681 	{NULL, 0, 0, 0, 0, 0 }
682 };
683 
684 #define DHD_IOVAR_BUF_SIZE	128
685 
686 #if defined(LINUX) || defined(linux) || defined(DHD_EFI)
687 fw_download_status_t
dhd_fw_download_status(dhd_pub_t * dhd_pub)688 dhd_fw_download_status(dhd_pub_t * dhd_pub)
689 {
690 	return dhd_pub->fw_download_status;
691 }
692 #endif /* defined(LINUX) || defined(linux) || defined(DHD_EFI) */
693 
694 bool
dhd_query_bus_erros(dhd_pub_t * dhdp)695 dhd_query_bus_erros(dhd_pub_t *dhdp)
696 {
697 	bool ret = FALSE;
698 
699 	if (dhdp->dongle_reset) {
700 		DHD_ERROR_RLMT(("%s: Dongle Reset occurred, cannot proceed\n",
701 			__FUNCTION__));
702 		ret = TRUE;
703 	}
704 
705 	if (dhdp->dongle_trap_occured) {
706 		DHD_ERROR_RLMT(("%s: FW TRAP has occurred, cannot proceed\n",
707 			__FUNCTION__));
708 		ret = TRUE;
709 #ifdef OEM_ANDROID
710 		dhdp->hang_reason = HANG_REASON_DONGLE_TRAP;
711 		dhd_os_send_hang_message(dhdp);
712 #endif /* OEM_ANDROID */
713 	}
714 
715 	if (dhdp->iovar_timeout_occured) {
716 		DHD_ERROR_RLMT(("%s: Resumed on timeout for previous IOVAR, cannot proceed\n",
717 			__FUNCTION__));
718 		ret = TRUE;
719 	}
720 
721 #ifdef PCIE_FULL_DONGLE
722 	if (dhdp->d3ack_timeout_occured) {
723 		DHD_ERROR_RLMT(("%s: Resumed on timeout for previous D3ACK, cannot proceed\n",
724 			__FUNCTION__));
725 		ret = TRUE;
726 	}
727 	if (dhdp->livelock_occured) {
728 		DHD_ERROR_RLMT(("%s: LIVELOCK occurred for previous msg, cannot proceed\n",
729 			__FUNCTION__));
730 		ret = TRUE;
731 	}
732 
733 	if (dhdp->pktid_audit_failed) {
734 		DHD_ERROR_RLMT(("%s: pktid_audit_failed, cannot proceed\n",
735 			__FUNCTION__));
736 		ret = TRUE;
737 	}
738 #endif /* PCIE_FULL_DONGLE */
739 
740 	if (dhdp->iface_op_failed) {
741 		DHD_ERROR_RLMT(("%s: iface_op_failed, cannot proceed\n",
742 			__FUNCTION__));
743 		ret = TRUE;
744 	}
745 
746 	if (dhdp->scan_timeout_occurred) {
747 		DHD_ERROR_RLMT(("%s: scan_timeout_occurred, cannot proceed\n",
748 			__FUNCTION__));
749 		ret = TRUE;
750 	}
751 
752 	if (dhdp->scan_busy_occurred) {
753 		DHD_ERROR_RLMT(("%s: scan_busy_occurred, cannot proceed\n",
754 			__FUNCTION__));
755 		ret = TRUE;
756 	}
757 
758 #ifdef DNGL_AXI_ERROR_LOGGING
759 	if (dhdp->axi_error) {
760 		DHD_ERROR_RLMT(("%s: AXI error occurred, cannot proceed\n",
761 			__FUNCTION__));
762 		ret = TRUE;
763 	}
764 #endif /* DNGL_AXI_ERROR_LOGGING */
765 
766 #if defined(BCMPCIE)
767 	if (dhd_bus_get_linkdown(dhdp)) {
768 		DHD_ERROR_RLMT(("%s : PCIE Link down occurred, cannot proceed\n",
769 			__FUNCTION__));
770 		ret = TRUE;
771 	}
772 
773 	if (dhd_bus_get_cto(dhdp)) {
774 		DHD_ERROR_RLMT(("%s : CTO Recovery reported, cannot proceed\n",
775 			__FUNCTION__));
776 		ret = TRUE;
777 	}
778 #endif
779 
780 	return ret;
781 }
782 
783 void
dhd_clear_bus_errors(dhd_pub_t * dhdp)784 dhd_clear_bus_errors(dhd_pub_t *dhdp)
785 {
786 	if (!dhdp)
787 		return;
788 
789 	dhdp->dongle_reset = FALSE;
790 	dhdp->dongle_trap_occured = FALSE;
791 	dhdp->iovar_timeout_occured = FALSE;
792 #ifdef PCIE_FULL_DONGLE
793 	dhdp->d3ack_timeout_occured = FALSE;
794 	dhdp->livelock_occured = FALSE;
795 	dhdp->pktid_audit_failed = FALSE;
796 #endif
797 	dhdp->iface_op_failed = FALSE;
798 	dhdp->scan_timeout_occurred = FALSE;
799 	dhdp->scan_busy_occurred = FALSE;
800 #ifdef BT_OVER_PCIE
801 	dhdp->dongle_trap_due_to_bt = FALSE;
802 #endif
803 }
804 
805 #ifdef DHD_SSSR_DUMP
806 
807 /* This can be overwritten by module parameter defined in dhd_linux.c */
808 uint sssr_enab = TRUE;
809 
810 #ifdef DHD_FIS_DUMP
811 uint fis_enab = TRUE;
812 #else
813 uint fis_enab = FALSE;
814 #endif /* DHD_FIS_DUMP */
815 
816 int
dhd_sssr_mempool_init(dhd_pub_t * dhd)817 dhd_sssr_mempool_init(dhd_pub_t *dhd)
818 {
819 	dhd->sssr_mempool = (uint8 *) MALLOCZ(dhd->osh, DHD_SSSR_MEMPOOL_SIZE);
820 	if (dhd->sssr_mempool == NULL) {
821 		DHD_ERROR(("%s: MALLOC of sssr_mempool failed\n",
822 			__FUNCTION__));
823 		return BCME_ERROR;
824 	}
825 	return BCME_OK;
826 }
827 
828 void
dhd_sssr_mempool_deinit(dhd_pub_t * dhd)829 dhd_sssr_mempool_deinit(dhd_pub_t *dhd)
830 {
831 	if (dhd->sssr_mempool) {
832 		MFREE(dhd->osh, dhd->sssr_mempool, DHD_SSSR_MEMPOOL_SIZE);
833 		dhd->sssr_mempool = NULL;
834 	}
835 }
836 
837 int
dhd_sssr_reg_info_init(dhd_pub_t * dhd)838 dhd_sssr_reg_info_init(dhd_pub_t *dhd)
839 {
840 	dhd->sssr_reg_info = (sssr_reg_info_cmn_t *) MALLOCZ(dhd->osh, sizeof(sssr_reg_info_cmn_t));
841 	if (dhd->sssr_reg_info == NULL) {
842 		DHD_ERROR(("%s: MALLOC of sssr_reg_info failed\n",
843 			__FUNCTION__));
844 		return BCME_ERROR;
845 	}
846 	return BCME_OK;
847 }
848 
849 void
dhd_sssr_reg_info_deinit(dhd_pub_t * dhd)850 dhd_sssr_reg_info_deinit(dhd_pub_t *dhd)
851 {
852 	if (dhd->sssr_reg_info) {
853 		MFREE(dhd->osh, dhd->sssr_reg_info, sizeof(sssr_reg_info_cmn_t));
854 		dhd->sssr_reg_info = NULL;
855 	}
856 }
857 
858 #ifdef DHD_PCIE_REG_ACCESS
859 static void
dhd_dump_sssr_reg_info_v2(dhd_pub_t * dhd)860 dhd_dump_sssr_reg_info_v2(dhd_pub_t *dhd)
861 {
862 	sssr_reg_info_cmn_t *sssr_reg_info_cmn = dhd->sssr_reg_info;
863 	sssr_reg_info_v2_t *sssr_reg_info = (sssr_reg_info_v2_t *)&sssr_reg_info_cmn->rev2;
864 	int i, j;
865 	uint8 num_d11cores = dhd_d11_slices_num_get(dhd);
866 	DHD_ERROR(("pmu_regs\n"));
867 	DHD_ERROR(("pmuintmask0=0x%x pmuintmask1=0x%x resreqtimer=0x%x "
868 		"macresreqtimer=0x%x macresreqtimer1=0x%x\n",
869 		sssr_reg_info->pmu_regs.base_regs.pmuintmask0,
870 		sssr_reg_info->pmu_regs.base_regs.pmuintmask1,
871 		sssr_reg_info->pmu_regs.base_regs.resreqtimer,
872 		sssr_reg_info->pmu_regs.base_regs.macresreqtimer,
873 		sssr_reg_info->pmu_regs.base_regs.macresreqtimer1));
874 	DHD_ERROR(("chipcommon_regs\n"));
875 	DHD_ERROR(("intmask=0x%x powerctrl=0x%x clockcontrolstatus=0x%x powerctrl_mask=0x%x\n",
876 		sssr_reg_info->chipcommon_regs.base_regs.intmask,
877 		sssr_reg_info->chipcommon_regs.base_regs.powerctrl,
878 		sssr_reg_info->chipcommon_regs.base_regs.clockcontrolstatus,
879 		sssr_reg_info->chipcommon_regs.base_regs.powerctrl_mask));
880 	DHD_ERROR(("arm_regs\n"));
881 	DHD_ERROR(("clockcontrolstatus=0x%x clockcontrolstatus_val=0x%x"
882 		" resetctrl=0x%x extrsrcreq=0x%x\n",
883 		sssr_reg_info->arm_regs.base_regs.clockcontrolstatus,
884 		sssr_reg_info->arm_regs.base_regs.clockcontrolstatus_val,
885 		sssr_reg_info->arm_regs.wrapper_regs.resetctrl,
886 		sssr_reg_info->arm_regs.wrapper_regs.extrsrcreq));
887 	DHD_ERROR(("pcie_regs\n"));
888 	DHD_ERROR(("ltrstate=0x%x clockcontrolstatus=0x%x "
889 		"clockcontrolstatus_val=0x%x extrsrcreq=0x%x\n",
890 		sssr_reg_info->pcie_regs.base_regs.ltrstate,
891 		sssr_reg_info->pcie_regs.base_regs.clockcontrolstatus,
892 		sssr_reg_info->pcie_regs.base_regs.clockcontrolstatus_val,
893 		sssr_reg_info->pcie_regs.wrapper_regs.extrsrcreq));
894 
895 	for (i = 0; i < num_d11cores; i++) {
896 		DHD_ERROR(("mac_regs core[%d]\n", i));
897 		DHD_ERROR(("xmtaddress=0x%x xmtdata=0x%x clockcontrolstatus=0x%x "
898 			"clockcontrolstatus_val=0x%x\n",
899 			sssr_reg_info->mac_regs[i].base_regs.xmtaddress,
900 			sssr_reg_info->mac_regs[i].base_regs.xmtdata,
901 			sssr_reg_info->mac_regs[i].base_regs.clockcontrolstatus,
902 			sssr_reg_info->mac_regs[i].base_regs.clockcontrolstatus_val));
903 		DHD_ERROR(("resetctrl=0x%x extrsrcreq=0x%x ioctrl=0x%x\n",
904 			sssr_reg_info->mac_regs[i].wrapper_regs.resetctrl,
905 			sssr_reg_info->mac_regs[i].wrapper_regs.extrsrcreq,
906 			sssr_reg_info->mac_regs[i].wrapper_regs.ioctrl));
907 		for (j = 0; j < SSSR_D11_RESET_SEQ_STEPS; j++) {
908 			DHD_ERROR(("ioctrl_resetseq_val[%d] 0x%x\n", j,
909 				sssr_reg_info->mac_regs[i].wrapper_regs.ioctrl_resetseq_val[j]));
910 		}
911 		DHD_ERROR(("sr_size=0x%x\n", sssr_reg_info->mac_regs[i].sr_size));
912 	}
913 	DHD_ERROR(("dig_regs\n"));
914 	DHD_ERROR(("dig_sr_addr=0x%x dig_sr_size=0x%x\n",
915 		sssr_reg_info->dig_mem_info.dig_sr_addr,
916 		sssr_reg_info->dig_mem_info.dig_sr_size));
917 }
918 
919 static void
dhd_dump_sssr_reg_info_v3(dhd_pub_t * dhd)920 dhd_dump_sssr_reg_info_v3(dhd_pub_t *dhd)
921 {
922 	sssr_reg_info_cmn_t *sssr_reg_info_cmn = dhd->sssr_reg_info;
923 	sssr_reg_info_v3_t *sssr_reg_info = (sssr_reg_info_v3_t *)&sssr_reg_info_cmn->rev3;
924 	int i;
925 
926 	dhd_dump_sssr_reg_info_v2(dhd);
927 
928 	DHD_ERROR(("FIS Enab in fw : %d\n", sssr_reg_info->fis_enab));
929 
930 	DHD_ERROR(("HWA regs for reset \n"));
931 	DHD_ERROR(("clkenable 0x%x, clkgatingenable 0x%x, clkext 0x%x, "
932 		"clkctlstatus 0x%x, ioctrl 0x%x, resetctrl 0x%x\n",
933 		sssr_reg_info->hwa_regs.base_regs.clkenable,
934 		sssr_reg_info->hwa_regs.base_regs.clkgatingenable,
935 		sssr_reg_info->hwa_regs.base_regs.clkext,
936 		sssr_reg_info->hwa_regs.base_regs.clkctlstatus,
937 		sssr_reg_info->hwa_regs.wrapper_regs.ioctrl,
938 		sssr_reg_info->hwa_regs.wrapper_regs.resetctrl));
939 	DHD_ERROR(("HWA regs value seq for reset \n"));
940 	for (i = 0; i < SSSR_HWA_RESET_SEQ_STEPS; i++) {
941 		DHD_ERROR(("hwa_resetseq_val[%d] 0x%x", i,
942 			sssr_reg_info->hwa_regs.hwa_resetseq_val[i]));
943 	}
944 }
945 
946 static void
dhd_dump_sssr_reg_info_v1(dhd_pub_t * dhd)947 dhd_dump_sssr_reg_info_v1(dhd_pub_t *dhd)
948 {
949 	sssr_reg_info_cmn_t *sssr_reg_info_cmn = dhd->sssr_reg_info;
950 	sssr_reg_info_v1_t *sssr_reg_info = (sssr_reg_info_v1_t *)&sssr_reg_info_cmn->rev1;
951 	int i, j;
952 	uint8 num_d11cores = dhd_d11_slices_num_get(dhd);
953 
954 	DHD_ERROR(("pmu_regs\n"));
955 	DHD_ERROR(("pmuintmask0=0x%x pmuintmask1=0x%x resreqtimer=0x%x "
956 		"macresreqtimer=0x%x macresreqtimer1=0x%x\n",
957 		sssr_reg_info->pmu_regs.base_regs.pmuintmask0,
958 		sssr_reg_info->pmu_regs.base_regs.pmuintmask1,
959 		sssr_reg_info->pmu_regs.base_regs.resreqtimer,
960 		sssr_reg_info->pmu_regs.base_regs.macresreqtimer,
961 		sssr_reg_info->pmu_regs.base_regs.macresreqtimer1));
962 	DHD_ERROR(("chipcommon_regs\n"));
963 	DHD_ERROR(("intmask=0x%x powerctrl=0x%x clockcontrolstatus=0x%x powerctrl_mask=0x%x\n",
964 		sssr_reg_info->chipcommon_regs.base_regs.intmask,
965 		sssr_reg_info->chipcommon_regs.base_regs.powerctrl,
966 		sssr_reg_info->chipcommon_regs.base_regs.clockcontrolstatus,
967 		sssr_reg_info->chipcommon_regs.base_regs.powerctrl_mask));
968 	DHD_ERROR(("arm_regs\n"));
969 	DHD_ERROR(("clockcontrolstatus=0x%x clockcontrolstatus_val=0x%x"
970 		" resetctrl=0x%x itopoobb=0x%x\n",
971 		sssr_reg_info->arm_regs.base_regs.clockcontrolstatus,
972 		sssr_reg_info->arm_regs.base_regs.clockcontrolstatus_val,
973 		sssr_reg_info->arm_regs.wrapper_regs.resetctrl,
974 		sssr_reg_info->arm_regs.wrapper_regs.itopoobb));
975 	DHD_ERROR(("pcie_regs\n"));
976 	DHD_ERROR(("ltrstate=0x%x clockcontrolstatus=0x%x "
977 		"clockcontrolstatus_val=0x%x itopoobb=0x%x\n",
978 		sssr_reg_info->pcie_regs.base_regs.ltrstate,
979 		sssr_reg_info->pcie_regs.base_regs.clockcontrolstatus,
980 		sssr_reg_info->pcie_regs.base_regs.clockcontrolstatus_val,
981 		sssr_reg_info->pcie_regs.wrapper_regs.itopoobb));
982 	DHD_ERROR(("vasip_regs\n"));
983 	DHD_ERROR(("ioctrl=0x%x vasip_sr_addr=0x%x vasip_sr_size=0x%x\n",
984 		sssr_reg_info->vasip_regs.wrapper_regs.ioctrl,
985 		sssr_reg_info->vasip_regs.vasip_sr_addr,
986 		sssr_reg_info->vasip_regs.vasip_sr_size));
987 
988 	for (i = 0; i < num_d11cores; i++) {
989 		DHD_ERROR(("mac_regs core[%d]\n", i));
990 		DHD_ERROR(("xmtaddress=0x%x xmtdata=0x%x clockcontrolstatus=0x%x "
991 			"clockcontrolstatus_val=0x%x\n",
992 			sssr_reg_info->mac_regs[i].base_regs.xmtaddress,
993 			sssr_reg_info->mac_regs[i].base_regs.xmtdata,
994 			sssr_reg_info->mac_regs[i].base_regs.clockcontrolstatus,
995 			sssr_reg_info->mac_regs[i].base_regs.clockcontrolstatus_val));
996 		DHD_ERROR(("resetctrl=0x%x itopoobb=0x%x ioctrl=0x%x\n",
997 			sssr_reg_info->mac_regs[i].wrapper_regs.resetctrl,
998 			sssr_reg_info->mac_regs[i].wrapper_regs.itopoobb,
999 			sssr_reg_info->mac_regs[i].wrapper_regs.ioctrl));
1000 		for (j = 0; j < SSSR_D11_RESET_SEQ_STEPS; j++) {
1001 			DHD_ERROR(("ioctrl_resetseq_val[%d] 0x%x\n", j,
1002 				sssr_reg_info->mac_regs[i].wrapper_regs.ioctrl_resetseq_val[j]));
1003 		}
1004 		DHD_ERROR(("sr_size=0x%x\n", sssr_reg_info->mac_regs[i].sr_size));
1005 	}
1006 }
1007 
1008 #endif /* DHD_PCIE_REG_ACCESS */
1009 
1010 void
dhd_dump_sssr_reg_info(dhd_pub_t * dhd)1011 dhd_dump_sssr_reg_info(dhd_pub_t *dhd)
1012 {
1013 #ifdef DHD_PCIE_REG_ACCESS
1014 	sssr_reg_info_cmn_t *sssr_reg_info_cmn = dhd->sssr_reg_info;
1015 	sssr_reg_info_v1_t *sssr_reg_info = (sssr_reg_info_v1_t *)&sssr_reg_info_cmn->rev1;
1016 
1017 	DHD_ERROR(("************** SSSR REG INFO start version:%d ****************\n",
1018 		sssr_reg_info->version));
1019 	switch (sssr_reg_info->version) {
1020 		case SSSR_REG_INFO_VER_3 :
1021 			dhd_dump_sssr_reg_info_v3(dhd);
1022 			break;
1023 		case SSSR_REG_INFO_VER_2 :
1024 			dhd_dump_sssr_reg_info_v2(dhd);
1025 			break;
1026 		default:
1027 			dhd_dump_sssr_reg_info_v1(dhd);
1028 			break;
1029 	}
1030 	DHD_ERROR(("************** SSSR REG INFO end ****************\n"));
1031 #endif /* DHD_PCIE_REG_ACCESS */
1032 }
1033 
1034 int
dhd_get_sssr_reg_info(dhd_pub_t * dhd)1035 dhd_get_sssr_reg_info(dhd_pub_t *dhd)
1036 {
1037 	int ret;
1038 	/* get sssr_reg_info from firmware */
1039 	ret = dhd_iovar(dhd, 0, "sssr_reg_info", NULL, 0,  (char *)dhd->sssr_reg_info,
1040 		sizeof(sssr_reg_info_cmn_t), FALSE);
1041 	if (ret < 0) {
1042 		DHD_ERROR(("%s: sssr_reg_info failed (error=%d)\n",
1043 			__FUNCTION__, ret));
1044 		return BCME_ERROR;
1045 	}
1046 
1047 	dhd_dump_sssr_reg_info(dhd);
1048 	return BCME_OK;
1049 }
1050 
1051 uint32
dhd_get_sssr_bufsize(dhd_pub_t * dhd)1052 dhd_get_sssr_bufsize(dhd_pub_t *dhd)
1053 {
1054 	int i;
1055 	uint32 sssr_bufsize = 0;
1056 	uint8 num_d11cores;
1057 
1058 	num_d11cores = dhd_d11_slices_num_get(dhd);
1059 
1060 	switch (dhd->sssr_reg_info->rev2.version) {
1061 		case SSSR_REG_INFO_VER_3 :
1062 			/* intentional fall through */
1063 		case SSSR_REG_INFO_VER_2 :
1064 			for (i = 0; i < num_d11cores; i++) {
1065 				sssr_bufsize += dhd->sssr_reg_info->rev2.mac_regs[i].sr_size;
1066 			}
1067 			if ((dhd->sssr_reg_info->rev2.length >
1068 			 OFFSETOF(sssr_reg_info_v2_t, dig_mem_info)) &&
1069 			 dhd->sssr_reg_info->rev2.dig_mem_info.dig_sr_addr) {
1070 				sssr_bufsize += 0; /* TBD */
1071 			}
1072 			break;
1073 		case SSSR_REG_INFO_VER_1 :
1074 			for (i = 0; i < num_d11cores; i++) {
1075 				sssr_bufsize += dhd->sssr_reg_info->rev1.mac_regs[i].sr_size;
1076 			}
1077 			if (dhd->sssr_reg_info->rev1.vasip_regs.vasip_sr_size) {
1078 				sssr_bufsize += dhd->sssr_reg_info->rev1.vasip_regs.vasip_sr_size;
1079 			} else if ((dhd->sssr_reg_info->rev1.length > OFFSETOF(sssr_reg_info_v1_t,
1080 				dig_mem_info)) && dhd->sssr_reg_info->rev1.
1081 				dig_mem_info.dig_sr_addr) {
1082 				sssr_bufsize += dhd->sssr_reg_info->rev1.dig_mem_info.dig_sr_size;
1083 			}
1084 			break;
1085 		case SSSR_REG_INFO_VER_0 :
1086 			for (i = 0; i < num_d11cores; i++) {
1087 				sssr_bufsize += dhd->sssr_reg_info->rev0.mac_regs[i].sr_size;
1088 			}
1089 			if (dhd->sssr_reg_info->rev0.vasip_regs.vasip_sr_size) {
1090 				sssr_bufsize += dhd->sssr_reg_info->rev0.vasip_regs.vasip_sr_size;
1091 			}
1092 			break;
1093 		default :
1094 			DHD_ERROR(("invalid sssr_reg_ver"));
1095 			return BCME_UNSUPPORTED;
1096 	}
1097 
1098 #ifdef DHD_SSSR_DUMP_BEFORE_SR
1099 	/* Double the size as different dumps will be saved before and after SR */
1100 	sssr_bufsize = 2 * sssr_bufsize;
1101 #endif /* DHD_SSSR_DUMP_BEFORE_SR */
1102 
1103 	return sssr_bufsize;
1104 }
1105 
1106 int
dhd_sssr_dump_init(dhd_pub_t * dhd)1107 dhd_sssr_dump_init(dhd_pub_t *dhd)
1108 {
1109 	int i;
1110 	uint32 sssr_bufsize;
1111 	uint32 mempool_used = 0;
1112 	uint8 num_d11cores = 0;
1113 	bool alloc_sssr = FALSE;
1114 	uint32 sr_size = 0;
1115 
1116 	dhd->sssr_inited = FALSE;
1117 	if (!sssr_enab) {
1118 		DHD_ERROR(("%s: sssr dump not inited as instructed by mod param\n", __FUNCTION__));
1119 		return BCME_OK;
1120 	}
1121 
1122 	/* check if sssr mempool is allocated */
1123 	if (dhd->sssr_mempool == NULL) {
1124 		DHD_ERROR(("%s: sssr_mempool is not allocated\n",
1125 			__FUNCTION__));
1126 		return BCME_ERROR;
1127 	}
1128 
1129 	/* check if sssr mempool is allocated */
1130 	if (dhd->sssr_reg_info == NULL) {
1131 		DHD_ERROR(("%s: sssr_reg_info is not allocated\n",
1132 			__FUNCTION__));
1133 		return BCME_ERROR;
1134 	}
1135 
1136 	/* Get SSSR reg info */
1137 	if (dhd_get_sssr_reg_info(dhd) != BCME_OK) {
1138 		DHD_ERROR(("%s: dhd_get_sssr_reg_info failed\n", __FUNCTION__));
1139 		printf("DEBUG_SSSr: %s: dhd_get_sssr_reg_info failed\n", __FUNCTION__);
1140 		return BCME_ERROR;
1141 	}
1142 
1143 	num_d11cores = dhd_d11_slices_num_get(dhd);
1144 	/* Validate structure version and length */
1145 	switch (dhd->sssr_reg_info->rev2.version) {
1146 		case SSSR_REG_INFO_VER_3 :
1147 			if (dhd->sssr_reg_info->rev3.length != sizeof(sssr_reg_info_v3_t)) {
1148 				DHD_ERROR(("%s: dhd->sssr_reg_info->rev2.length (%d : %d)"
1149 					 "mismatch on rev2\n", __FUNCTION__,
1150 					 (int)dhd->sssr_reg_info->rev3.length,
1151 					 (int)sizeof(sssr_reg_info_v3_t)));
1152 				return BCME_ERROR;
1153 			}
1154 			break;
1155 		case SSSR_REG_INFO_VER_2 :
1156 			if (dhd->sssr_reg_info->rev2.length != sizeof(sssr_reg_info_v2_t)) {
1157 				DHD_ERROR(("%s: dhd->sssr_reg_info->rev2.length (%d : %d)"
1158 					 "mismatch on rev2\n", __FUNCTION__,
1159 					 (int)dhd->sssr_reg_info->rev2.length,
1160 					 (int)sizeof(sssr_reg_info_v2_t)));
1161 				return BCME_ERROR;
1162 			}
1163 			break;
1164 		case SSSR_REG_INFO_VER_1 :
1165 			if (dhd->sssr_reg_info->rev1.length != sizeof(sssr_reg_info_v1_t)) {
1166 				DHD_ERROR(("%s: dhd->sssr_reg_info->rev1.length (%d : %d)"
1167 					 "mismatch on rev1\n", __FUNCTION__,
1168 					 (int)dhd->sssr_reg_info->rev1.length,
1169 					 (int)sizeof(sssr_reg_info_v1_t)));
1170 				return BCME_ERROR;
1171 			}
1172 			break;
1173 		case SSSR_REG_INFO_VER_0 :
1174 			if (dhd->sssr_reg_info->rev0.length != sizeof(sssr_reg_info_v0_t)) {
1175 				DHD_ERROR(("%s: dhd->sssr_reg_info->rev0.length (%d : %d)"
1176 					 "mismatch on rev0\n", __FUNCTION__,
1177 					 (int)dhd->sssr_reg_info->rev0.length,
1178 					 (int)sizeof(sssr_reg_info_v0_t)));
1179 				return BCME_ERROR;
1180 			}
1181 			break;
1182 		default :
1183 			DHD_ERROR(("invalid sssr_reg_ver"));
1184 			return BCME_UNSUPPORTED;
1185 	}
1186 
1187 	/* validate fifo size */
1188 	sssr_bufsize = dhd_get_sssr_bufsize(dhd);
1189 	if (sssr_bufsize > DHD_SSSR_MEMPOOL_SIZE) {
1190 		DHD_ERROR(("%s: sssr_bufsize(%d) is greater than sssr_mempool(%d)\n",
1191 			__FUNCTION__, (int)sssr_bufsize, DHD_SSSR_MEMPOOL_SIZE));
1192 		return BCME_ERROR;
1193 	}
1194 
1195 	/* init all pointers to NULL */
1196 	for (i = 0; i < num_d11cores; i++) {
1197 #ifdef DHD_SSSR_DUMP_BEFORE_SR
1198 		dhd->sssr_d11_before[i] = NULL;
1199 #endif /* DHD_SSSR_DUMP_BEFORE_SR */
1200 		dhd->sssr_d11_after[i] = NULL;
1201 	}
1202 
1203 #ifdef DHD_SSSR_DUMP_BEFORE_SR
1204 	dhd->sssr_dig_buf_before = NULL;
1205 #endif /* DHD_SSSR_DUMP_BEFORE_SR */
1206 	dhd->sssr_dig_buf_after = NULL;
1207 
1208 	/* Allocate memory */
1209 	for (i = 0; i < num_d11cores; i++) {
1210 		alloc_sssr = FALSE;
1211 		sr_size = 0;
1212 
1213 		switch (dhd->sssr_reg_info->rev2.version) {
1214 			case SSSR_REG_INFO_VER_3 :
1215 				/* intentional fall through */
1216 			case SSSR_REG_INFO_VER_2 :
1217 				if (dhd->sssr_reg_info->rev2.mac_regs[i].sr_size) {
1218 					alloc_sssr = TRUE;
1219 					sr_size = dhd->sssr_reg_info->rev2.mac_regs[i].sr_size;
1220 				}
1221 				break;
1222 			case SSSR_REG_INFO_VER_1 :
1223 				if (dhd->sssr_reg_info->rev1.mac_regs[i].sr_size) {
1224 					alloc_sssr = TRUE;
1225 					sr_size = dhd->sssr_reg_info->rev1.mac_regs[i].sr_size;
1226 				}
1227 				break;
1228 			case SSSR_REG_INFO_VER_0 :
1229 				if (dhd->sssr_reg_info->rev0.mac_regs[i].sr_size) {
1230 					alloc_sssr = TRUE;
1231 					sr_size = dhd->sssr_reg_info->rev0.mac_regs[i].sr_size;
1232 				}
1233 				break;
1234 			default :
1235 				DHD_ERROR(("invalid sssr_reg_ver"));
1236 				return BCME_UNSUPPORTED;
1237 		}
1238 
1239 		if (alloc_sssr) {
1240 #ifdef DHD_SSSR_DUMP_BEFORE_SR
1241 			dhd->sssr_d11_before[i] = (uint32 *)(dhd->sssr_mempool + mempool_used);
1242 			mempool_used += sr_size;
1243 #endif /* DHD_SSSR_DUMP_BEFORE_SR */
1244 
1245 			dhd->sssr_d11_after[i] = (uint32 *)(dhd->sssr_mempool + mempool_used);
1246 			mempool_used += sr_size;
1247 		}
1248 	}
1249 
1250 	/* Allocate dump memory for VASIP (version 0 or 1) or digital core (version 0, 1, or 2) */
1251 	alloc_sssr = FALSE;
1252 	sr_size = 0;
1253 	switch (dhd->sssr_reg_info->rev2.version) {
1254 		case SSSR_REG_INFO_VER_3 :
1255 			/* intentional fall through */
1256 		case SSSR_REG_INFO_VER_2 :
1257 			if ((dhd->sssr_reg_info->rev2.length >
1258 			 OFFSETOF(sssr_reg_info_v2_t, dig_mem_info)) &&
1259 			 dhd->sssr_reg_info->rev2.dig_mem_info.dig_sr_addr) {
1260 				alloc_sssr = TRUE;
1261 				sr_size = dhd->sssr_reg_info->rev2.dig_mem_info.dig_sr_size;
1262 			}
1263 			break;
1264 		case SSSR_REG_INFO_VER_1 :
1265 			if (dhd->sssr_reg_info->rev1.vasip_regs.vasip_sr_size) {
1266 				alloc_sssr = TRUE;
1267 				sr_size = dhd->sssr_reg_info->rev1.vasip_regs.vasip_sr_size;
1268 			} else if ((dhd->sssr_reg_info->rev1.length > OFFSETOF(sssr_reg_info_v1_t,
1269 				dig_mem_info)) && dhd->sssr_reg_info->rev1.
1270 				dig_mem_info.dig_sr_addr) {
1271 				alloc_sssr = TRUE;
1272 				sr_size = dhd->sssr_reg_info->rev1.dig_mem_info.dig_sr_size;
1273 			}
1274 			break;
1275 		case SSSR_REG_INFO_VER_0 :
1276 			if (dhd->sssr_reg_info->rev0.vasip_regs.vasip_sr_size) {
1277 				alloc_sssr = TRUE;
1278 				sr_size = dhd->sssr_reg_info->rev0.vasip_regs.vasip_sr_size;
1279 			}
1280 			break;
1281 		default :
1282 			DHD_ERROR(("invalid sssr_reg_ver"));
1283 			return BCME_UNSUPPORTED;
1284 	}
1285 
1286 	if (alloc_sssr) {
1287 		dhd->sssr_dig_buf_after = (uint32 *)(dhd->sssr_mempool + mempool_used);
1288 		mempool_used += sr_size;
1289 
1290 #ifdef DHD_SSSR_DUMP_BEFORE_SR
1291 		/* DIG dump before suspend is not applicable. */
1292 		dhd->sssr_dig_buf_before = NULL;
1293 #endif /* DHD_SSSR_DUMP_BEFORE_SR */
1294 	}
1295 
1296 	dhd->sssr_inited = TRUE;
1297 
1298 	return BCME_OK;
1299 
1300 }
1301 
1302 void
dhd_sssr_dump_deinit(dhd_pub_t * dhd)1303 dhd_sssr_dump_deinit(dhd_pub_t *dhd)
1304 {
1305 	int i;
1306 
1307 	dhd->sssr_inited = FALSE;
1308 	/* init all pointers to NULL */
1309 	for (i = 0; i < MAX_NUM_D11_CORES_WITH_SCAN; i++) {
1310 #ifdef DHD_SSSR_DUMP_BEFORE_SR
1311 		dhd->sssr_d11_before[i] = NULL;
1312 #endif /* DHD_SSSR_DUMP_BEFORE_SR */
1313 		dhd->sssr_d11_after[i] = NULL;
1314 	}
1315 #ifdef DHD_SSSR_DUMP_BEFORE_SR
1316 	dhd->sssr_dig_buf_before = NULL;
1317 #endif /* DHD_SSSR_DUMP_BEFORE_SR */
1318 	dhd->sssr_dig_buf_after = NULL;
1319 
1320 	return;
1321 }
1322 
1323 void
dhd_sssr_print_filepath(dhd_pub_t * dhd,char * path)1324 dhd_sssr_print_filepath(dhd_pub_t *dhd, char *path)
1325 {
1326 	bool print_info = FALSE;
1327 	int dump_mode;
1328 
1329 	if (!dhd || !path) {
1330 		DHD_ERROR(("%s: dhd or memdump_path is NULL\n",
1331 			__FUNCTION__));
1332 		return;
1333 	}
1334 
1335 	if (!dhd->sssr_dump_collected) {
1336 		/* SSSR dump is not collected */
1337 		return;
1338 	}
1339 
1340 	dump_mode = dhd->sssr_dump_mode;
1341 
1342 	if (bcmstrstr(path, "core_0_before")) {
1343 		if (dhd->sssr_d11_outofreset[0] &&
1344 			dump_mode == SSSR_DUMP_MODE_SSSR) {
1345 			print_info = TRUE;
1346 		}
1347 	} else if (bcmstrstr(path, "core_0_after")) {
1348 		if (dhd->sssr_d11_outofreset[0]) {
1349 			print_info = TRUE;
1350 		}
1351 	} else if (bcmstrstr(path, "core_1_before")) {
1352 		if (dhd->sssr_d11_outofreset[1] &&
1353 			dump_mode == SSSR_DUMP_MODE_SSSR) {
1354 			print_info = TRUE;
1355 		}
1356 	} else if (bcmstrstr(path, "core_1_after")) {
1357 		if (dhd->sssr_d11_outofreset[1]) {
1358 			print_info = TRUE;
1359 		}
1360 	} else if (bcmstrstr(path, "core_2_before")) {
1361 		if (dhd->sssr_d11_outofreset[2] &&
1362 			dump_mode == SSSR_DUMP_MODE_SSSR) {
1363 			print_info = TRUE;
1364 		}
1365 	} else if (bcmstrstr(path, "core_2_after")) {
1366 		if (dhd->sssr_d11_outofreset[2]) {
1367 			print_info = TRUE;
1368 		}
1369 	} else {
1370 		print_info = TRUE;
1371 	}
1372 
1373 	if (print_info) {
1374 		DHD_ERROR(("%s: file_path = %s%s\n", __FUNCTION__,
1375 			path, FILE_NAME_HAL_TAG));
1376 	}
1377 }
1378 #endif /* DHD_SSSR_DUMP */
1379 
1380 #ifdef DHD_SDTC_ETB_DUMP
1381 /*
1382  * sdtc: system debug trace controller
1383  * etb: embedded trace buf
1384  */
1385 void
dhd_sdtc_etb_init(dhd_pub_t * dhd)1386 dhd_sdtc_etb_init(dhd_pub_t *dhd)
1387 {
1388 	bcm_iov_buf_t *iov_req = NULL;
1389 	etb_addr_info_t *p_etb_addr_info = NULL;
1390 	bcm_iov_buf_t *iov_resp = NULL;
1391 	uint8 *buf = NULL;
1392 	int ret = 0;
1393 	uint16 iovlen = 0;
1394 	uint16 version = 0;
1395 
1396 	BCM_REFERENCE(p_etb_addr_info);
1397 	dhd->sdtc_etb_inited = FALSE;
1398 
1399 	iov_req = MALLOCZ(dhd->osh, WLC_IOCTL_SMLEN);
1400 	if (iov_req == NULL) {
1401 		DHD_ERROR(("%s: Failed to alloc buffer for iovar request\n", __FUNCTION__));
1402 		goto exit;
1403 	}
1404 
1405 	buf = MALLOCZ(dhd->osh, WLC_IOCTL_MAXLEN);
1406 	if (buf == NULL) {
1407 		DHD_ERROR(("%s: Failed to alloc buffer for iovar response\n", __FUNCTION__));
1408 		goto exit;
1409 	}
1410 
1411 	/* fill header */
1412 	iov_req->version = WL_SDTC_IOV_VERSION;
1413 	iov_req->id = WL_SDTC_CMD_ETB_INFO;
1414 	iov_req->len = sizeof(etb_addr_info_t);
1415 	iovlen = OFFSETOF(bcm_iov_buf_t, data) + iov_req->len;
1416 
1417 	ret = dhd_iovar(dhd, 0, "sdtc", (char *)iov_req, iovlen,
1418 		(char *)buf, WLC_IOCTL_MAXLEN, FALSE);
1419 	if (ret < 0) {
1420 		DHD_ERROR(("%s failed to get sdtc etb_info %d\n", __FUNCTION__, ret));
1421 		goto exit;
1422 	}
1423 
1424 	version = dtoh16(*(uint16 *)buf);
1425 	/* Check for version */
1426 	if (version != WL_SDTC_IOV_VERSION) {
1427 		DHD_ERROR(("%s WL_SDTC_IOV_VERSION mis match\n", __FUNCTION__));
1428 		goto exit;
1429 	}
1430 	iov_resp = (bcm_iov_buf_t *)buf;
1431 	if (iov_resp->id == iov_req->id) {
1432 		p_etb_addr_info = (etb_addr_info_t*)iov_resp->data;
1433 		dhd->etb_addr_info.version = p_etb_addr_info->version;
1434 		dhd->etb_addr_info.len = p_etb_addr_info->len;
1435 		dhd->etb_addr_info.etbinfo_addr = p_etb_addr_info->etbinfo_addr;
1436 
1437 		DHD_ERROR(("%s etb_addr_info: ver:%d, len:%d, addr:0x%x\n", __FUNCTION__,
1438 			dhd->etb_addr_info.version, dhd->etb_addr_info.len,
1439 			dhd->etb_addr_info.etbinfo_addr));
1440 	} else {
1441 		DHD_ERROR(("%s Unknown CMD-ID (%d) as  response for request ID %d\n",
1442 			__FUNCTION__, iov_resp->id, iov_req->id));
1443 		goto exit;
1444 	}
1445 
1446 	/* since all the requirements for SDTC and ETB are met mark the capability as TRUE */
1447 	dhd->sdtc_etb_inited = TRUE;
1448 	DHD_ERROR(("%s sdtc_etb_inited: %d\n", __FUNCTION__, dhd->sdtc_etb_inited));
1449 exit:
1450 	if (iov_req) {
1451 		MFREE(dhd->osh, iov_req, WLC_IOCTL_SMLEN);
1452 	}
1453 	if (buf) {
1454 		MFREE(dhd->osh, buf, WLC_IOCTL_MAXLEN);
1455 	}
1456 	return;
1457 }
1458 
1459 void
dhd_sdtc_etb_deinit(dhd_pub_t * dhd)1460 dhd_sdtc_etb_deinit(dhd_pub_t *dhd)
1461 {
1462 	dhd->sdtc_etb_inited = FALSE;
1463 }
1464 
1465 int
dhd_sdtc_etb_mempool_init(dhd_pub_t * dhd)1466 dhd_sdtc_etb_mempool_init(dhd_pub_t *dhd)
1467 {
1468 	dhd->sdtc_etb_mempool = (uint8 *) MALLOCZ(dhd->osh, DHD_SDTC_ETB_MEMPOOL_SIZE);
1469 	if (dhd->sdtc_etb_mempool == NULL) {
1470 		DHD_ERROR(("%s: MALLOC of sdtc_etb_mempool failed\n",
1471 			__FUNCTION__));
1472 		return BCME_ERROR;
1473 	}
1474 	return BCME_OK;
1475 }
1476 
1477 void
dhd_sdtc_etb_mempool_deinit(dhd_pub_t * dhd)1478 dhd_sdtc_etb_mempool_deinit(dhd_pub_t *dhd)
1479 {
1480 	if (dhd->sdtc_etb_mempool) {
1481 		MFREE(dhd->osh, dhd->sdtc_etb_mempool, DHD_SDTC_ETB_MEMPOOL_SIZE);
1482 		dhd->sdtc_etb_mempool = NULL;
1483 	}
1484 }
1485 #endif /* DHD_SDTC_ETB_DUMP */
1486 
1487 #ifdef DHD_FW_COREDUMP
dhd_get_fwdump_buf(dhd_pub_t * dhd_pub,uint32 length)1488 void* dhd_get_fwdump_buf(dhd_pub_t *dhd_pub, uint32 length)
1489 {
1490 	if (!dhd_pub->soc_ram) {
1491 #if defined(CONFIG_DHD_USE_STATIC_BUF) && defined(DHD_USE_STATIC_MEMDUMP)
1492 		dhd_pub->soc_ram = (uint8*)DHD_OS_PREALLOC(dhd_pub,
1493 			DHD_PREALLOC_MEMDUMP_RAM, length);
1494 #else
1495 		dhd_pub->soc_ram = (uint8*) MALLOC(dhd_pub->osh, length);
1496 
1497 		if ((dhd_pub->soc_ram == NULL) && CAN_SLEEP()) {
1498 			DHD_ERROR(("%s: Try to allocate virtual memory for fw crash snap shot.\n",
1499 				__FUNCTION__));
1500 			dhd_pub->soc_ram = (uint8*) VMALLOC(dhd_pub->osh, length);
1501 		}
1502 #endif /* CONFIG_DHD_USE_STATIC_BUF && DHD_USE_STATIC_MEMDUMP */
1503 	}
1504 
1505 	if (dhd_pub->soc_ram == NULL) {
1506 		DHD_ERROR(("%s: Failed to allocate memory for fw crash snap shot.\n",
1507 			__FUNCTION__));
1508 		dhd_pub->soc_ram_length = 0;
1509 	} else {
1510 		memset(dhd_pub->soc_ram, 0, length);
1511 		dhd_pub->soc_ram_length = length;
1512 	}
1513 
1514 	/* soc_ram free handled in dhd_{free,clear} */
1515 	return dhd_pub->soc_ram;
1516 }
1517 #endif /* DHD_FW_COREDUMP */
1518 
1519 /* to NDIS developer, the structure dhd_common is redundant,
1520  * please do NOT merge it back from other branches !!!
1521  */
1522 
1523 int
dhd_common_socram_dump(dhd_pub_t * dhdp)1524 dhd_common_socram_dump(dhd_pub_t *dhdp)
1525 {
1526 #ifdef BCMDBUS
1527 	return 0;
1528 #else
1529 	return dhd_socram_dump(dhdp->bus);
1530 #endif /* BCMDBUS */
1531 }
1532 
1533 int
dhd_dump(dhd_pub_t * dhdp,char * buf,int buflen)1534 dhd_dump(dhd_pub_t *dhdp, char *buf, int buflen)
1535 {
1536 	struct bcmstrbuf b;
1537 	struct bcmstrbuf *strbuf = &b;
1538 #ifdef DHD_MEM_STATS
1539 	uint64 malloc_mem = 0;
1540 	uint64 total_txpath_mem = 0;
1541 	uint64 txpath_bkpq_len = 0;
1542 	uint64 txpath_bkpq_mem = 0;
1543 	uint64 total_dhd_mem = 0;
1544 #endif /* DHD_MEM_STATS */
1545 
1546 	if (!dhdp || !dhdp->prot || !buf) {
1547 		return BCME_ERROR;
1548 	}
1549 
1550 	bcm_binit(strbuf, buf, buflen);
1551 
1552 	/* Base DHD info */
1553 	bcm_bprintf(strbuf, "%s\n", dhd_version);
1554 	bcm_bprintf(strbuf, "\n");
1555 	bcm_bprintf(strbuf, "pub.up %d pub.txoff %d pub.busstate %d\n",
1556 	            dhdp->up, dhdp->txoff, dhdp->busstate);
1557 	bcm_bprintf(strbuf, "pub.hdrlen %u pub.maxctl %u pub.rxsz %u\n",
1558 	            dhdp->hdrlen, dhdp->maxctl, dhdp->rxsz);
1559 	bcm_bprintf(strbuf, "pub.iswl %d pub.drv_version %ld pub.mac "MACDBG"\n",
1560 	            dhdp->iswl, dhdp->drv_version, MAC2STRDBG(&dhdp->mac));
1561 	bcm_bprintf(strbuf, "pub.bcmerror %d tickcnt %u\n", dhdp->bcmerror, dhdp->tickcnt);
1562 
1563 	bcm_bprintf(strbuf, "dongle stats:\n");
1564 	bcm_bprintf(strbuf, "tx_packets %lu tx_bytes %lu tx_errors %lu tx_dropped %lu\n",
1565 	            dhdp->dstats.tx_packets, dhdp->dstats.tx_bytes,
1566 	            dhdp->dstats.tx_errors, dhdp->dstats.tx_dropped);
1567 	bcm_bprintf(strbuf, "rx_packets %lu rx_bytes %lu rx_errors %lu rx_dropped %lu\n",
1568 	            dhdp->dstats.rx_packets, dhdp->dstats.rx_bytes,
1569 	            dhdp->dstats.rx_errors, dhdp->dstats.rx_dropped);
1570 	bcm_bprintf(strbuf, "multicast %lu\n", dhdp->dstats.multicast);
1571 
1572 	bcm_bprintf(strbuf, "bus stats:\n");
1573 	bcm_bprintf(strbuf, "tx_packets %lu  tx_dropped %lu tx_multicast %lu tx_errors %lu\n",
1574 	            dhdp->tx_packets, dhdp->tx_dropped, dhdp->tx_multicast, dhdp->tx_errors);
1575 	bcm_bprintf(strbuf, "tx_ctlpkts %lu tx_ctlerrs %lu\n",
1576 	            dhdp->tx_ctlpkts, dhdp->tx_ctlerrs);
1577 	bcm_bprintf(strbuf, "rx_packets %lu rx_multicast %lu rx_errors %lu \n",
1578 	            dhdp->rx_packets, dhdp->rx_multicast, dhdp->rx_errors);
1579 	bcm_bprintf(strbuf, "rx_ctlpkts %lu rx_ctlerrs %lu rx_dropped %lu\n",
1580 	            dhdp->rx_ctlpkts, dhdp->rx_ctlerrs, dhdp->rx_dropped);
1581 	bcm_bprintf(strbuf, "rx_readahead_cnt %lu tx_realloc %lu\n",
1582 	            dhdp->rx_readahead_cnt, dhdp->tx_realloc);
1583 	bcm_bprintf(strbuf, "tx_pktgetfail %lu rx_pktgetfail %lu\n",
1584 	            dhdp->tx_pktgetfail, dhdp->rx_pktgetfail);
1585 	bcm_bprintf(strbuf, "tx_big_packets %lu\n",
1586 	            dhdp->tx_big_packets);
1587 	bcm_bprintf(strbuf, "\n");
1588 #ifdef DMAMAP_STATS
1589 	/* Add DMA MAP info */
1590 	bcm_bprintf(strbuf, "DMA MAP stats: \n");
1591 	bcm_bprintf(strbuf, "txdata: %lu size: %luK, rxdata: %lu size: %luK\n",
1592 			dhdp->dma_stats.txdata, KB(dhdp->dma_stats.txdata_sz),
1593 			dhdp->dma_stats.rxdata, KB(dhdp->dma_stats.rxdata_sz));
1594 #ifndef IOCTLRESP_USE_CONSTMEM
1595 	bcm_bprintf(strbuf, "IOCTL RX: %lu size: %luK ,",
1596 			dhdp->dma_stats.ioctl_rx, KB(dhdp->dma_stats.ioctl_rx_sz));
1597 #endif /* !IOCTLRESP_USE_CONSTMEM */
1598 	bcm_bprintf(strbuf, "EVENT RX: %lu size: %luK, INFO RX: %lu size: %luK, "
1599 			"TSBUF RX: %lu size %luK\n",
1600 			dhdp->dma_stats.event_rx, KB(dhdp->dma_stats.event_rx_sz),
1601 			dhdp->dma_stats.info_rx, KB(dhdp->dma_stats.info_rx_sz),
1602 			dhdp->dma_stats.tsbuf_rx, KB(dhdp->dma_stats.tsbuf_rx_sz));
1603 	bcm_bprintf(strbuf, "Total : %luK \n",
1604 			KB(dhdp->dma_stats.txdata_sz + dhdp->dma_stats.rxdata_sz +
1605 			dhdp->dma_stats.ioctl_rx_sz + dhdp->dma_stats.event_rx_sz +
1606 			dhdp->dma_stats.tsbuf_rx_sz));
1607 #endif /* DMAMAP_STATS */
1608 	bcm_bprintf(strbuf, "dhd_induce_error : %u\n", dhdp->dhd_induce_error);
1609 	/* Add any prot info */
1610 	dhd_prot_dump(dhdp, strbuf);
1611 	bcm_bprintf(strbuf, "\n");
1612 
1613 	/* Add any bus info */
1614 	dhd_bus_dump(dhdp, strbuf);
1615 #if defined(BCM_ROUTER_DHD) && defined(HNDCTF)
1616 	/* Add ctf info */
1617 	dhd_ctf_dump(dhdp, strbuf);
1618 #endif /* BCM_ROUTER_DHD && HNDCTF */
1619 
1620 #if defined(DHD_LB_STATS)
1621 	dhd_lb_stats_dump(dhdp, strbuf);
1622 #endif /* DHD_LB_STATS */
1623 
1624 #ifdef DHD_MEM_STATS
1625 
1626 	malloc_mem = MALLOCED(dhdp->osh);
1627 
1628 	txpath_bkpq_len = dhd_active_tx_flowring_bkpq_len(dhdp);
1629 	/*
1630 	 * Instead of traversing the entire queue to find the skbs length,
1631 	 * considering MAX_MTU_SZ as lenth of each skb.
1632 	 */
1633 	txpath_bkpq_mem = (txpath_bkpq_len* MAX_MTU_SZ);
1634 	total_txpath_mem = dhdp->txpath_mem + txpath_bkpq_mem;
1635 
1636 	bcm_bprintf(strbuf, "\nDHD malloc memory_usage: %llubytes %lluKB\n",
1637 		malloc_mem, (malloc_mem / 1024));
1638 
1639 	bcm_bprintf(strbuf, "\nDHD tx-bkpq len: %llu memory_usage: %llubytes %lluKB\n",
1640 		txpath_bkpq_len, txpath_bkpq_mem, (txpath_bkpq_mem / 1024));
1641 	bcm_bprintf(strbuf, "DHD tx-path memory_usage: %llubytes %lluKB\n",
1642 		total_txpath_mem, (total_txpath_mem / 1024));
1643 
1644 	total_dhd_mem = malloc_mem + total_txpath_mem;
1645 #if defined(DHD_LB_STATS)
1646 	total_dhd_mem += dhd_lb_mem_usage(dhdp, strbuf);
1647 #endif /* DHD_LB_STATS */
1648 	bcm_bprintf(strbuf, "\nDHD Totoal memory_usage: %llubytes %lluKB \n",
1649 		total_dhd_mem, (total_dhd_mem / 1024));
1650 #endif /* DHD_MEM_STATS */
1651 #if defined(DHD_LB_STATS)
1652 	bcm_bprintf(strbuf, "\nlb_rxp_stop_thr_hitcnt: %llu lb_rxp_strt_thr_hitcnt: %llu\n",
1653 		dhdp->lb_rxp_stop_thr_hitcnt, dhdp->lb_rxp_strt_thr_hitcnt);
1654 	bcm_bprintf(strbuf, "\nlb_rxp_napi_sched_cnt: %llu lb_rxp_napi_complete_cnt: %llu\n",
1655 		dhdp->lb_rxp_napi_sched_cnt, dhdp->lb_rxp_napi_complete_cnt);
1656 #endif /* DHD_LB_STATS */
1657 
1658 #if defined(DHD_MQ) && defined(DHD_MQ_STATS)
1659 	dhd_mqstats_dump(dhdp, strbuf);
1660 #endif
1661 
1662 #ifdef DHD_WET
1663 	if (dhd_get_wet_mode(dhdp)) {
1664 		bcm_bprintf(strbuf, "Wet Dump:\n");
1665 		dhd_wet_dump(dhdp, strbuf);
1666 		}
1667 #endif /* DHD_WET */
1668 
1669 	DHD_ERROR(("%s bufsize: %d free: %d", __FUNCTION__, buflen, strbuf->size));
1670 	/* return remaining buffer length */
1671 	return (!strbuf->size ? BCME_BUFTOOSHORT : strbuf->size);
1672 }
1673 
1674 void
dhd_dump_to_kernelog(dhd_pub_t * dhdp)1675 dhd_dump_to_kernelog(dhd_pub_t *dhdp)
1676 {
1677 	char buf[512];
1678 
1679 	DHD_ERROR(("F/W version: %s\n", fw_version));
1680 	bcm_bprintf_bypass = TRUE;
1681 	dhd_dump(dhdp, buf, sizeof(buf));
1682 	bcm_bprintf_bypass = FALSE;
1683 }
1684 
1685 int
dhd_wl_ioctl_cmd(dhd_pub_t * dhd_pub,int cmd,void * arg,int len,uint8 set,int ifidx)1686 dhd_wl_ioctl_cmd(dhd_pub_t *dhd_pub, int cmd, void *arg, int len, uint8 set, int ifidx)
1687 {
1688 	wl_ioctl_t ioc;
1689 
1690 	ioc.cmd = cmd;
1691 	ioc.buf = arg;
1692 	ioc.len = len;
1693 	ioc.set = set;
1694 
1695 	return dhd_wl_ioctl(dhd_pub, ifidx, &ioc, arg, len);
1696 }
1697 
1698 int
dhd_wl_ioctl_get_intiovar(dhd_pub_t * dhd_pub,char * name,uint * pval,int cmd,uint8 set,int ifidx)1699 dhd_wl_ioctl_get_intiovar(dhd_pub_t *dhd_pub, char *name, uint *pval,
1700 	int cmd, uint8 set, int ifidx)
1701 {
1702 	char iovbuf[WLC_IOCTL_SMLEN];
1703 	int ret = -1;
1704 
1705 	memset(iovbuf, 0, sizeof(iovbuf));
1706 	if (bcm_mkiovar(name, NULL, 0, iovbuf, sizeof(iovbuf))) {
1707 		ret = dhd_wl_ioctl_cmd(dhd_pub, cmd, iovbuf, sizeof(iovbuf), set, ifidx);
1708 		if (!ret) {
1709 			*pval = ltoh32(*((uint*)iovbuf));
1710 		} else {
1711 			DHD_ERROR(("%s: get int iovar %s failed, ERR %d\n",
1712 				__FUNCTION__, name, ret));
1713 		}
1714 	} else {
1715 		DHD_ERROR(("%s: mkiovar %s failed\n",
1716 			__FUNCTION__, name));
1717 	}
1718 
1719 	return ret;
1720 }
1721 
1722 int
dhd_wl_ioctl_set_intiovar(dhd_pub_t * dhd_pub,char * name,uint val,int cmd,uint8 set,int ifidx)1723 dhd_wl_ioctl_set_intiovar(dhd_pub_t *dhd_pub, char *name, uint val,
1724 	int cmd, uint8 set, int ifidx)
1725 {
1726 	char iovbuf[WLC_IOCTL_SMLEN];
1727 	int ret = -1;
1728 	int lval = htol32(val);
1729 	uint len;
1730 
1731 	len = bcm_mkiovar(name, (char*)&lval, sizeof(lval), iovbuf, sizeof(iovbuf));
1732 
1733 	if (len) {
1734 		ret = dhd_wl_ioctl_cmd(dhd_pub, cmd, iovbuf, len, set, ifidx);
1735 		if (ret) {
1736 			DHD_ERROR(("%s: set int iovar %s failed, ERR %d\n",
1737 				__FUNCTION__, name, ret));
1738 		}
1739 	} else {
1740 		DHD_ERROR(("%s: mkiovar %s failed\n",
1741 			__FUNCTION__, name));
1742 	}
1743 
1744 	return ret;
1745 }
1746 
1747 static struct ioctl2str_s {
1748 	uint32 ioctl;
1749 	char *name;
1750 } ioctl2str_array[] = {
1751 	{WLC_UP, "UP"},
1752 	{WLC_DOWN, "DOWN"},
1753 	{WLC_SET_PROMISC, "SET_PROMISC"},
1754 	{WLC_SET_INFRA, "SET_INFRA"},
1755 	{WLC_SET_AUTH, "SET_AUTH"},
1756 	{WLC_SET_SSID, "SET_SSID"},
1757 	{WLC_RESTART, "RESTART"},
1758 	{WLC_SET_CHANNEL, "SET_CHANNEL"},
1759 	{WLC_SET_RATE_PARAMS, "SET_RATE_PARAMS"},
1760 	{WLC_SET_KEY, "SET_KEY"},
1761 	{WLC_SCAN, "SCAN"},
1762 	{WLC_DISASSOC, "DISASSOC"},
1763 	{WLC_REASSOC, "REASSOC"},
1764 	{WLC_SET_COUNTRY, "SET_COUNTRY"},
1765 	{WLC_SET_WAKE, "SET_WAKE"},
1766 	{WLC_SET_SCANSUPPRESS, "SET_SCANSUPPRESS"},
1767 	{WLC_SCB_DEAUTHORIZE, "SCB_DEAUTHORIZE"},
1768 	{WLC_SET_WSEC, "SET_WSEC"},
1769 	{WLC_SET_INTERFERENCE_MODE, "SET_INTERFERENCE_MODE"},
1770 	{WLC_SET_RADAR, "SET_RADAR"},
1771 	{0, NULL}
1772 };
1773 
1774 static char *
ioctl2str(uint32 ioctl)1775 ioctl2str(uint32 ioctl)
1776 {
1777 	struct ioctl2str_s *p = ioctl2str_array;
1778 
1779 	while (p->name != NULL) {
1780 		if (p->ioctl == ioctl) {
1781 			return p->name;
1782 		}
1783 		p++;
1784 	}
1785 
1786 	return "";
1787 }
1788 
1789 /**
1790  * @param ioc          IO control struct, members are partially used by this function.
1791  * @param buf [inout]  Contains parameters to send to dongle, contains dongle response on return.
1792  * @param len          Maximum number of bytes that dongle is allowed to write into 'buf'.
1793  */
1794 int
dhd_wl_ioctl(dhd_pub_t * dhd_pub,int ifidx,wl_ioctl_t * ioc,void * buf,int len)1795 dhd_wl_ioctl(dhd_pub_t *dhd_pub, int ifidx, wl_ioctl_t *ioc, void *buf, int len)
1796 {
1797 	int ret = BCME_ERROR;
1798 	unsigned long flags;
1799 #ifdef DUMP_IOCTL_IOV_LIST
1800 	dhd_iov_li_t *iov_li;
1801 #endif /* DUMP_IOCTL_IOV_LIST */
1802 #ifdef REPORT_FATAL_TIMEOUTS
1803 	wl_escan_params_t *eparams;
1804 	uint8 *buf_ptr = (uint8 *)buf;
1805 	uint16 action = 0;
1806 #endif /* REPORT_FATAL_TIMEOUTS */
1807 	int hostsleep_set = 0;
1808 	int hostsleep_val = 0;
1809 
1810 	if (dhd_query_bus_erros(dhd_pub)) {
1811 		return -ENODEV;
1812 	}
1813 
1814 #ifdef DHD_PCIE_NATIVE_RUNTIMEPM
1815 	DHD_OS_WAKE_LOCK(dhd_pub);
1816 	if (pm_runtime_get_sync(dhd_bus_to_dev(dhd_pub->bus)) < 0) {
1817 		DHD_RPM(("%s: pm_runtime_get_sync error. \n", __FUNCTION__));
1818 		DHD_OS_WAKE_UNLOCK(dhd_pub);
1819 		return BCME_ERROR;
1820 	}
1821 #endif /* DHD_PCIE_NATIVE_RUNTIMEPM */
1822 
1823 #ifdef KEEPIF_ON_DEVICE_RESET
1824 		if (ioc->cmd == WLC_GET_VAR) {
1825 			dbus_config_t config;
1826 			config.general_param = 0;
1827 			if (buf) {
1828 				if (!strcmp(buf, "wowl_activate")) {
1829 					 /* 1 (TRUE) after decreased by 1 */
1830 					config.general_param = 2;
1831 				} else if (!strcmp(buf, "wowl_clear")) {
1832 					 /* 0 (FALSE) after decreased by 1 */
1833 					config.general_param = 1;
1834 				}
1835 			}
1836 			if (config.general_param) {
1837 				config.config_id = DBUS_CONFIG_ID_KEEPIF_ON_DEVRESET;
1838 				config.general_param--;
1839 				dbus_set_config(dhd_pub->dbus, &config);
1840 			}
1841 		}
1842 #endif /* KEEPIF_ON_DEVICE_RESET */
1843 
1844 	if (dhd_os_proto_block(dhd_pub))
1845 	{
1846 #ifdef DHD_LOG_DUMP
1847 		int slen, val, lval, min_len;
1848 		char *msg, tmp[64];
1849 
1850 		/* WLC_GET_VAR */
1851 		if (ioc->cmd == WLC_GET_VAR && buf) {
1852 			min_len = MIN(sizeof(tmp) - 1, strlen(buf));
1853 			memset(tmp, 0, sizeof(tmp));
1854 			bcopy(buf, tmp, min_len);
1855 			tmp[min_len] = '\0';
1856 		}
1857 #endif /* DHD_LOG_DUMP */
1858 
1859 #ifdef DHD_DISCONNECT_TRACE
1860 		if (WLC_DISASSOC == ioc->cmd || WLC_DOWN == ioc->cmd ||
1861 			WLC_DISASSOC_MYAP == ioc->cmd) {
1862 			DHD_ERROR(("IOCTL Disconnect WiFi: %d\n", ioc->cmd));
1863 		}
1864 #endif /* HW_DISCONNECT_TRACE */
1865 		/* logging of iovars that are send to the dongle, ./dhd msglevel +iovar */
1866 		if (ioc->set == TRUE) {
1867 			char *pars = (char *)buf; // points at user buffer
1868 			if (ioc->cmd == WLC_SET_VAR && buf) {
1869 				DHD_DNGL_IOVAR_SET(("iovar:%d: set %s", ifidx, pars));
1870 				if (ioc->len > 1 + sizeof(uint32)) {
1871 					// skip iovar name:
1872 					pars += strnlen(pars, ioc->len - 1 - sizeof(uint32));
1873 					pars++;               // skip NULL character
1874 				}
1875 			} else {
1876 				DHD_DNGL_IOVAR_SET(("ioctl:%d: set %d %s",
1877 					ifidx, ioc->cmd, ioctl2str(ioc->cmd)));
1878 			}
1879 			if (pars != NULL) {
1880 				DHD_DNGL_IOVAR_SET((" 0x%x\n", *(uint32*)pars));
1881 			} else {
1882 				DHD_DNGL_IOVAR_SET((" NULL\n"));
1883 			}
1884 		}
1885 
1886 		DHD_LINUX_GENERAL_LOCK(dhd_pub, flags);
1887 		if (DHD_BUS_CHECK_DOWN_OR_DOWN_IN_PROGRESS(dhd_pub)) {
1888 #ifdef DHD_EFI
1889 			DHD_INFO(("%s: returning as busstate=%d\n",
1890 				__FUNCTION__, dhd_pub->busstate));
1891 #else
1892 			DHD_INFO(("%s: returning as busstate=%d\n",
1893 				__FUNCTION__, dhd_pub->busstate));
1894 #endif /* DHD_EFI */
1895 			DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
1896 			dhd_os_proto_unblock(dhd_pub);
1897 			return -ENODEV;
1898 		}
1899 		DHD_BUS_BUSY_SET_IN_IOVAR(dhd_pub);
1900 		DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
1901 
1902 #ifdef DHD_PCIE_RUNTIMEPM
1903 		dhdpcie_runtime_bus_wake(dhd_pub, TRUE, dhd_wl_ioctl);
1904 #endif /* DHD_PCIE_RUNTIMEPM */
1905 
1906 		DHD_LINUX_GENERAL_LOCK(dhd_pub, flags);
1907 		if (DHD_BUS_CHECK_SUSPEND_OR_ANY_SUSPEND_IN_PROGRESS(dhd_pub) ||
1908 			dhd_pub->dhd_induce_error == DHD_INDUCE_IOCTL_SUSPEND_ERROR) {
1909 			DHD_ERROR(("%s: bus is in suspend(%d) or suspending(0x%x) state!!\n",
1910 				__FUNCTION__, dhd_pub->busstate, dhd_pub->dhd_bus_busy_state));
1911 #ifdef DHD_SEND_HANG_IOCTL_SUSPEND_ERROR
1912 			ioctl_suspend_error++;
1913 			if (ioctl_suspend_error > MAX_IOCTL_SUSPEND_ERROR) {
1914 				dhd_pub->hang_reason = HANG_REASON_IOCTL_SUSPEND_ERROR;
1915 				dhd_os_send_hang_message(dhd_pub);
1916 				ioctl_suspend_error = 0;
1917 			}
1918 #endif /* DHD_SEND_HANG_IOCTL_SUSPEND_ERROR */
1919 			DHD_BUS_BUSY_CLEAR_IN_IOVAR(dhd_pub);
1920 			dhd_os_busbusy_wake(dhd_pub);
1921 			DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
1922 			dhd_os_proto_unblock(dhd_pub);
1923 			return -ENODEV;
1924 		}
1925 #ifdef DHD_SEND_HANG_IOCTL_SUSPEND_ERROR
1926 		ioctl_suspend_error = 0;
1927 #endif /* DHD_SEND_HANG_IOCTL_SUSPEND_ERROR */
1928 		DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
1929 
1930 #if defined(WL_WLC_SHIM)
1931 		{
1932 			struct wl_shim_node *shim = dhd_pub_shim(dhd_pub);
1933 
1934 			wl_io_pport_t io_pport;
1935 			io_pport.dhd_pub = dhd_pub;
1936 			io_pport.ifidx = ifidx;
1937 
1938 			ret = wl_shim_ioctl(shim, ioc, len, &io_pport);
1939 			if (ret != BCME_OK) {
1940 				DHD_TRACE(("%s: wl_shim_ioctl(%d) ERR %d\n",
1941 					__FUNCTION__, ioc->cmd, ret));
1942 			}
1943 		}
1944 #else
1945 #ifdef DUMP_IOCTL_IOV_LIST
1946 		if (ioc->cmd != WLC_GET_MAGIC && ioc->cmd != WLC_GET_VERSION && buf) {
1947 			if (!(iov_li = MALLOC(dhd_pub->osh, sizeof(*iov_li)))) {
1948 				DHD_ERROR(("iovar dump list item allocation Failed\n"));
1949 			} else {
1950 				iov_li->cmd = ioc->cmd;
1951 				if (buf)
1952 					bcopy((char *)buf, iov_li->buff, strlen((char *)buf)+1);
1953 				dhd_iov_li_append(dhd_pub, &dhd_pub->dump_iovlist_head,
1954 						&iov_li->list);
1955 			}
1956 		}
1957 #endif /* DUMP_IOCTL_IOV_LIST */
1958 
1959 #ifdef REPORT_FATAL_TIMEOUTS
1960 		/* fill in the sync_id to ensure that the scan timeout is always for the
1961 		* current running escan in the FW - the wl app does not fill in an
1962 		* incrementing number for sync_id, it only fills in a random number which
1963 		* increases the chance of 2 consecutive escans having the same sync id
1964 		* This should happen here after dhd_proto_block()
1965 		* is called, so that sync_id does not
1966 		* get incremented if 2 consecutive escans are fired in quick succession
1967 		*/
1968 		if ((ioc->cmd == WLC_SET_VAR &&
1969 				buf != NULL &&
1970 				strcmp("escan", buf) == 0)) {
1971 			eparams = (wl_escan_params_t *) (buf_ptr + strlen("escan") + 1);
1972 			action = dtoh16(eparams->action);
1973 			if (action == WL_SCAN_ACTION_START) {
1974 				++dhd_pub->esync_id;
1975 				/* sync id of 0 is not used for escan,
1976 				* it is used to indicate
1977 				* a normal scan timer is running, so as
1978 				* to ensure that escan abort event
1979 				* does not cancel a normal scan timeout
1980 				*/
1981 				if (dhd_pub->esync_id == 0)
1982 					++dhd_pub->esync_id;
1983 				DHD_INFO(("%s:escan sync id set to = %u \n",
1984 					__FUNCTION__, dhd_pub->esync_id));
1985 				eparams->sync_id = htod16(dhd_pub->esync_id);
1986 			}
1987 		}
1988 #endif /* REPORT_FATAL_TIMEOUTS */
1989 
1990 		if (dhd_conf_check_hostsleep(dhd_pub, ioc->cmd, ioc->buf, len,
1991 				&hostsleep_set, &hostsleep_val, &ret))
1992 			goto exit;
1993 		ret = dhd_prot_ioctl(dhd_pub, ifidx, ioc, buf, len);
1994 		dhd_conf_get_hostsleep(dhd_pub, hostsleep_set, hostsleep_val, ret);
1995 
1996 #ifdef DUMP_IOCTL_IOV_LIST
1997 		if (ret == -ETIMEDOUT) {
1998 			DHD_ERROR(("Last %d issued commands: Latest one is at bottom.\n",
1999 				IOV_LIST_MAX_LEN));
2000 			dhd_iov_li_print(&dhd_pub->dump_iovlist_head);
2001 		}
2002 #endif /* DUMP_IOCTL_IOV_LIST */
2003 #endif /* defined(WL_WLC_SHIM) */
2004 #ifdef WL_CFGVENDOR_SEND_HANG_EVENT
2005 		if (ret == -ETIMEDOUT) {
2006 			copy_hang_info_ioctl_timeout(dhd_pub, ifidx, ioc);
2007 		}
2008 #endif /* WL_CFGVENDOR_SEND_HANG_EVENT */
2009 #ifdef DHD_LOG_DUMP
2010 		if ((ioc->cmd == WLC_GET_VAR || ioc->cmd == WLC_SET_VAR) &&
2011 				buf != NULL) {
2012 			if (buf) {
2013 				lval = 0;
2014 				slen = strlen(buf) + 1;
2015 				msg = (char*)buf;
2016 				if (len >= slen + sizeof(lval)) {
2017 					if (ioc->cmd == WLC_GET_VAR) {
2018 						msg = tmp;
2019 						lval = *(int*)buf;
2020 					} else {
2021 						min_len = MIN(ioc->len - slen, sizeof(int));
2022 						bcopy((msg + slen), &lval, min_len);
2023 					}
2024 					if (!strncmp(msg, "cur_etheraddr",
2025 						strlen("cur_etheraddr"))) {
2026 						lval = 0;
2027 					}
2028 				}
2029 				DHD_IOVAR_MEM((
2030 					"%s: cmd: %d, msg: %s val: 0x%x,"
2031 					" len: %d, set: %d, txn-id: %d\n",
2032 					ioc->cmd == WLC_GET_VAR ?
2033 					"WLC_GET_VAR" : "WLC_SET_VAR",
2034 					ioc->cmd, msg, lval, ioc->len, ioc->set,
2035 					dhd_prot_get_ioctl_trans_id(dhd_pub)));
2036 			} else {
2037 				DHD_IOVAR_MEM(("%s: cmd: %d, len: %d, set: %d, txn-id: %d\n",
2038 					ioc->cmd == WLC_GET_VAR ? "WLC_GET_VAR" : "WLC_SET_VAR",
2039 					ioc->cmd, ioc->len, ioc->set,
2040 					dhd_prot_get_ioctl_trans_id(dhd_pub)));
2041 			}
2042 		} else {
2043 			slen = ioc->len;
2044 			if (buf != NULL && slen != 0) {
2045 				if (slen >= 4) {
2046 					val = *(int*)buf;
2047 				} else if (slen >= 2) {
2048 					val = *(short*)buf;
2049 				} else {
2050 					val = *(char*)buf;
2051 				}
2052 				/* Do not dump for WLC_GET_MAGIC and WLC_GET_VERSION */
2053 				if (ioc->cmd != WLC_GET_MAGIC && ioc->cmd != WLC_GET_VERSION) {
2054 					DHD_IOVAR_MEM(("WLC_IOCTL: cmd: %d, val: %d, len: %d, "
2055 						"set: %d\n", ioc->cmd, val, ioc->len, ioc->set));
2056 				}
2057 			} else {
2058 				DHD_IOVAR_MEM(("WLC_IOCTL: cmd: %d, buf is NULL\n", ioc->cmd));
2059 			}
2060 		}
2061 #endif /* DHD_LOG_DUMP */
2062 #if defined(OEM_ANDROID)
2063 		if (ret && dhd_pub->up) {
2064 			/* Send hang event only if dhd_open() was success */
2065 			dhd_os_check_hang(dhd_pub, ifidx, ret);
2066 		}
2067 
2068 		if (ret == -ETIMEDOUT && !dhd_pub->up) {
2069 			DHD_ERROR(("%s: 'resumed on timeout' error is "
2070 				"occurred before the interface does not"
2071 				" bring up\n", __FUNCTION__));
2072 		}
2073 #endif /* defined(OEM_ANDROID) */
2074 
2075 exit:
2076 		DHD_LINUX_GENERAL_LOCK(dhd_pub, flags);
2077 		DHD_BUS_BUSY_CLEAR_IN_IOVAR(dhd_pub);
2078 		dhd_os_busbusy_wake(dhd_pub);
2079 		DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
2080 
2081 #ifdef REPORT_FATAL_TIMEOUTS
2082 		if ((ret == BCME_OK && ioc->cmd == WLC_SET_VAR &&
2083 				buf != NULL &&
2084 				strcmp("escan", buf) == 0)) {
2085 			if (action == WL_SCAN_ACTION_START)
2086 				dhd_start_scan_timer(dhd_pub, TRUE);
2087 		}
2088 #endif /* REPORT_FATAL_TIMEOUTS */
2089 
2090 		dhd_os_proto_unblock(dhd_pub);
2091 
2092 #ifdef DETAIL_DEBUG_LOG_FOR_IOCTL
2093 		if (ret < 0) {
2094 			if ((ioc->cmd == WLC_GET_VAR || ioc->cmd == WLC_SET_VAR) &&
2095 					buf != NULL) {
2096 				if (ret == BCME_UNSUPPORTED || ret == BCME_NOTASSOCIATED) {
2097 					DHD_ERROR_MEM(("%s: %s: %s, %s\n",
2098 						__FUNCTION__, ioc->cmd == WLC_GET_VAR ?
2099 						"WLC_GET_VAR" : "WLC_SET_VAR",
2100 						buf? (char *)buf:"NO MESSAGE",
2101 						ret == BCME_UNSUPPORTED ? "UNSUPPORTED"
2102 						: "NOT ASSOCIATED"));
2103 				} else {
2104 					DHD_ERROR_MEM(("%s: %s: %s, ret = %d\n",
2105 						__FUNCTION__, ioc->cmd == WLC_GET_VAR ?
2106 						"WLC_GET_VAR" : "WLC_SET_VAR",
2107 						(char *)buf, ret));
2108 				}
2109 			} else {
2110 				if (ret == BCME_UNSUPPORTED || ret == BCME_NOTASSOCIATED) {
2111 					DHD_ERROR_MEM(("%s: WLC_IOCTL: cmd: %d, %s\n",
2112 						__FUNCTION__, ioc->cmd,
2113 						ret == BCME_UNSUPPORTED ? "UNSUPPORTED" :
2114 						"NOT ASSOCIATED"));
2115 				} else {
2116 					DHD_ERROR_MEM(("%s: WLC_IOCTL: cmd: %d, ret = %d\n",
2117 						__FUNCTION__, ioc->cmd, ret));
2118 				}
2119 			}
2120 		}
2121 #endif /* DETAIL_DEBUG_LOG_FOR_IOCTL */
2122 	}
2123 
2124 #ifdef DHD_PCIE_NATIVE_RUNTIMEPM
2125 	pm_runtime_mark_last_busy(dhd_bus_to_dev(dhd_pub->bus));
2126 	pm_runtime_put_autosuspend(dhd_bus_to_dev(dhd_pub->bus));
2127 
2128 	DHD_OS_WAKE_UNLOCK(dhd_pub);
2129 #endif /* DHD_PCIE_NATIVE_RUNTIMEPM */
2130 
2131 #ifdef WL_MONITOR
2132 	/* Intercept monitor ioctl here, add/del monitor if */
2133 	if (ret == BCME_OK && ioc->cmd == WLC_SET_MONITOR) {
2134 		int val = 0;
2135 		if (buf != NULL && len != 0) {
2136 			if (len >= 4) {
2137 				val = *(int*)buf;
2138 			} else if (len >= 2) {
2139 				val = *(short*)buf;
2140 			} else {
2141 				val = *(char*)buf;
2142 			}
2143 		}
2144 		dhd_set_monitor(dhd_pub, ifidx, val);
2145 	}
2146 #endif /* WL_MONITOR */
2147 
2148 	return ret;
2149 }
2150 
wl_get_port_num(wl_io_pport_t * io_pport)2151 uint wl_get_port_num(wl_io_pport_t *io_pport)
2152 {
2153 	return 0;
2154 }
2155 
2156 /* Get bssidx from iovar params
2157  * Input:   dhd_pub - pointer to dhd_pub_t
2158  *	    params  - IOVAR params
2159  * Output:  idx	    - BSS index
2160  *	    val	    - ponter to the IOVAR arguments
2161  */
2162 static int
dhd_iovar_parse_bssidx(dhd_pub_t * dhd_pub,const char * params,uint32 * idx,const char ** val)2163 dhd_iovar_parse_bssidx(dhd_pub_t *dhd_pub, const char *params, uint32 *idx, const char **val)
2164 {
2165 	char *prefix = "bsscfg:";
2166 	uint32	bssidx;
2167 
2168 	if (!(strncmp(params, prefix, strlen(prefix)))) {
2169 		/* per bss setting should be prefixed with 'bsscfg:' */
2170 		const char *p = params + strlen(prefix);
2171 
2172 		/* Skip Name */
2173 		while (*p != '\0')
2174 			p++;
2175 		/* consider null */
2176 		p = p + 1;
2177 		bcopy(p, &bssidx, sizeof(uint32));
2178 		/* Get corresponding dhd index */
2179 		bssidx = dhd_bssidx2idx(dhd_pub, htod32(bssidx));
2180 
2181 		if (bssidx >= DHD_MAX_IFS) {
2182 			DHD_ERROR(("%s Wrong bssidx provided\n", __FUNCTION__));
2183 			return BCME_ERROR;
2184 		}
2185 
2186 		/* skip bss idx */
2187 		p += sizeof(uint32);
2188 		*val = p;
2189 		*idx = bssidx;
2190 	} else {
2191 		DHD_ERROR(("%s: bad parameter for per bss iovar\n", __FUNCTION__));
2192 		return BCME_ERROR;
2193 	}
2194 
2195 	return BCME_OK;
2196 }
2197 
2198 #if defined(DHD_DEBUG) && defined(BCMDBUS)
2199 /* USB Device console input function */
dhd_bus_console_in(dhd_pub_t * dhd,uchar * msg,uint msglen)2200 int dhd_bus_console_in(dhd_pub_t *dhd, uchar *msg, uint msglen)
2201 {
2202 	DHD_TRACE(("%s \n", __FUNCTION__));
2203 
2204 	return dhd_iovar(dhd, 0, "cons", msg, msglen, NULL, 0, TRUE);
2205 
2206 }
2207 #endif /* DHD_DEBUG && BCMDBUS  */
2208 
2209 #ifdef DHD_DEBUG
2210 int
dhd_mem_debug(dhd_pub_t * dhd,uchar * msg,uint msglen)2211 dhd_mem_debug(dhd_pub_t *dhd, uchar *msg, uint msglen)
2212 {
2213 	unsigned long int_arg = 0;
2214 	char *p;
2215 	char *end_ptr = NULL;
2216 	dhd_dbg_mwli_t *mw_li;
2217 	dll_t *item, *next;
2218 	/* check if mwalloc, mwquery or mwfree was supplied arguement with space */
2219 	p = bcmstrstr((char *)msg, " ");
2220 	if (p != NULL) {
2221 		/* space should be converted to null as separation flag for firmware */
2222 		*p = '\0';
2223 		/* store the argument in int_arg */
2224 		int_arg = bcm_strtoul(p+1, &end_ptr, 10);
2225 	}
2226 
2227 	if (!p && !strcmp(msg, "query")) {
2228 		/* lets query the list inetrnally */
2229 		if (dll_empty(dll_head_p(&dhd->mw_list_head))) {
2230 			DHD_ERROR(("memwaste list is empty, call mwalloc < size > to allocate\n"));
2231 		} else {
2232 			for (item = dll_head_p(&dhd->mw_list_head);
2233 					!dll_end(&dhd->mw_list_head, item); item = next) {
2234 				next = dll_next_p(item);
2235 				mw_li = (dhd_dbg_mwli_t *)CONTAINEROF(item, dhd_dbg_mwli_t, list);
2236 				DHD_ERROR(("item: <id=%d, size=%d>\n", mw_li->id, mw_li->size));
2237 			}
2238 		}
2239 	} else if (p && end_ptr && (*end_ptr == '\0') && !strcmp(msg, "alloc")) {
2240 		int32 alloc_handle;
2241 		/* convert size into KB and append as integer */
2242 		*((int32 *)(p+1)) = int_arg*1024;
2243 		*(p+1+sizeof(int32)) = '\0';
2244 
2245 		/* recalculated length -> 5 bytes for "alloc" + 4 bytes for size +
2246 		 * 1 bytes for null caracter
2247 		 */
2248 		msglen = strlen(msg) + sizeof(int32) + 1;
2249 		if (dhd_wl_ioctl_cmd(dhd, WLC_GET_VAR, msg, msglen+1, FALSE, 0) < 0) {
2250 			DHD_ERROR(("IOCTL failed for memdebug alloc\n"));
2251 		}
2252 
2253 		/* returned allocated handle from dongle, basically address of the allocated unit */
2254 		alloc_handle = *((int32 *)msg);
2255 
2256 		/* add a node in the list with tuple <id, handle, size> */
2257 		if (alloc_handle == 0) {
2258 			DHD_ERROR(("Reuqested size could not be allocated\n"));
2259 		} else if (!(mw_li = MALLOC(dhd->osh, sizeof(*mw_li)))) {
2260 			DHD_ERROR(("mw list item allocation Failed\n"));
2261 		} else {
2262 			mw_li->id = dhd->mw_id++;
2263 			mw_li->handle = alloc_handle;
2264 			mw_li->size = int_arg;
2265 			/* append the node in the list */
2266 			dll_append(&dhd->mw_list_head, &mw_li->list);
2267 		}
2268 	} else if (p && end_ptr && (*end_ptr == '\0') && !strcmp(msg, "free")) {
2269 		/* inform dongle to free wasted chunk */
2270 		int handle = 0;
2271 		int size = 0;
2272 		for (item = dll_head_p(&dhd->mw_list_head);
2273 				!dll_end(&dhd->mw_list_head, item); item = next) {
2274 			next = dll_next_p(item);
2275 			mw_li = (dhd_dbg_mwli_t *)CONTAINEROF(item, dhd_dbg_mwli_t, list);
2276 
2277 			if (mw_li->id == (int)int_arg) {
2278 				handle = mw_li->handle;
2279 				size = mw_li->size;
2280 				dll_delete(item);
2281 				MFREE(dhd->osh, mw_li, sizeof(*mw_li));
2282 				if (dll_empty(dll_head_p(&dhd->mw_list_head))) {
2283 					/* reset the id */
2284 					dhd->mw_id = 0;
2285 				}
2286 			}
2287 		}
2288 		if (handle) {
2289 			int len;
2290 			/* append the free handle and the chunk size in first 8 bytes
2291 			 * after the command and null character
2292 			 */
2293 			*((int32 *)(p+1)) = handle;
2294 			*((int32 *)((p+1)+sizeof(int32))) = size;
2295 			/* append null as terminator */
2296 			*(p+1+2*sizeof(int32)) = '\0';
2297 			/* recalculated length -> 4 bytes for "free" + 8 bytes for hadnle and size
2298 			 * + 1 bytes for null caracter
2299 			 */
2300 			len = strlen(msg) + 2*sizeof(int32) + 1;
2301 			/* send iovar to free the chunk */
2302 			if (dhd_wl_ioctl_cmd(dhd, WLC_GET_VAR, msg, len, FALSE, 0) < 0) {
2303 				DHD_ERROR(("IOCTL failed for memdebug free\n"));
2304 			}
2305 		} else {
2306 			DHD_ERROR(("specified id does not exist\n"));
2307 		}
2308 	} else {
2309 		/* for all the wrong argument formats */
2310 		return BCME_BADARG;
2311 	}
2312 	return 0;
2313 }
2314 extern void
dhd_mw_list_delete(dhd_pub_t * dhd,dll_t * list_head)2315 dhd_mw_list_delete(dhd_pub_t *dhd, dll_t *list_head)
2316 {
2317 	dll_t *item;
2318 	dhd_dbg_mwli_t *mw_li;
2319 	while (!(dll_empty(list_head))) {
2320 		item = dll_head_p(list_head);
2321 		mw_li = (dhd_dbg_mwli_t *)CONTAINEROF(item, dhd_dbg_mwli_t, list);
2322 		dll_delete(item);
2323 		MFREE(dhd->osh, mw_li, sizeof(*mw_li));
2324 	}
2325 }
2326 #ifdef BCMPCIE
2327 int
dhd_flow_ring_debug(dhd_pub_t * dhd,char * msg,uint msglen)2328 dhd_flow_ring_debug(dhd_pub_t *dhd, char *msg, uint msglen)
2329 {
2330 	flow_ring_table_t *flow_ring_table;
2331 	char *cmd;
2332 	char *end_ptr = NULL;
2333 	uint8 prio;
2334 	uint16 flowid;
2335 	int i;
2336 	int ret = 0;
2337 	cmd = bcmstrstr(msg, " ");
2338 	BCM_REFERENCE(prio);
2339 	if (cmd != NULL) {
2340 		/* in order to use string operations append null */
2341 		*cmd = '\0';
2342 	} else {
2343 		DHD_ERROR(("missing: create/delete args\n"));
2344 		return BCME_ERROR;
2345 	}
2346 	if (cmd && !strcmp(msg, "create")) {
2347 		/* extract <"source address", "destination address", "priority"> */
2348 		uint8 sa[ETHER_ADDR_LEN], da[ETHER_ADDR_LEN];
2349 		BCM_REFERENCE(sa);
2350 		BCM_REFERENCE(da);
2351 		msg = msg + strlen("create") + 1;
2352 		/* fill ethernet source address */
2353 		for (i = 0; i < ETHER_ADDR_LEN; i++) {
2354 			sa[i] = (uint8)bcm_strtoul(msg, &end_ptr, 16);
2355 			if (*end_ptr == ':') {
2356 				msg = (end_ptr + 1);
2357 			} else if (i != 5) {
2358 				DHD_ERROR(("not a valid source mac addr\n"));
2359 				return BCME_ERROR;
2360 			}
2361 		}
2362 		if (*end_ptr != ' ') {
2363 			DHD_ERROR(("missing: destiantion mac id\n"));
2364 			return BCME_ERROR;
2365 		} else {
2366 			/* skip space */
2367 			msg = end_ptr + 1;
2368 		}
2369 		/* fill ethernet destination address */
2370 		for (i = 0; i < ETHER_ADDR_LEN; i++) {
2371 			da[i] = (uint8)bcm_strtoul(msg, &end_ptr, 16);
2372 			if (*end_ptr == ':') {
2373 				msg = (end_ptr + 1);
2374 			} else if (i != 5) {
2375 				DHD_ERROR(("not a valid destination  mac addr\n"));
2376 				return BCME_ERROR;
2377 			}
2378 		}
2379 		if (*end_ptr != ' ') {
2380 			DHD_ERROR(("missing: priority\n"));
2381 			return BCME_ERROR;
2382 		} else {
2383 			msg = end_ptr + 1;
2384 		}
2385 		/* parse priority */
2386 		prio = (uint8)bcm_strtoul(msg, &end_ptr, 10);
2387 		if (prio > MAXPRIO) {
2388 			DHD_ERROR(("%s: invalid priority. Must be between 0-7 inclusive\n",
2389 				__FUNCTION__));
2390 			return BCME_ERROR;
2391 		}
2392 
2393 		if (*end_ptr != '\0') {
2394 			DHD_ERROR(("msg not truncated with NULL character\n"));
2395 			return BCME_ERROR;
2396 		}
2397 		ret = dhd_flowid_debug_create(dhd, 0, prio, (char *)sa, (char *)da, &flowid);
2398 		if (ret != BCME_OK) {
2399 			DHD_ERROR(("%s: flowring creation failed ret: %d\n", __FUNCTION__, ret));
2400 			return BCME_ERROR;
2401 		}
2402 		return BCME_OK;
2403 
2404 	} else if (cmd && !strcmp(msg, "delete")) {
2405 		msg = msg + strlen("delete") + 1;
2406 		/* parse flowid */
2407 		flowid = (uint16)bcm_strtoul(msg, &end_ptr, 10);
2408 		if (*end_ptr != '\0') {
2409 			DHD_ERROR(("msg not truncated with NULL character\n"));
2410 			return BCME_ERROR;
2411 		}
2412 
2413 		/* Find flowid from ifidx 0 since this IOVAR creating flowring with ifidx 0 */
2414 		if (dhd_flowid_find_by_ifidx(dhd, 0, flowid) != BCME_OK)
2415 		{
2416 			DHD_ERROR(("%s : Deleting not created flowid: %u\n", __FUNCTION__, flowid));
2417 			return BCME_ERROR;
2418 		}
2419 
2420 		flow_ring_table = (flow_ring_table_t *)dhd->flow_ring_table;
2421 		ret = dhd_bus_flow_ring_delete_request(dhd->bus, (void *)&flow_ring_table[flowid]);
2422 		if (ret != BCME_OK) {
2423 			DHD_ERROR(("%s: flowring deletion failed ret: %d\n", __FUNCTION__, ret));
2424 			return BCME_ERROR;
2425 		}
2426 		return BCME_OK;
2427 	}
2428 	DHD_ERROR(("%s: neither create nor delete\n", __FUNCTION__));
2429 	return BCME_ERROR;
2430 }
2431 #endif /* BCMPCIE */
2432 #endif /* DHD_DEBUG */
2433 
2434 static int
dhd_doiovar(dhd_pub_t * dhd_pub,const bcm_iovar_t * vi,uint32 actionid,const char * name,void * params,int plen,void * arg,uint len,int val_size)2435 dhd_doiovar(dhd_pub_t *dhd_pub, const bcm_iovar_t *vi, uint32 actionid, const char *name,
2436             void *params, int plen, void *arg, uint len, int val_size)
2437 {
2438 	int bcmerror = 0;
2439 	int32 int_val = 0;
2440 	uint32 dhd_ver_len, bus_api_rev_len;
2441 
2442 	DHD_TRACE(("%s: Enter\n", __FUNCTION__));
2443 	DHD_TRACE(("%s: actionid = %d; name %s\n", __FUNCTION__, actionid, name));
2444 
2445 	if ((bcmerror = bcm_iovar_lencheck(vi, arg, len, IOV_ISSET(actionid))) != 0)
2446 		goto exit;
2447 
2448 	if (plen >= (int)sizeof(int_val))
2449 		bcopy(params, &int_val, sizeof(int_val));
2450 
2451 	switch (actionid) {
2452 	case IOV_GVAL(IOV_VERSION):
2453 		/* Need to have checked buffer length */
2454 		dhd_ver_len = sizeof(dhd_version) - 1;
2455 		bus_api_rev_len = strlen(bus_api_revision);
2456 		if (len > dhd_ver_len + bus_api_rev_len) {
2457 			bcmerror = memcpy_s((char *)arg, len, dhd_version, dhd_ver_len);
2458 			if (bcmerror != BCME_OK) {
2459 				break;
2460 			}
2461 			bcmerror = memcpy_s((char *)arg + dhd_ver_len, len - dhd_ver_len,
2462 				bus_api_revision, bus_api_rev_len);
2463 			if (bcmerror != BCME_OK) {
2464 				break;
2465 			}
2466 			*((char *)arg + dhd_ver_len + bus_api_rev_len) = '\0';
2467 		}
2468 #if defined(BCMSDIO) && defined(PKT_STATICS)
2469 		dhd_bus_clear_txpktstatics(dhd_pub->bus);
2470 #endif
2471 		break;
2472 
2473 	case IOV_GVAL(IOV_WLMSGLEVEL):
2474 		printf("android_msg_level=0x%x\n", android_msg_level);
2475 		printf("config_msg_level=0x%x\n", config_msg_level);
2476 #if defined(WL_WIRELESS_EXT)
2477 		int_val = (int32)iw_msg_level;
2478 		bcopy(&int_val, arg, val_size);
2479 		printf("iw_msg_level=0x%x\n", iw_msg_level);
2480 #endif
2481 #ifdef WL_CFG80211
2482 		int_val = (int32)wl_dbg_level;
2483 		bcopy(&int_val, arg, val_size);
2484 		printf("cfg_msg_level=0x%x\n", wl_dbg_level);
2485 #endif
2486 		break;
2487 
2488 	case IOV_SVAL(IOV_WLMSGLEVEL):
2489 		if (int_val & DHD_ANDROID_VAL) {
2490 			android_msg_level = (uint)(int_val & 0xFFFF);
2491 			printf("android_msg_level=0x%x\n", android_msg_level);
2492 		}
2493 		if (int_val & DHD_CONFIG_VAL) {
2494 			config_msg_level = (uint)(int_val & 0xFFFF);
2495 			printf("config_msg_level=0x%x\n", config_msg_level);
2496 		}
2497 #if defined(WL_WIRELESS_EXT)
2498 		if (int_val & DHD_IW_VAL) {
2499 			iw_msg_level = (uint)(int_val & 0xFFFF);
2500 			printf("iw_msg_level=0x%x\n", iw_msg_level);
2501 		}
2502 #endif
2503 #ifdef WL_CFG80211
2504 		if (int_val & DHD_CFG_VAL) {
2505 			wl_cfg80211_enable_trace((u32)(int_val & 0xFFFF));
2506 		}
2507 #endif
2508 		break;
2509 
2510 	case IOV_GVAL(IOV_MSGLEVEL):
2511 		int_val = (int32)dhd_msg_level;
2512 		bcopy(&int_val, arg, val_size);
2513 #if defined(BCMSDIO) && defined(PKT_STATICS)
2514 		dhd_bus_dump_txpktstatics(dhd_pub->bus);
2515 #endif
2516 		break;
2517 
2518 	case IOV_SVAL(IOV_MSGLEVEL):
2519 		dhd_msg_level = int_val;
2520 		break;
2521 
2522 	case IOV_GVAL(IOV_BCMERRORSTR):
2523 		bcm_strncpy_s((char *)arg, len, bcmerrorstr(dhd_pub->bcmerror), BCME_STRLEN);
2524 		((char *)arg)[BCME_STRLEN - 1] = 0x00;
2525 		break;
2526 
2527 	case IOV_GVAL(IOV_BCMERROR):
2528 		int_val = (int32)dhd_pub->bcmerror;
2529 		bcopy(&int_val, arg, val_size);
2530 		break;
2531 
2532 #ifndef BCMDBUS
2533 	case IOV_GVAL(IOV_WDTICK):
2534 		int_val = (int32)dhd_watchdog_ms;
2535 		bcopy(&int_val, arg, val_size);
2536 		break;
2537 #endif /* !BCMDBUS */
2538 
2539 	case IOV_SVAL(IOV_WDTICK):
2540 		if (!dhd_pub->up) {
2541 			bcmerror = BCME_NOTUP;
2542 			break;
2543 		}
2544 
2545 		dhd_watchdog_ms = (uint)int_val;
2546 
2547 		dhd_os_wd_timer(dhd_pub, (uint)int_val);
2548 		break;
2549 
2550 	case IOV_GVAL(IOV_DUMP):
2551 		if (dhd_dump(dhd_pub, arg, len) <= 0)
2552 			bcmerror = BCME_ERROR;
2553 		else
2554 			bcmerror = BCME_OK;
2555 		break;
2556 
2557 #ifndef BCMDBUS
2558 	case IOV_GVAL(IOV_DCONSOLE_POLL):
2559 		int_val = (int32)dhd_pub->dhd_console_ms;
2560 		bcopy(&int_val, arg, val_size);
2561 		break;
2562 
2563 	case IOV_SVAL(IOV_DCONSOLE_POLL):
2564 		dhd_pub->dhd_console_ms = (uint)int_val;
2565 		break;
2566 
2567 #if defined(DHD_DEBUG)
2568 	case IOV_SVAL(IOV_CONS):
2569 		if (len > 0) {
2570 #ifdef CONSOLE_DPC
2571 			bcmerror = dhd_bus_txcons(dhd_pub, arg, len - 1);
2572 #else
2573 			bcmerror = dhd_bus_console_in(dhd_pub, arg, len - 1);
2574 #endif
2575 		}
2576 		break;
2577 #endif /* DHD_DEBUG */
2578 #endif /* !BCMDBUS */
2579 
2580 	case IOV_SVAL(IOV_CLEARCOUNTS):
2581 		dhd_pub->tx_packets = dhd_pub->rx_packets = 0;
2582 		dhd_pub->tx_errors = dhd_pub->rx_errors = 0;
2583 		dhd_pub->tx_ctlpkts = dhd_pub->rx_ctlpkts = 0;
2584 		dhd_pub->tx_ctlerrs = dhd_pub->rx_ctlerrs = 0;
2585 		dhd_pub->tx_dropped = 0;
2586 		dhd_pub->rx_dropped = 0;
2587 		dhd_pub->tx_pktgetfail = 0;
2588 		dhd_pub->rx_pktgetfail = 0;
2589 		dhd_pub->rx_readahead_cnt = 0;
2590 		dhd_pub->tx_realloc = 0;
2591 		dhd_pub->wd_dpc_sched = 0;
2592 		dhd_pub->tx_big_packets = 0;
2593 		memset(&dhd_pub->dstats, 0, sizeof(dhd_pub->dstats));
2594 		dhd_bus_clearcounts(dhd_pub);
2595 #ifdef PROP_TXSTATUS
2596 		/* clear proptxstatus related counters */
2597 		dhd_wlfc_clear_counts(dhd_pub);
2598 #endif /* PROP_TXSTATUS */
2599 #if defined(DHD_LB_STATS)
2600 		DHD_LB_STATS_RESET(dhd_pub);
2601 #endif /* DHD_LB_STATS */
2602 		break;
2603 
2604 #ifdef BCMPERFSTATS
2605 	case IOV_GVAL(IOV_LOGDUMP): {
2606 		bcmdumplog((char*)arg, len);
2607 		break;
2608 	}
2609 
2610 	case IOV_SVAL(IOV_LOGCAL): {
2611 		bcmlog("Starting OSL_DELAY (%d usecs)", (uint)int_val, 0);
2612 		OSL_DELAY((uint)int_val);
2613 		bcmlog("Finished OSL_DELAY (%d usecs)", (uint)int_val, 0);
2614 		break;
2615 	}
2616 
2617 	case IOV_SVAL(IOV_LOGSTAMP): {
2618 		int int_val2;
2619 
2620 		if (plen >= 2 * sizeof(int)) {
2621 			bcopy((char *)params + sizeof(int_val), &int_val2, sizeof(int_val2));
2622 			bcmlog("User message %d %d", (uint)int_val, (uint)int_val2);
2623 		} else if (plen >= sizeof(int)) {
2624 			bcmlog("User message %d", (uint)int_val, 0);
2625 		} else {
2626 			bcmlog("User message", 0, 0);
2627 		}
2628 		break;
2629 	}
2630 #endif /* BCMPERFSTATS */
2631 
2632 	case IOV_GVAL(IOV_IOCTLTIMEOUT): {
2633 		int_val = (int32)dhd_os_get_ioctl_resp_timeout();
2634 		bcopy(&int_val, arg, sizeof(int_val));
2635 		break;
2636 	}
2637 
2638 	case IOV_SVAL(IOV_IOCTLTIMEOUT): {
2639 		if (int_val <= 0)
2640 			bcmerror = BCME_BADARG;
2641 		else
2642 			dhd_os_set_ioctl_resp_timeout((unsigned int)int_val);
2643 		break;
2644 	}
2645 
2646 #ifdef PROP_TXSTATUS
2647 	case IOV_GVAL(IOV_PROPTXSTATUS_ENABLE): {
2648 		bool wlfc_enab = FALSE;
2649 		bcmerror = dhd_wlfc_get_enable(dhd_pub, &wlfc_enab);
2650 		if (bcmerror != BCME_OK)
2651 			goto exit;
2652 		int_val = wlfc_enab ? 1 : 0;
2653 		bcopy(&int_val, arg, val_size);
2654 		break;
2655 	}
2656 	case IOV_SVAL(IOV_PROPTXSTATUS_ENABLE): {
2657 		bool wlfc_enab = FALSE;
2658 		bcmerror = dhd_wlfc_get_enable(dhd_pub, &wlfc_enab);
2659 		if (bcmerror != BCME_OK)
2660 			goto exit;
2661 
2662 		/* wlfc is already set as desired */
2663 		if (wlfc_enab == (int_val == 0 ? FALSE : TRUE))
2664 			goto exit;
2665 
2666 		if (int_val == TRUE)
2667 			bcmerror = dhd_wlfc_init(dhd_pub);
2668 		else
2669 			bcmerror = dhd_wlfc_deinit(dhd_pub);
2670 
2671 		break;
2672 	}
2673 	case IOV_GVAL(IOV_PROPTXSTATUS_MODE):
2674 		bcmerror = dhd_wlfc_get_mode(dhd_pub, &int_val);
2675 		if (bcmerror != BCME_OK)
2676 			goto exit;
2677 		bcopy(&int_val, arg, val_size);
2678 		break;
2679 
2680 	case IOV_SVAL(IOV_PROPTXSTATUS_MODE):
2681 		dhd_wlfc_set_mode(dhd_pub, int_val);
2682 		break;
2683 #ifdef QMONITOR
2684 	case IOV_GVAL(IOV_QMON_TIME_THRES): {
2685 		int_val = dhd_qmon_thres(dhd_pub, FALSE, 0);
2686 		bcopy(&int_val, arg, val_size);
2687 		break;
2688 	}
2689 
2690 	case IOV_SVAL(IOV_QMON_TIME_THRES): {
2691 		dhd_qmon_thres(dhd_pub, TRUE, int_val);
2692 		break;
2693 	}
2694 
2695 	case IOV_GVAL(IOV_QMON_TIME_PERCENT): {
2696 		int_val = dhd_qmon_getpercent(dhd_pub);
2697 		bcopy(&int_val, arg, val_size);
2698 		break;
2699 	}
2700 #endif /* QMONITOR */
2701 
2702 	case IOV_GVAL(IOV_PROPTXSTATUS_MODULE_IGNORE):
2703 		bcmerror = dhd_wlfc_get_module_ignore(dhd_pub, &int_val);
2704 		if (bcmerror != BCME_OK)
2705 			goto exit;
2706 		bcopy(&int_val, arg, val_size);
2707 		break;
2708 
2709 	case IOV_SVAL(IOV_PROPTXSTATUS_MODULE_IGNORE):
2710 		dhd_wlfc_set_module_ignore(dhd_pub, int_val);
2711 		break;
2712 
2713 	case IOV_GVAL(IOV_PROPTXSTATUS_CREDIT_IGNORE):
2714 		bcmerror = dhd_wlfc_get_credit_ignore(dhd_pub, &int_val);
2715 		if (bcmerror != BCME_OK)
2716 			goto exit;
2717 		bcopy(&int_val, arg, val_size);
2718 		break;
2719 
2720 	case IOV_SVAL(IOV_PROPTXSTATUS_CREDIT_IGNORE):
2721 		dhd_wlfc_set_credit_ignore(dhd_pub, int_val);
2722 		break;
2723 
2724 	case IOV_GVAL(IOV_PROPTXSTATUS_TXSTATUS_IGNORE):
2725 		bcmerror = dhd_wlfc_get_txstatus_ignore(dhd_pub, &int_val);
2726 		if (bcmerror != BCME_OK)
2727 			goto exit;
2728 		bcopy(&int_val, arg, val_size);
2729 		break;
2730 
2731 	case IOV_SVAL(IOV_PROPTXSTATUS_TXSTATUS_IGNORE):
2732 		dhd_wlfc_set_txstatus_ignore(dhd_pub, int_val);
2733 		break;
2734 
2735 	case IOV_GVAL(IOV_PROPTXSTATUS_RXPKT_CHK):
2736 		bcmerror = dhd_wlfc_get_rxpkt_chk(dhd_pub, &int_val);
2737 		if (bcmerror != BCME_OK)
2738 			goto exit;
2739 		bcopy(&int_val, arg, val_size);
2740 		break;
2741 
2742 	case IOV_SVAL(IOV_PROPTXSTATUS_RXPKT_CHK):
2743 		dhd_wlfc_set_rxpkt_chk(dhd_pub, int_val);
2744 		break;
2745 
2746 #endif /* PROP_TXSTATUS */
2747 
2748 	case IOV_GVAL(IOV_BUS_TYPE):
2749 		/* The dhd application queries the driver to check if its usb or sdio.  */
2750 #ifdef BCMDBUS
2751 		int_val = BUS_TYPE_USB;
2752 #endif
2753 #ifdef BCMSDIO
2754 		int_val = BUS_TYPE_SDIO;
2755 #endif
2756 #ifdef PCIE_FULL_DONGLE
2757 		int_val = BUS_TYPE_PCIE;
2758 #endif
2759 		bcopy(&int_val, arg, val_size);
2760 		break;
2761 
2762 	case IOV_SVAL(IOV_CHANGEMTU):
2763 		int_val &= 0xffff;
2764 		bcmerror = dhd_change_mtu(dhd_pub, int_val, 0);
2765 		break;
2766 
2767 	case IOV_GVAL(IOV_HOSTREORDER_FLOWS):
2768 	{
2769 		uint i = 0;
2770 		uint8 *ptr = (uint8 *)arg;
2771 		uint8 count = 0;
2772 
2773 		ptr++;
2774 		for (i = 0; i < WLHOST_REORDERDATA_MAXFLOWS; i++) {
2775 			if (dhd_pub->reorder_bufs[i] != NULL) {
2776 				*ptr = dhd_pub->reorder_bufs[i]->flow_id;
2777 				ptr++;
2778 				count++;
2779 			}
2780 		}
2781 		ptr = (uint8 *)arg;
2782 		*ptr = count;
2783 		break;
2784 	}
2785 #ifdef DHDTCPACK_SUPPRESS
2786 	case IOV_GVAL(IOV_TCPACK_SUPPRESS): {
2787 		int_val = (uint32)dhd_pub->tcpack_sup_mode;
2788 		bcopy(&int_val, arg, val_size);
2789 		break;
2790 	}
2791 	case IOV_SVAL(IOV_TCPACK_SUPPRESS): {
2792 		bcmerror = dhd_tcpack_suppress_set(dhd_pub, (uint8)int_val);
2793 		break;
2794 	}
2795 #endif /* DHDTCPACK_SUPPRESS */
2796 #ifdef DHD_WMF
2797 	case IOV_GVAL(IOV_WMF_BSS_ENAB): {
2798 		uint32	bssidx;
2799 		dhd_wmf_t *wmf;
2800 		const char *val;
2801 
2802 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
2803 			DHD_ERROR(("%s: wmf_bss_enable: bad parameter\n", __FUNCTION__));
2804 			bcmerror = BCME_BADARG;
2805 			break;
2806 		}
2807 
2808 		wmf = dhd_wmf_conf(dhd_pub, bssidx);
2809 		int_val = wmf->wmf_enable ? 1 :0;
2810 		bcopy(&int_val, arg, val_size);
2811 		break;
2812 	}
2813 	case IOV_SVAL(IOV_WMF_BSS_ENAB): {
2814 		/* Enable/Disable WMF */
2815 		uint32	bssidx;
2816 		dhd_wmf_t *wmf;
2817 		const char *val;
2818 
2819 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
2820 			DHD_ERROR(("%s: wmf_bss_enable: bad parameter\n", __FUNCTION__));
2821 			bcmerror = BCME_BADARG;
2822 			break;
2823 		}
2824 
2825 		ASSERT(val);
2826 		bcopy(val, &int_val, sizeof(uint32));
2827 		wmf = dhd_wmf_conf(dhd_pub, bssidx);
2828 		if (wmf->wmf_enable == int_val)
2829 			break;
2830 		if (int_val) {
2831 			/* Enable WMF */
2832 			if (dhd_wmf_instance_add(dhd_pub, bssidx) != BCME_OK) {
2833 				DHD_ERROR(("%s: Error in creating WMF instance\n",
2834 				__FUNCTION__));
2835 				break;
2836 			}
2837 			if (dhd_wmf_start(dhd_pub, bssidx) != BCME_OK) {
2838 				DHD_ERROR(("%s: Failed to start WMF\n", __FUNCTION__));
2839 				break;
2840 			}
2841 			wmf->wmf_enable = TRUE;
2842 		} else {
2843 			/* Disable WMF */
2844 			wmf->wmf_enable = FALSE;
2845 			dhd_wmf_stop(dhd_pub, bssidx);
2846 			dhd_wmf_instance_del(dhd_pub, bssidx);
2847 		}
2848 		break;
2849 	}
2850 	case IOV_GVAL(IOV_WMF_UCAST_IGMP):
2851 		int_val = dhd_pub->wmf_ucast_igmp ? 1 : 0;
2852 		bcopy(&int_val, arg, val_size);
2853 		break;
2854 	case IOV_SVAL(IOV_WMF_UCAST_IGMP):
2855 		if (dhd_pub->wmf_ucast_igmp == int_val)
2856 			break;
2857 
2858 		if (int_val >= OFF && int_val <= ON)
2859 			dhd_pub->wmf_ucast_igmp = int_val;
2860 		else
2861 			bcmerror = BCME_RANGE;
2862 		break;
2863 	case IOV_GVAL(IOV_WMF_MCAST_DATA_SENDUP):
2864 		int_val = dhd_wmf_mcast_data_sendup(dhd_pub, 0, FALSE, FALSE);
2865 		bcopy(&int_val, arg, val_size);
2866 		break;
2867 	case IOV_SVAL(IOV_WMF_MCAST_DATA_SENDUP):
2868 		dhd_wmf_mcast_data_sendup(dhd_pub, 0, TRUE, int_val);
2869 		break;
2870 
2871 #ifdef WL_IGMP_UCQUERY
2872 	case IOV_GVAL(IOV_WMF_UCAST_IGMP_QUERY):
2873 		int_val = dhd_pub->wmf_ucast_igmp_query ? 1 : 0;
2874 		bcopy(&int_val, arg, val_size);
2875 		break;
2876 	case IOV_SVAL(IOV_WMF_UCAST_IGMP_QUERY):
2877 		if (dhd_pub->wmf_ucast_igmp_query == int_val)
2878 			break;
2879 
2880 		if (int_val >= OFF && int_val <= ON)
2881 			dhd_pub->wmf_ucast_igmp_query = int_val;
2882 		else
2883 			bcmerror = BCME_RANGE;
2884 		break;
2885 #endif /* WL_IGMP_UCQUERY */
2886 #ifdef DHD_UCAST_UPNP
2887 	case IOV_GVAL(IOV_WMF_UCAST_UPNP):
2888 		int_val = dhd_pub->wmf_ucast_upnp ? 1 : 0;
2889 		bcopy(&int_val, arg, val_size);
2890 		break;
2891 	case IOV_SVAL(IOV_WMF_UCAST_UPNP):
2892 		if (dhd_pub->wmf_ucast_upnp == int_val)
2893 			break;
2894 
2895 		if (int_val >= OFF && int_val <= ON)
2896 			dhd_pub->wmf_ucast_upnp = int_val;
2897 		else
2898 			bcmerror = BCME_RANGE;
2899 		break;
2900 #endif /* DHD_UCAST_UPNP */
2901 
2902 	case IOV_GVAL(IOV_WMF_PSTA_DISABLE): {
2903 		uint32	bssidx;
2904 		const char *val;
2905 
2906 		if (dhd_iovar_parse_bssidx(dhd_pub, (char *)name, &bssidx, &val) != BCME_OK) {
2907 			DHD_ERROR(("%s: ap isoalate: bad parameter\n", __FUNCTION__));
2908 			bcmerror = BCME_BADARG;
2909 			break;
2910 		}
2911 
2912 		int_val = dhd_get_wmf_psta_disable(dhd_pub, bssidx);
2913 		bcopy(&int_val, arg, val_size);
2914 		break;
2915 	}
2916 
2917 	case IOV_SVAL(IOV_WMF_PSTA_DISABLE): {
2918 		uint32	bssidx;
2919 		const char *val;
2920 
2921 		if (dhd_iovar_parse_bssidx(dhd_pub, (char *)name, &bssidx, &val) != BCME_OK) {
2922 			DHD_ERROR(("%s: ap isolate: bad parameter\n", __FUNCTION__));
2923 			bcmerror = BCME_BADARG;
2924 			break;
2925 		}
2926 
2927 		ASSERT(val);
2928 		bcopy(val, &int_val, sizeof(uint32));
2929 		dhd_set_wmf_psta_disable(dhd_pub, bssidx, int_val);
2930 		break;
2931 	}
2932 #endif /* DHD_WMF */
2933 
2934 #if defined(BCM_ROUTER_DHD)
2935 	case IOV_SVAL(IOV_TRAFFIC_MGMT_DWM): {
2936 			trf_mgmt_filter_list_t   *trf_mgmt_filter_list =
2937 				(trf_mgmt_filter_list_t *)(arg);
2938 			bcmerror = traffic_mgmt_add_dwm_filter(dhd_pub, trf_mgmt_filter_list, len);
2939 		}
2940 		break;
2941 #endif /* BCM_ROUTER_DHD */
2942 
2943 #ifdef DHD_L2_FILTER
2944 	case IOV_GVAL(IOV_DHCP_UNICAST): {
2945 		uint32 bssidx;
2946 		const char *val;
2947 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
2948 			DHD_ERROR(("%s: IOV_DHCP_UNICAST: bad parameterand name = %s\n",
2949 				__FUNCTION__, name));
2950 			bcmerror = BCME_BADARG;
2951 			break;
2952 		}
2953 		int_val = dhd_get_dhcp_unicast_status(dhd_pub, bssidx);
2954 		memcpy(arg, &int_val, val_size);
2955 		break;
2956 	}
2957 	case IOV_SVAL(IOV_DHCP_UNICAST): {
2958 		uint32	bssidx;
2959 		const char *val;
2960 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
2961 			DHD_ERROR(("%s: IOV_DHCP_UNICAST: bad parameterand name = %s\n",
2962 				__FUNCTION__, name));
2963 			bcmerror = BCME_BADARG;
2964 			break;
2965 		}
2966 		memcpy(&int_val, val, sizeof(int_val));
2967 		bcmerror = dhd_set_dhcp_unicast_status(dhd_pub, bssidx, int_val ? 1 : 0);
2968 		break;
2969 	}
2970 	case IOV_GVAL(IOV_BLOCK_PING): {
2971 		uint32 bssidx;
2972 		const char *val;
2973 
2974 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
2975 			DHD_ERROR(("%s: IOV_BLOCK_PING: bad parameter\n", __FUNCTION__));
2976 			bcmerror = BCME_BADARG;
2977 			break;
2978 		}
2979 		int_val = dhd_get_block_ping_status(dhd_pub, bssidx);
2980 		memcpy(arg, &int_val, val_size);
2981 		break;
2982 	}
2983 	case IOV_SVAL(IOV_BLOCK_PING): {
2984 		uint32	bssidx;
2985 		const char *val;
2986 
2987 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
2988 			DHD_ERROR(("%s: IOV_BLOCK_PING: bad parameter\n", __FUNCTION__));
2989 			bcmerror = BCME_BADARG;
2990 			break;
2991 		}
2992 		memcpy(&int_val, val, sizeof(int_val));
2993 		bcmerror = dhd_set_block_ping_status(dhd_pub, bssidx, int_val ? 1 : 0);
2994 		break;
2995 	}
2996 	case IOV_GVAL(IOV_PROXY_ARP): {
2997 		uint32	bssidx;
2998 		const char *val;
2999 
3000 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
3001 			DHD_ERROR(("%s: IOV_PROXY_ARP: bad parameter\n", __FUNCTION__));
3002 			bcmerror = BCME_BADARG;
3003 			break;
3004 		}
3005 		int_val = dhd_get_parp_status(dhd_pub, bssidx);
3006 		bcopy(&int_val, arg, val_size);
3007 		break;
3008 	}
3009 	case IOV_SVAL(IOV_PROXY_ARP): {
3010 		uint32	bssidx;
3011 		const char *val;
3012 
3013 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
3014 			DHD_ERROR(("%s: IOV_PROXY_ARP: bad parameter\n", __FUNCTION__));
3015 			bcmerror = BCME_BADARG;
3016 			break;
3017 		}
3018 		bcopy(val, &int_val, sizeof(int_val));
3019 
3020 		/* Issue a iovar request to WL to update the proxy arp capability bit
3021 		 * in the Extended Capability IE of beacons/probe responses.
3022 		 */
3023 		bcmerror = dhd_iovar(dhd_pub, bssidx, "proxy_arp_advertise", val, sizeof(int_val),
3024 				NULL, 0, TRUE);
3025 		if (bcmerror == BCME_OK) {
3026 			dhd_set_parp_status(dhd_pub, bssidx, int_val ? 1 : 0);
3027 		}
3028 		break;
3029 	}
3030 	case IOV_GVAL(IOV_GRAT_ARP): {
3031 		uint32 bssidx;
3032 		const char *val;
3033 
3034 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
3035 			DHD_ERROR(("%s: IOV_GRAT_ARP: bad parameter\n", __FUNCTION__));
3036 			bcmerror = BCME_BADARG;
3037 			break;
3038 		}
3039 		int_val = dhd_get_grat_arp_status(dhd_pub, bssidx);
3040 		memcpy(arg, &int_val, val_size);
3041 		break;
3042 	}
3043 	case IOV_SVAL(IOV_GRAT_ARP): {
3044 		uint32	bssidx;
3045 		const char *val;
3046 
3047 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
3048 			DHD_ERROR(("%s: IOV_GRAT_ARP: bad parameter\n", __FUNCTION__));
3049 			bcmerror = BCME_BADARG;
3050 			break;
3051 		}
3052 		memcpy(&int_val, val, sizeof(int_val));
3053 		bcmerror = dhd_set_grat_arp_status(dhd_pub, bssidx, int_val ? 1 : 0);
3054 		break;
3055 	}
3056 	case IOV_GVAL(IOV_BLOCK_TDLS): {
3057 		uint32 bssidx;
3058 		const char *val;
3059 
3060 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
3061 			DHD_ERROR(("%s: IOV_BLOCK_TDLS: bad parameter\n", __FUNCTION__));
3062 			bcmerror = BCME_BADARG;
3063 			break;
3064 		}
3065 		int_val = dhd_get_block_tdls_status(dhd_pub, bssidx);
3066 		memcpy(arg, &int_val, val_size);
3067 		break;
3068 	}
3069 	case IOV_SVAL(IOV_BLOCK_TDLS): {
3070 		uint32	bssidx;
3071 		const char *val;
3072 
3073 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
3074 			DHD_ERROR(("%s: IOV_BLOCK_TDLS: bad parameter\n", __FUNCTION__));
3075 			bcmerror = BCME_BADARG;
3076 			break;
3077 		}
3078 		memcpy(&int_val, val, sizeof(int_val));
3079 		bcmerror = dhd_set_block_tdls_status(dhd_pub, bssidx, int_val ? 1 : 0);
3080 		break;
3081 	}
3082 #endif /* DHD_L2_FILTER */
3083 	case IOV_SVAL(IOV_DHD_IE): {
3084 		uint32	bssidx;
3085 		const char *val;
3086 #if (defined(BCM_ROUTER_DHD) && defined(QOS_MAP_SET))
3087 		uint8 ie_type;
3088 		bcm_tlv_t *qos_map_ie = NULL;
3089 		ie_setbuf_t *ie_getbufp = (ie_setbuf_t *)(arg+4);
3090 		ie_type = ie_getbufp->ie_buffer.ie_list[0].ie_data.id;
3091 #endif /* BCM_ROUTER_DHD && QOS_MAP_SET */
3092 
3093 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
3094 			DHD_ERROR(("%s: dhd ie: bad parameter\n", __FUNCTION__));
3095 			bcmerror = BCME_BADARG;
3096 			break;
3097 		}
3098 
3099 #if (defined(BCM_ROUTER_DHD) && defined(QOS_MAP_SET))
3100 		qos_map_ie = (bcm_tlv_t *)(&(ie_getbufp->ie_buffer.ie_list[0].ie_data));
3101 		if (qos_map_ie != NULL && (ie_type == DOT11_MNG_QOS_MAP_ID)) {
3102 				bcmerror = dhd_set_qosmap_up_table(dhd_pub, bssidx, qos_map_ie);
3103 		}
3104 #endif /* BCM_ROUTER_DHD && QOS_MAP_SET */
3105 		break;
3106 	}
3107 	case IOV_GVAL(IOV_AP_ISOLATE): {
3108 		uint32	bssidx;
3109 		const char *val;
3110 
3111 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
3112 			DHD_ERROR(("%s: ap isoalate: bad parameter\n", __FUNCTION__));
3113 			bcmerror = BCME_BADARG;
3114 			break;
3115 		}
3116 
3117 		int_val = dhd_get_ap_isolate(dhd_pub, bssidx);
3118 		bcopy(&int_val, arg, val_size);
3119 		break;
3120 	}
3121 	case IOV_SVAL(IOV_AP_ISOLATE): {
3122 		uint32	bssidx;
3123 		const char *val;
3124 
3125 		if (dhd_iovar_parse_bssidx(dhd_pub, name, &bssidx, &val) != BCME_OK) {
3126 			DHD_ERROR(("%s: ap isolate: bad parameter\n", __FUNCTION__));
3127 			bcmerror = BCME_BADARG;
3128 			break;
3129 		}
3130 
3131 		ASSERT(val);
3132 		bcopy(val, &int_val, sizeof(uint32));
3133 		dhd_set_ap_isolate(dhd_pub, bssidx, int_val);
3134 		break;
3135 	}
3136 #ifdef DHD_PSTA
3137 	case IOV_GVAL(IOV_PSTA): {
3138 		int_val = dhd_get_psta_mode(dhd_pub);
3139 		bcopy(&int_val, arg, val_size);
3140 		break;
3141 		}
3142 	case IOV_SVAL(IOV_PSTA): {
3143 		if (int_val >= DHD_MODE_PSTA_DISABLED && int_val <= DHD_MODE_PSR) {
3144 			dhd_set_psta_mode(dhd_pub, int_val);
3145 		} else {
3146 			bcmerror = BCME_RANGE;
3147 		}
3148 		break;
3149 		}
3150 #endif /* DHD_PSTA */
3151 #ifdef DHD_WET
3152 	case IOV_GVAL(IOV_WET):
3153 		 int_val = dhd_get_wet_mode(dhd_pub);
3154 		 bcopy(&int_val, arg, val_size);
3155 		 break;
3156 
3157 	case IOV_SVAL(IOV_WET):
3158 		 if (int_val == 0 || int_val == 1) {
3159 			 dhd_set_wet_mode(dhd_pub, int_val);
3160 			 /* Delete the WET DB when disabled */
3161 			 if (!int_val) {
3162 				 dhd_wet_sta_delete_list(dhd_pub);
3163 			 }
3164 		 } else {
3165 			 bcmerror = BCME_RANGE;
3166 		 }
3167 				 break;
3168 	case IOV_SVAL(IOV_WET_HOST_IPV4):
3169 			dhd_set_wet_host_ipv4(dhd_pub, params, plen);
3170 			break;
3171 	case IOV_SVAL(IOV_WET_HOST_MAC):
3172 			dhd_set_wet_host_mac(dhd_pub, params, plen);
3173 		break;
3174 #endif /* DHD_WET */
3175 #ifdef DHD_MCAST_REGEN
3176 	case IOV_GVAL(IOV_MCAST_REGEN_BSS_ENABLE): {
3177 		uint32	bssidx;
3178 		const char *val;
3179 
3180 		if (dhd_iovar_parse_bssidx(dhd_pub, (char *)name, &bssidx, &val) != BCME_OK) {
3181 			DHD_ERROR(("%s: mcast_regen_bss_enable: bad parameter\n", __FUNCTION__));
3182 			bcmerror = BCME_BADARG;
3183 			break;
3184 		}
3185 
3186 		int_val = dhd_get_mcast_regen_bss_enable(dhd_pub, bssidx);
3187 		bcopy(&int_val, arg, val_size);
3188 		break;
3189 	}
3190 
3191 	case IOV_SVAL(IOV_MCAST_REGEN_BSS_ENABLE): {
3192 		uint32	bssidx;
3193 		const char *val;
3194 
3195 		if (dhd_iovar_parse_bssidx(dhd_pub, (char *)name, &bssidx, &val) != BCME_OK) {
3196 			DHD_ERROR(("%s: mcast_regen_bss_enable: bad parameter\n", __FUNCTION__));
3197 			bcmerror = BCME_BADARG;
3198 			break;
3199 		}
3200 
3201 		ASSERT(val);
3202 		bcopy(val, &int_val, sizeof(uint32));
3203 		dhd_set_mcast_regen_bss_enable(dhd_pub, bssidx, int_val);
3204 		break;
3205 	}
3206 #endif /* DHD_MCAST_REGEN */
3207 
3208 	case IOV_GVAL(IOV_CFG80211_OPMODE): {
3209 		int_val = (int32)dhd_pub->op_mode;
3210 		bcopy(&int_val, arg, sizeof(int_val));
3211 		break;
3212 		}
3213 	case IOV_SVAL(IOV_CFG80211_OPMODE): {
3214 		if (int_val <= 0)
3215 			bcmerror = BCME_BADARG;
3216 		else
3217 			dhd_pub->op_mode = int_val;
3218 		break;
3219 	}
3220 
3221 	case IOV_GVAL(IOV_ASSERT_TYPE):
3222 		int_val = g_assert_type;
3223 		bcopy(&int_val, arg, val_size);
3224 		break;
3225 
3226 	case IOV_SVAL(IOV_ASSERT_TYPE):
3227 		g_assert_type = (uint32)int_val;
3228 		break;
3229 
3230 #if defined(NDIS)
3231 	case IOV_GVAL(IOV_WAKEIND):
3232 		dhd_os_wakeind(dhd_pub, &int_val);
3233 		bcopy(&int_val, arg, val_size);
3234 		break;
3235 #endif /* NDIS */
3236 
3237 #if !defined(NDIS) && !defined(BCM_ROUTER_DHD)
3238 	case IOV_GVAL(IOV_LMTEST): {
3239 		*(uint32 *)arg = (uint32)lmtest;
3240 		break;
3241 	}
3242 
3243 	case IOV_SVAL(IOV_LMTEST): {
3244 		uint32 val = *(uint32 *)arg;
3245 		if (val > 50)
3246 			bcmerror = BCME_BADARG;
3247 		else {
3248 			lmtest = (uint)val;
3249 			DHD_ERROR(("%s: lmtest %s\n",
3250 				__FUNCTION__, (lmtest == FALSE)? "OFF" : "ON"));
3251 		}
3252 		break;
3253 	}
3254 #endif /* !NDIS && !BCM_ROUTER_DHD */
3255 #ifdef BCMDBG
3256 	case IOV_GVAL(IOV_MACDBG_PD11REGS):
3257 		bcmerror = dhd_macdbg_pd11regs(dhd_pub, params, plen, arg, len);
3258 		break;
3259 	case IOV_GVAL(IOV_MACDBG_REGLIST):
3260 		bcmerror = dhd_macdbg_reglist(dhd_pub, arg, len);
3261 		break;
3262 	case IOV_GVAL(IOV_MACDBG_PSVMPMEMS):
3263 		bcmerror = dhd_macdbg_psvmpmems(dhd_pub, params, plen, arg, len);
3264 		break;
3265 #endif /* BCMDBG */
3266 
3267 #ifdef SHOW_LOGTRACE
3268 	case IOV_GVAL(IOV_DUMP_TRACE_LOG): {
3269 		trace_buf_info_t *trace_buf_info = (trace_buf_info_t *)arg;
3270 		dhd_dbg_ring_t *dbg_verbose_ring = NULL;
3271 
3272 		dbg_verbose_ring = dhd_dbg_get_ring_from_ring_id(dhd_pub, FW_VERBOSE_RING_ID);
3273 		if (dbg_verbose_ring == NULL) {
3274 			DHD_ERROR(("dbg_verbose_ring is NULL\n"));
3275 			bcmerror = BCME_UNSUPPORTED;
3276 			break;
3277 		}
3278 
3279 		if (trace_buf_info != NULL) {
3280 			bzero(trace_buf_info, sizeof(trace_buf_info_t));
3281 			dhd_dbg_read_ring_into_trace_buf(dbg_verbose_ring, trace_buf_info);
3282 		} else {
3283 			DHD_ERROR(("%s: arg is NULL\n", __FUNCTION__));
3284 			bcmerror = BCME_NOMEM;
3285 		}
3286 		break;
3287 	}
3288 #endif /* SHOW_LOGTRACE */
3289 #ifdef BTLOG
3290 	case IOV_GVAL(IOV_DUMP_BT_LOG): {
3291 		bt_log_buf_info_t *bt_log_buf_info = (bt_log_buf_info_t *)arg;
3292 		uint32 rlen;
3293 
3294 		rlen = dhd_dbg_pull_single_from_ring(dhd_pub, BT_LOG_RING_ID, bt_log_buf_info->buf,
3295 			BT_LOG_BUF_MAX_SIZE, TRUE);
3296 		bt_log_buf_info->size = rlen;
3297 		bt_log_buf_info->availability = BT_LOG_NEXT_BUF_NOT_AVAIL;
3298 		if (rlen == 0) {
3299 			bt_log_buf_info->availability = BT_LOG_BUF_NOT_AVAILABLE;
3300 		} else {
3301 			dhd_dbg_ring_status_t ring_status;
3302 			dhd_dbg_get_ring_status(dhd_pub, BT_LOG_RING_ID, &ring_status);
3303 			if (ring_status.written_bytes != ring_status.read_bytes) {
3304 				bt_log_buf_info->availability = BT_LOG_NEXT_BUF_AVAIL;
3305 			}
3306 		}
3307 		break;
3308 	}
3309 	case IOV_GVAL(IOV_BTLOG):
3310 	{
3311 		uint32 btlog_val = dhd_pub->bt_logging_enabled ? 1 : 0;
3312 		bcopy(&btlog_val, arg, val_size);
3313 	}
3314 		break;
3315 	case IOV_SVAL(IOV_BTLOG):
3316 	{
3317 		if (dhd_pub->busstate != DHD_BUS_DOWN) {
3318 			DHD_ERROR(("%s: Can change only when bus down (before FW download)\n",
3319 				__FUNCTION__));
3320 			bcmerror = BCME_NOTDOWN;
3321 			break;
3322 		}
3323 		if (int_val)
3324 			dhd_pub->bt_logging_enabled = TRUE;
3325 		else
3326 			dhd_pub->bt_logging_enabled = FALSE;
3327 	}
3328 		break;
3329 
3330 #endif	/* BTLOG */
3331 #ifdef SNAPSHOT_UPLOAD
3332 	case IOV_SVAL(IOV_BT_MEM_DUMP): {
3333 		dhd_prot_send_snapshot_request(dhd_pub, SNAPSHOT_TYPE_BT, int_val);
3334 		break;
3335 	}
3336 	case IOV_GVAL(IOV_BT_UPLOAD): {
3337 		int status;
3338 		bt_mem_req_t req;
3339 		bt_log_buf_info_t *mem_info = (bt_log_buf_info_t *)arg;
3340 		uint32 size;
3341 		bool is_more;
3342 
3343 		memcpy(&req, params, sizeof(req));
3344 
3345 		status = dhd_prot_get_snapshot(dhd_pub, SNAPSHOT_TYPE_BT, req.offset,
3346 			req.buf_size, mem_info->buf, &size, &is_more);
3347 		if (status == BCME_OK) {
3348 			mem_info->size = size;
3349 			mem_info->availability = is_more ?
3350 				BT_LOG_NEXT_BUF_AVAIL : BT_LOG_NEXT_BUF_NOT_AVAIL;
3351 		} else if (status == BCME_NOTREADY) {
3352 			mem_info->size = 0;
3353 			mem_info->availability = BT_LOG_NOT_READY;
3354 		} else {
3355 			mem_info->size = 0;
3356 			mem_info->availability = BT_LOG_BUF_NOT_AVAILABLE;
3357 		}
3358 		break;
3359 	}
3360 #endif	/* SNAPSHOT_UPLOAD */
3361 #ifdef REPORT_FATAL_TIMEOUTS
3362 	case IOV_GVAL(IOV_SCAN_TO): {
3363 		dhd_get_scan_to_val(dhd_pub, (uint32 *)&int_val);
3364 		bcopy(&int_val, arg, val_size);
3365 		break;
3366 	}
3367 	case IOV_SVAL(IOV_SCAN_TO): {
3368 		dhd_set_scan_to_val(dhd_pub, (uint32)int_val);
3369 		break;
3370 	}
3371 	case IOV_GVAL(IOV_JOIN_TO): {
3372 		dhd_get_join_to_val(dhd_pub, (uint32 *)&int_val);
3373 		bcopy(&int_val, arg, val_size);
3374 		break;
3375 	}
3376 	case IOV_SVAL(IOV_JOIN_TO): {
3377 		dhd_set_join_to_val(dhd_pub, (uint32)int_val);
3378 		break;
3379 	}
3380 	case IOV_GVAL(IOV_CMD_TO): {
3381 		dhd_get_cmd_to_val(dhd_pub, (uint32 *)&int_val);
3382 		bcopy(&int_val, arg, val_size);
3383 		break;
3384 	}
3385 	case IOV_SVAL(IOV_CMD_TO): {
3386 		dhd_set_cmd_to_val(dhd_pub, (uint32)int_val);
3387 		break;
3388 	}
3389 	case IOV_GVAL(IOV_OQS_TO): {
3390 		dhd_get_bus_to_val(dhd_pub, (uint32 *)&int_val);
3391 		bcopy(&int_val, arg, val_size);
3392 		break;
3393 	}
3394 	case IOV_SVAL(IOV_OQS_TO): {
3395 		dhd_set_bus_to_val(dhd_pub, (uint32)int_val);
3396 		break;
3397 	}
3398 #endif /* REPORT_FATAL_TIMEOUTS */
3399 	case IOV_GVAL(IOV_DONGLE_TRAP_TYPE):
3400 		if (dhd_pub->dongle_trap_occured)
3401 			int_val = ltoh32(dhd_pub->last_trap_info.type);
3402 		else
3403 			int_val = 0;
3404 		bcopy(&int_val, arg, val_size);
3405 		break;
3406 
3407 	case IOV_GVAL(IOV_DONGLE_TRAP_INFO):
3408 	{
3409 		struct bcmstrbuf strbuf;
3410 		bcm_binit(&strbuf, arg, len);
3411 		if (dhd_pub->dongle_trap_occured == FALSE) {
3412 			bcm_bprintf(&strbuf, "no trap recorded\n");
3413 			break;
3414 		}
3415 #ifndef BCMDBUS
3416 		dhd_bus_dump_trap_info(dhd_pub->bus, &strbuf);
3417 #endif /* BCMDBUS */
3418 		break;
3419 	}
3420 #ifdef DHD_DEBUG
3421 #if defined(BCMSDIO) || defined(BCMPCIE)
3422 
3423 	case IOV_GVAL(IOV_BPADDR):
3424 		{
3425 			sdreg_t sdreg;
3426 			uint32 addr, size;
3427 
3428 			memcpy(&sdreg, params, sizeof(sdreg));
3429 
3430 			addr = sdreg.offset;
3431 			size = sdreg.func;
3432 
3433 			bcmerror = dhd_bus_readwrite_bp_addr(dhd_pub, addr, size,
3434 				(uint *)&int_val, TRUE);
3435 
3436 			memcpy(arg, &int_val, sizeof(int32));
3437 
3438 			break;
3439 		}
3440 
3441 	case IOV_SVAL(IOV_BPADDR):
3442 		{
3443 			sdreg_t sdreg;
3444 			uint32 addr, size;
3445 
3446 			memcpy(&sdreg, params, sizeof(sdreg));
3447 
3448 			addr = sdreg.offset;
3449 			size = sdreg.func;
3450 
3451 			bcmerror = dhd_bus_readwrite_bp_addr(dhd_pub, addr, size,
3452 				(uint *)&sdreg.value,
3453 				FALSE);
3454 
3455 			break;
3456 		}
3457 #endif /* BCMSDIO || BCMPCIE */
3458 #ifdef BCMPCIE
3459 	case IOV_SVAL(IOV_FLOW_RING_DEBUG):
3460 		{
3461 			bcmerror = dhd_flow_ring_debug(dhd_pub, arg, len);
3462 			break;
3463 		}
3464 #endif /* BCMPCIE */
3465 	case IOV_SVAL(IOV_MEM_DEBUG):
3466 		if (len > 0) {
3467 			bcmerror = dhd_mem_debug(dhd_pub, arg, len - 1);
3468 		}
3469 		break;
3470 #endif /* DHD_DEBUG */
3471 #if defined(DHD_LOG_DUMP)
3472 #if defined(DHD_EFI)
3473 	case IOV_GVAL(IOV_LOG_CAPTURE_ENABLE):
3474 		{
3475 			int_val = dhd_pub->log_capture_enable;
3476 			bcopy(&int_val, arg, val_size);
3477 			break;
3478 		}
3479 
3480 	case IOV_SVAL(IOV_LOG_CAPTURE_ENABLE):
3481 		{
3482 			dhd_pub->log_capture_enable = (uint8)int_val;
3483 			break;
3484 		}
3485 #endif /* DHD_EFI */
3486 	case IOV_GVAL(IOV_LOG_DUMP):
3487 		{
3488 			dhd_prot_debug_info_print(dhd_pub);
3489 			dhd_log_dump_trigger(dhd_pub, CMD_DEFAULT);
3490 			break;
3491 		}
3492 #endif /* DHD_LOG_DUMP */
3493 
3494 	case IOV_GVAL(IOV_TPUT_TEST):
3495 		{
3496 			tput_test_t *tput_data = NULL;
3497 			if (params && plen >= sizeof(tput_test_t)) {
3498 				tput_data = (tput_test_t *)params;
3499 				bcmerror = dhd_tput_test(dhd_pub, tput_data);
3500 			} else {
3501 				DHD_ERROR(("%s: tput test - no input params ! \n", __FUNCTION__));
3502 				bcmerror = BCME_BADARG;
3503 			}
3504 			break;
3505 		}
3506 #if defined(DHD_AWDL) && defined(AWDL_SLOT_STATS)
3507 	case IOV_SVAL(IOV_PKT_LATENCY):
3508 		dhd_pub->pkt_latency = (uint32)int_val;
3509 		break;
3510 	case IOV_GVAL(IOV_PKT_LATENCY):
3511 		int_val = (int32)dhd_pub->pkt_latency;
3512 		bcopy(&int_val, arg, val_size);
3513 		break;
3514 #endif /* defined(DHD_AWDL) && defined(AWDL_SLOT_STATS)  */
3515 	case IOV_GVAL(IOV_DEBUG_BUF_DEST_STAT):
3516 		{
3517 			if (dhd_pub->debug_buf_dest_support) {
3518 				debug_buf_dest_stat_t *debug_buf_dest_stat =
3519 					(debug_buf_dest_stat_t *)arg;
3520 				memcpy(debug_buf_dest_stat, dhd_pub->debug_buf_dest_stat,
3521 					sizeof(dhd_pub->debug_buf_dest_stat));
3522 			} else {
3523 				bcmerror = BCME_DISABLED;
3524 			}
3525 			break;
3526 		}
3527 
3528 #ifdef DHD_PKTTS
3529 	case IOV_GVAL(IOV_PKTTS_ENAB): {
3530 		int_val = dhd_get_pktts_enab(dhd_pub);
3531 		(void)memcpy_s(arg, val_size, &int_val, sizeof(int_val));
3532 		break;
3533 	}
3534 	case IOV_SVAL(IOV_PKTTS_ENAB): {
3535 		dhd_set_pktts_enab(dhd_pub, !!int_val);
3536 		break;
3537 	}
3538 
3539 	case IOV_GVAL(IOV_PKTTS_FLOW): {
3540 		bcmerror = dhd_get_pktts_flow(dhd_pub, arg, len);
3541 		break;
3542 	}
3543 	case IOV_SVAL(IOV_PKTTS_FLOW): {
3544 		bcmerror = dhd_set_pktts_flow(dhd_pub, params, plen);
3545 		break;
3546 	}
3547 #endif /* DHD_PKTTS */
3548 
3549 #if defined(DHD_EFI)
3550 	case IOV_SVAL(IOV_INTR_POLL):
3551 		bcmerror = dhd_intr_poll(dhd_pub->bus, arg, len, TRUE);
3552 		break;
3553 
3554 	case IOV_GVAL(IOV_INTR_POLL):
3555 		bcmerror = dhd_intr_poll(dhd_pub->bus, params, plen, FALSE);
3556 		break;
3557 #endif /* DHD_EFI */
3558 
3559 #if defined(DHD_SSSR_DUMP)
3560 	case IOV_GVAL(IOV_FIS_TRIGGER):
3561 		bcmerror = dhd_bus_fis_trigger(dhd_pub);
3562 
3563 		if (bcmerror == BCME_OK) {
3564 			bcmerror = dhd_bus_fis_dump(dhd_pub);
3565 		}
3566 
3567 		int_val = bcmerror;
3568 		bcopy(&int_val, arg, val_size);
3569 		break;
3570 #endif /* defined(DHD_SSSR_DUMP) */
3571 
3572 #ifdef DHD_DEBUG
3573 	case IOV_SVAL(IOV_INDUCE_ERROR): {
3574 		if (int_val >= DHD_INDUCE_ERROR_MAX) {
3575 			DHD_ERROR(("%s: Invalid command : %u\n", __FUNCTION__, (uint16)int_val));
3576 		} else {
3577 			dhd_pub->dhd_induce_error = (uint16)int_val;
3578 #ifdef BCMPCIE
3579 			if (dhd_pub->dhd_induce_error == DHD_INDUCE_BH_CBP_HANG) {
3580 				dhdpcie_induce_cbp_hang(dhd_pub);
3581 			}
3582 #endif /* BCMPCIE */
3583 		}
3584 		break;
3585 	}
3586 #endif /* DHD_DEBUG */
3587 #ifdef WL_IFACE_MGMT_CONF
3588 #ifdef WL_CFG80211
3589 #ifdef WL_NANP2P
3590 	case IOV_GVAL(IOV_CONC_DISC): {
3591 		int_val = wl_cfg80211_get_iface_conc_disc(
3592 			dhd_linux_get_primary_netdev(dhd_pub));
3593 		bcopy(&int_val, arg, sizeof(int_val));
3594 		break;
3595 	}
3596 	case IOV_SVAL(IOV_CONC_DISC): {
3597 		bcmerror = wl_cfg80211_set_iface_conc_disc(
3598 			dhd_linux_get_primary_netdev(dhd_pub), (uint8)int_val);
3599 		break;
3600 	}
3601 #endif /* WL_NANP2P */
3602 #ifdef WL_IFACE_MGMT
3603 	case IOV_GVAL(IOV_IFACE_POLICY): {
3604 		int_val = wl_cfg80211_get_iface_policy(
3605 		dhd_linux_get_primary_netdev(dhd_pub));
3606 		bcopy(&int_val, arg, sizeof(int_val));
3607 		break;
3608 	}
3609 	case IOV_SVAL(IOV_IFACE_POLICY): {
3610 		bcmerror = wl_cfg80211_set_iface_policy(
3611 			dhd_linux_get_primary_netdev(dhd_pub),
3612 			arg, len);
3613 		break;
3614 	}
3615 #endif /* WL_IFACE_MGMT */
3616 #endif /* WL_CFG80211 */
3617 #endif /* WL_IFACE_MGMT_CONF */
3618 #ifdef RTT_GEOFENCE_CONT
3619 #if defined (RTT_SUPPORT) && defined (WL_NAN)
3620 	case IOV_GVAL(IOV_RTT_GEOFENCE_TYPE_OVRD): {
3621 		bool enable = 0;
3622 		dhd_rtt_get_geofence_cont_ind(dhd_pub, &enable);
3623 		int_val = enable ? 1 : 0;
3624 		bcopy(&int_val, arg, val_size);
3625 		break;
3626 	}
3627 	case IOV_SVAL(IOV_RTT_GEOFENCE_TYPE_OVRD): {
3628 		bool enable = *(bool *)arg;
3629 		dhd_rtt_set_geofence_cont_ind(dhd_pub, enable);
3630 		break;
3631 	}
3632 #endif /* RTT_SUPPORT && WL_NAN */
3633 #endif /* RTT_GEOFENCE_CONT */
3634 	case IOV_GVAL(IOV_FW_VBS): {
3635 		*(uint32 *)arg = (uint32)dhd_dbg_get_fwverbose(dhd_pub);
3636 		break;
3637 	}
3638 
3639 	case IOV_SVAL(IOV_FW_VBS): {
3640 		if (int_val < 0) {
3641 			int_val = 0;
3642 		}
3643 		dhd_dbg_set_fwverbose(dhd_pub, (uint32)int_val);
3644 		break;
3645 	}
3646 
3647 #ifdef DHD_TX_PROFILE
3648 	case IOV_SVAL(IOV_TX_PROFILE_TAG):
3649 	{
3650 		/* note: under the current implementation only one type of packet may be
3651 		 * tagged per profile
3652 		 */
3653 		const dhd_tx_profile_protocol_t *protocol = NULL;
3654 		/* for example, we might have a profile of profile_index 6, but at
3655 		 * offset 2 from dhd_pub->protocol_filters.
3656 		 */
3657 		uint8 offset;
3658 
3659 		if (params == NULL) {
3660 			bcmerror = BCME_ERROR;
3661 			break;
3662 		}
3663 
3664 		protocol = (dhd_tx_profile_protocol_t *)params;
3665 
3666 		/* validate */
3667 		if (protocol->version != DHD_TX_PROFILE_VERSION) {
3668 			bcmerror = BCME_VERSION;
3669 			break;
3670 		}
3671 		if (protocol->profile_index > DHD_MAX_PROFILE_INDEX) {
3672 			DHD_ERROR(("%s:\tprofile index must be between 0 and %d\n",
3673 					__FUNCTION__, DHD_MAX_PROFILE_INDEX));
3674 			bcmerror = BCME_RANGE;
3675 			break;
3676 		}
3677 		if (protocol->layer != DHD_TX_PROFILE_DATA_LINK_LAYER && protocol->layer
3678 				!= DHD_TX_PROFILE_NETWORK_LAYER) {
3679 			DHD_ERROR(("%s:\tlayer must be %d or %d\n", __FUNCTION__,
3680 					DHD_TX_PROFILE_DATA_LINK_LAYER,
3681 					DHD_TX_PROFILE_NETWORK_LAYER));
3682 			bcmerror = BCME_BADARG;
3683 			break;
3684 		}
3685 		if (protocol->protocol_number > __UINT16_MAX__) {
3686 			DHD_ERROR(("%s:\tprotocol number must be <= %d\n", __FUNCTION__,
3687 					__UINT16_MAX__));
3688 			bcmerror = BCME_BADLEN;
3689 			break;
3690 		}
3691 
3692 		/* find the dhd_tx_profile_protocol_t */
3693 		for (offset = 0; offset < dhd_pub->num_profiles; offset++) {
3694 			if (dhd_pub->protocol_filters[offset].profile_index ==
3695 					protocol->profile_index) {
3696 				break;
3697 			}
3698 		}
3699 
3700 		if (offset >= DHD_MAX_PROFILES) {
3701 #if DHD_MAX_PROFILES > 1
3702 			DHD_ERROR(("%s:\tonly %d profiles supported at present\n",
3703 					__FUNCTION__, DHD_MAX_PROFILES));
3704 #else /* DHD_MAX_PROFILES > 1 */
3705 			DHD_ERROR(("%s:\tonly %d profile supported at present\n",
3706 					__FUNCTION__, DHD_MAX_PROFILES));
3707 			DHD_ERROR(("%s:\tthere is a profile of index %d\n", __FUNCTION__,
3708 					dhd_pub->protocol_filters->profile_index));
3709 #endif /* DHD_MAX_PROFILES > 1 */
3710 			bcmerror = BCME_NOMEM;
3711 			break;
3712 		}
3713 
3714 		/* memory already allocated in dhd_attach; just assign the value */
3715 		dhd_pub->protocol_filters[offset] = *protocol;
3716 
3717 		if (offset >= dhd_pub->num_profiles) {
3718 			dhd_pub->num_profiles = offset + 1;
3719 		}
3720 
3721 		break;
3722 	}
3723 
3724 	case IOV_SVAL(IOV_TX_PROFILE_ENABLE):
3725 		dhd_pub->tx_profile_enab = int_val ? TRUE : FALSE;
3726 		break;
3727 
3728 	case IOV_GVAL(IOV_TX_PROFILE_ENABLE):
3729 		int_val = dhd_pub->tx_profile_enab;
3730 		bcmerror = memcpy_s(arg, val_size, &int_val, sizeof(int_val));
3731 		break;
3732 
3733 	case IOV_SVAL(IOV_TX_PROFILE_DUMP):
3734 	{
3735 		const dhd_tx_profile_protocol_t *protocol = NULL;
3736 		uint8 offset;
3737 		char *format = "%s:\ttx_profile %s: %d\n";
3738 
3739 		for (offset = 0; offset < dhd_pub->num_profiles; offset++) {
3740 			if (dhd_pub->protocol_filters[offset].profile_index == int_val) {
3741 				protocol = &(dhd_pub->protocol_filters[offset]);
3742 				break;
3743 			}
3744 		}
3745 
3746 		if (protocol == NULL) {
3747 			DHD_ERROR(("%s:\tno profile with index %d\n", __FUNCTION__,
3748 					int_val));
3749 			bcmerror = BCME_ERROR;
3750 			break;
3751 		}
3752 
3753 		printf(format, __FUNCTION__, "profile_index", protocol->profile_index);
3754 		printf(format, __FUNCTION__, "layer", protocol->layer);
3755 		printf(format, __FUNCTION__, "protocol_number", protocol->protocol_number);
3756 		printf(format, __FUNCTION__, "src_port", protocol->src_port);
3757 		printf(format, __FUNCTION__, "dest_port", protocol->dest_port);
3758 
3759 		break;
3760 	}
3761 #endif /* defined(DHD_TX_PROFILE) */
3762 
3763 	case IOV_GVAL(IOV_CHECK_TRAP_ROT): {
3764 		int_val = dhd_pub->check_trap_rot? 1 : 0;
3765 		(void)memcpy_s(arg, val_size, &int_val, sizeof(int_val));
3766 		break;
3767 	}
3768 	case IOV_SVAL(IOV_CHECK_TRAP_ROT): {
3769 		dhd_pub->check_trap_rot = *(bool *)arg;
3770 		break;
3771 	}
3772 
3773 #if defined(DHD_AWDL)
3774 	case IOV_SVAL(IOV_AWDL_LLC_ENABLE): {
3775 		bool bval = *(bool *)arg;
3776 		if (bval != 0 && bval != 1)
3777 			bcmerror = BCME_ERROR;
3778 		else
3779 			dhd_pub->awdl_llc_enabled = bval;
3780 		break;
3781 	}
3782 	case IOV_GVAL(IOV_AWDL_LLC_ENABLE):
3783 		int_val = dhd_pub->awdl_llc_enabled;
3784 		(void)memcpy_s(arg, val_size, &int_val, sizeof(int_val));
3785 		break;
3786 #endif
3787 #ifdef WLEASYMESH
3788 	case IOV_SVAL(IOV_1905_AL_UCAST): {
3789 		uint32  bssidx;
3790 		const char *val;
3791 		uint8 ea[6] = {0};
3792 		if (dhd_iovar_parse_bssidx(dhd_pub, (char *)name, &bssidx, &val) != BCME_OK) {
3793 			DHD_ERROR(("%s: 1905_al_ucast: bad parameter\n", __FUNCTION__));
3794 			bcmerror = BCME_BADARG;
3795 			break;
3796 		}
3797 		bcopy(val, ea, ETHER_ADDR_LEN);
3798 		printf("IOV_1905_AL_UCAST:" MACDBG "\n", MAC2STRDBG(ea));
3799 		bcmerror = dhd_set_1905_almac(dhd_pub, bssidx, ea, FALSE);
3800 		break;
3801 	}
3802 	case IOV_GVAL(IOV_1905_AL_UCAST): {
3803 		uint32  bssidx;
3804 		const char *val;
3805 		if (dhd_iovar_parse_bssidx(dhd_pub, (char *)name, &bssidx, &val) != BCME_OK) {
3806 			DHD_ERROR(("%s: 1905_al_ucast: bad parameter\n", __FUNCTION__));
3807 			bcmerror = BCME_BADARG;
3808 			break;
3809 		}
3810 
3811 		bcmerror = dhd_get_1905_almac(dhd_pub, bssidx, arg, FALSE);
3812 		break;
3813 	}
3814 	case IOV_SVAL(IOV_1905_AL_MCAST): {
3815 		uint32  bssidx;
3816 		const char *val;
3817 		uint8 ea[6] = {0};
3818 		if (dhd_iovar_parse_bssidx(dhd_pub, (char *)name, &bssidx, &val) != BCME_OK) {
3819 			DHD_ERROR(("%s: 1905_al_mcast: bad parameter\n", __FUNCTION__));
3820 			bcmerror = BCME_BADARG;
3821 			break;
3822 		}
3823 		bcopy(val, ea, ETHER_ADDR_LEN);
3824 		printf("IOV_1905_AL_MCAST:" MACDBG "\n", MAC2STRDBG(ea));
3825 		bcmerror = dhd_set_1905_almac(dhd_pub, bssidx, ea, TRUE);
3826 		break;
3827 	}
3828 	case IOV_GVAL(IOV_1905_AL_MCAST): {
3829 		uint32  bssidx;
3830 		const char *val;
3831 		if (dhd_iovar_parse_bssidx(dhd_pub, (char *)name, &bssidx, &val) != BCME_OK) {
3832 			DHD_ERROR(("%s: 1905_al_mcast: bad parameter\n", __FUNCTION__));
3833 			bcmerror = BCME_BADARG;
3834 			break;
3835 		}
3836 
3837 		bcmerror = dhd_get_1905_almac(dhd_pub, bssidx, arg, TRUE);
3838 		break;
3839 	}
3840 #endif /* WLEASYMESH */
3841 
3842 	default:
3843 		bcmerror = BCME_UNSUPPORTED;
3844 		break;
3845 	}
3846 
3847 exit:
3848 	DHD_TRACE(("%s: actionid %d, bcmerror %d\n", __FUNCTION__, actionid, bcmerror));
3849 	return bcmerror;
3850 }
3851 
3852 #ifdef BCMDONGLEHOST
3853 /* Store the status of a connection attempt for later retrieval by an iovar */
3854 void
dhd_store_conn_status(uint32 event,uint32 status,uint32 reason)3855 dhd_store_conn_status(uint32 event, uint32 status, uint32 reason)
3856 {
3857 	/* Do not overwrite a WLC_E_PRUNE with a WLC_E_SET_SSID
3858 	 * because an encryption/rsn mismatch results in both events, and
3859 	 * the important information is in the WLC_E_PRUNE.
3860 	 */
3861 	if (!(event == WLC_E_SET_SSID && status == WLC_E_STATUS_FAIL &&
3862 	      dhd_conn_event == WLC_E_PRUNE)) {
3863 		dhd_conn_event = event;
3864 		dhd_conn_status = status;
3865 		dhd_conn_reason = reason;
3866 	}
3867 }
3868 #else
3869 #error "BCMDONGLEHOST not defined"
3870 #endif /* BCMDONGLEHOST */
3871 
3872 bool
dhd_prec_enq(dhd_pub_t * dhdp,struct pktq * q,void * pkt,int prec)3873 dhd_prec_enq(dhd_pub_t *dhdp, struct pktq *q, void *pkt, int prec)
3874 {
3875 	void *p;
3876 	int eprec = -1;		/* precedence to evict from */
3877 	bool discard_oldest;
3878 
3879 	/* Fast case, precedence queue is not full and we are also not
3880 	 * exceeding total queue length
3881 	 */
3882 	if (!pktqprec_full(q, prec) && !pktq_full(q)) {
3883 		pktq_penq(q, prec, pkt);
3884 		return TRUE;
3885 	}
3886 
3887 	/* Determine precedence from which to evict packet, if any */
3888 	if (pktqprec_full(q, prec))
3889 		eprec = prec;
3890 	else if (pktq_full(q)) {
3891 		p = pktq_peek_tail(q, &eprec);
3892 		ASSERT(p);
3893 		if (eprec > prec || eprec < 0)
3894 			return FALSE;
3895 	}
3896 
3897 	/* Evict if needed */
3898 	if (eprec >= 0) {
3899 		/* Detect queueing to unconfigured precedence */
3900 		ASSERT(!pktqprec_empty(q, eprec));
3901 		discard_oldest = AC_BITMAP_TST(dhdp->wme_dp, eprec);
3902 		if (eprec == prec && !discard_oldest)
3903 			return FALSE;		/* refuse newer (incoming) packet */
3904 		/* Evict packet according to discard policy */
3905 		p = discard_oldest ? pktq_pdeq(q, eprec) : pktq_pdeq_tail(q, eprec);
3906 		ASSERT(p);
3907 #ifdef DHDTCPACK_SUPPRESS
3908 		if (dhd_tcpack_check_xmit(dhdp, p) == BCME_ERROR) {
3909 			DHD_ERROR(("%s %d: tcpack_suppress ERROR!!! Stop using it\n",
3910 				__FUNCTION__, __LINE__));
3911 			dhd_tcpack_suppress_set(dhdp, TCPACK_SUP_OFF);
3912 		}
3913 #endif /* DHDTCPACK_SUPPRESS */
3914 		PKTFREE(dhdp->osh, p, TRUE);
3915 	}
3916 
3917 	/* Enqueue */
3918 	p = pktq_penq(q, prec, pkt);
3919 	ASSERT(p);
3920 
3921 	return TRUE;
3922 }
3923 
3924 /*
3925  * Functions to drop proper pkts from queue:
3926  *	If one pkt in queue is non-fragmented, drop first non-fragmented pkt only
3927  *	If all pkts in queue are all fragmented, find and drop one whole set fragmented pkts
3928  *	If can't find pkts matching upper 2 cases, drop first pkt anyway
3929  */
3930 bool
dhd_prec_drop_pkts(dhd_pub_t * dhdp,struct pktq * pq,int prec,f_droppkt_t fn)3931 dhd_prec_drop_pkts(dhd_pub_t *dhdp, struct pktq *pq, int prec, f_droppkt_t fn)
3932 {
3933 	struct pktq_prec *q = NULL;
3934 	void *p, *prev = NULL, *next = NULL, *first = NULL, *last = NULL, *prev_first = NULL;
3935 	pkt_frag_t frag_info;
3936 
3937 	ASSERT(dhdp && pq);
3938 	ASSERT(prec >= 0 && prec < pq->num_prec);
3939 
3940 	q = &pq->q[prec];
3941 	p = q->head;
3942 
3943 	if (p == NULL)
3944 		return FALSE;
3945 
3946 	while (p) {
3947 		frag_info = pkt_frag_info(dhdp->osh, p);
3948 		if (frag_info == DHD_PKT_FRAG_NONE) {
3949 			break;
3950 		} else if (frag_info == DHD_PKT_FRAG_FIRST) {
3951 			if (first) {
3952 				/* No last frag pkt, use prev as last */
3953 				last = prev;
3954 				break;
3955 			} else {
3956 				first = p;
3957 				prev_first = prev;
3958 			}
3959 		} else if (frag_info == DHD_PKT_FRAG_LAST) {
3960 			if (first) {
3961 				last = p;
3962 				break;
3963 			}
3964 		}
3965 
3966 		prev = p;
3967 		p = PKTLINK(p);
3968 	}
3969 
3970 	if ((p == NULL) || ((frag_info != DHD_PKT_FRAG_NONE) && !(first && last))) {
3971 		/* Not found matching pkts, use oldest */
3972 		prev = NULL;
3973 		p = q->head;
3974 		frag_info = 0;
3975 	}
3976 
3977 	if (frag_info == DHD_PKT_FRAG_NONE) {
3978 		first = last = p;
3979 		prev_first = prev;
3980 	}
3981 
3982 	p = first;
3983 	while (p) {
3984 		next = PKTLINK(p);
3985 		q->n_pkts--;
3986 		pq->n_pkts_tot--;
3987 
3988 #ifdef WL_TXQ_STALL
3989 		q->dequeue_count++;
3990 #endif
3991 
3992 		PKTSETLINK(p, NULL);
3993 
3994 		if (fn)
3995 			fn(dhdp, prec, p, TRUE);
3996 
3997 		if (p == last)
3998 			break;
3999 
4000 		p = next;
4001 	}
4002 
4003 	if (prev_first == NULL) {
4004 		if ((q->head = next) == NULL)
4005 			q->tail = NULL;
4006 	} else {
4007 		PKTSETLINK(prev_first, next);
4008 		if (!next)
4009 			q->tail = prev_first;
4010 	}
4011 
4012 	return TRUE;
4013 }
4014 
4015 static int
dhd_iovar_op(dhd_pub_t * dhd_pub,const char * name,void * params,int plen,void * arg,uint len,bool set)4016 dhd_iovar_op(dhd_pub_t *dhd_pub, const char *name,
4017 	void *params, int plen, void *arg, uint len, bool set)
4018 {
4019 	int bcmerror = 0;
4020 	uint val_size;
4021 	const bcm_iovar_t *vi = NULL;
4022 	uint32 actionid;
4023 
4024 	DHD_TRACE(("%s: Enter\n", __FUNCTION__));
4025 
4026 	ASSERT(name);
4027 
4028 	/* Get MUST have return space */
4029 	ASSERT(set || (arg && len));
4030 
4031 	/* Set does NOT take qualifiers */
4032 	ASSERT(!set || (!params && !plen));
4033 
4034 	if ((vi = bcm_iovar_lookup(dhd_iovars, name)) == NULL) {
4035 		bcmerror = BCME_UNSUPPORTED;
4036 		goto exit;
4037 	}
4038 
4039 	DHD_CTL(("%s: %s %s, len %d plen %d\n", __FUNCTION__,
4040 		name, (set ? "set" : "get"), len, plen));
4041 
4042 	/* set up 'params' pointer in case this is a set command so that
4043 	 * the convenience int and bool code can be common to set and get
4044 	 */
4045 	if (params == NULL) {
4046 		params = arg;
4047 		plen = len;
4048 	}
4049 
4050 	if (vi->type == IOVT_VOID)
4051 		val_size = 0;
4052 	else if (vi->type == IOVT_BUFFER)
4053 		val_size = len;
4054 	else
4055 		/* all other types are integer sized */
4056 		val_size = sizeof(int);
4057 
4058 	actionid = set ? IOV_SVAL(vi->varid) : IOV_GVAL(vi->varid);
4059 
4060 	bcmerror = dhd_doiovar(dhd_pub, vi, actionid, name, params, plen, arg, len, val_size);
4061 
4062 exit:
4063 	return bcmerror;
4064 }
4065 
4066 int
dhd_ioctl(dhd_pub_t * dhd_pub,dhd_ioctl_t * ioc,void * buf,uint buflen)4067 dhd_ioctl(dhd_pub_t * dhd_pub, dhd_ioctl_t *ioc, void *buf, uint buflen)
4068 {
4069 	int bcmerror = 0;
4070 	unsigned long flags;
4071 
4072 	DHD_TRACE(("%s: Enter\n", __FUNCTION__));
4073 
4074 	if (!buf) {
4075 		return BCME_BADARG;
4076 	}
4077 
4078 	dhd_os_dhdiovar_lock(dhd_pub);
4079 	switch (ioc->cmd) {
4080 		case DHD_GET_MAGIC:
4081 			if (buflen < sizeof(int))
4082 				bcmerror = BCME_BUFTOOSHORT;
4083 			else
4084 				*(int*)buf = DHD_IOCTL_MAGIC;
4085 			break;
4086 
4087 		case DHD_GET_VERSION:
4088 			if (buflen < sizeof(int))
4089 				bcmerror = BCME_BUFTOOSHORT;
4090 			else
4091 				*(int*)buf = DHD_IOCTL_VERSION;
4092 			break;
4093 
4094 		case DHD_GET_VAR:
4095 		case DHD_SET_VAR:
4096 			{
4097 				char *arg;
4098 				uint arglen;
4099 
4100 				DHD_LINUX_GENERAL_LOCK(dhd_pub, flags);
4101 				if (DHD_BUS_CHECK_DOWN_OR_DOWN_IN_PROGRESS(dhd_pub) &&
4102 					bcmstricmp((char *)buf, "devreset")) {
4103 					/* In platforms like FC19, the FW download is done via IOCTL
4104 					 * and should not return error for IOCTLs fired before FW
4105 					 * Download is done
4106 					 */
4107 					if (dhd_fw_download_status(dhd_pub) == FW_DOWNLOAD_DONE) {
4108 						DHD_ERROR(("%s: return as fw_download_status=%d\n",
4109 							__FUNCTION__,
4110 							dhd_fw_download_status(dhd_pub)));
4111 						DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
4112 						dhd_os_dhdiovar_unlock(dhd_pub);
4113 						return -ENODEV;
4114 					}
4115 				}
4116 				DHD_BUS_BUSY_SET_IN_DHD_IOVAR(dhd_pub);
4117 				DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
4118 
4119 #ifdef DHD_PCIE_RUNTIMEPM
4120 				dhdpcie_runtime_bus_wake(dhd_pub, TRUE, dhd_ioctl);
4121 #endif /* DHD_PCIE_RUNTIMEPM */
4122 
4123 				DHD_LINUX_GENERAL_LOCK(dhd_pub, flags);
4124 				if (DHD_BUS_CHECK_SUSPEND_OR_SUSPEND_IN_PROGRESS(dhd_pub)) {
4125 					/* If Suspend/Resume is tested via pcie_suspend IOVAR
4126 					 * then continue to execute the IOVAR, return from here for
4127 					 * other IOVARs, also include pciecfgreg and devreset to go
4128 					 * through.
4129 					 */
4130 #ifdef DHD_EFI
4131 					if (bcmstricmp((char *)buf, "pcie_suspend") &&
4132 						bcmstricmp((char *)buf, "pciecfgreg") &&
4133 						bcmstricmp((char *)buf, "devreset") &&
4134 						bcmstricmp((char *)buf, "sdio_suspend") &&
4135 						bcmstricmp((char *)buf, "control_signal"))
4136 #else
4137 					if (bcmstricmp((char *)buf, "pcie_suspend") &&
4138 					    bcmstricmp((char *)buf, "pciecfgreg") &&
4139 					    bcmstricmp((char *)buf, "devreset") &&
4140 					    bcmstricmp((char *)buf, "sdio_suspend"))
4141 #endif /* DHD_EFI */
4142 					{
4143 						DHD_ERROR(("%s: bus is in suspend(%d)"
4144 							"or suspending(0x%x) state\n",
4145 							__FUNCTION__, dhd_pub->busstate,
4146 							dhd_pub->dhd_bus_busy_state));
4147 						DHD_BUS_BUSY_CLEAR_IN_DHD_IOVAR(dhd_pub);
4148 						dhd_os_busbusy_wake(dhd_pub);
4149 						DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
4150 						dhd_os_dhdiovar_unlock(dhd_pub);
4151 						return -ENODEV;
4152 					}
4153 				}
4154 				/* During devreset ioctl, we call dhdpcie_advertise_bus_cleanup,
4155 				 * which will wait for all the busy contexts to get over for
4156 				 * particular time and call ASSERT if timeout happens. As during
4157 				 * devreset ioctal, we made DHD_BUS_BUSY_SET_IN_DHD_IOVAR,
4158 				 * to avoid ASSERT, clear the IOCTL busy state. "devreset" ioctl is
4159 				 * not used in Production platforms but only used in FC19 setups.
4160 				 */
4161 				if (!bcmstricmp((char *)buf, "devreset") ||
4162 #ifdef BCMPCIE
4163 				    (dhd_bus_is_multibp_capable(dhd_pub->bus) &&
4164 				    !bcmstricmp((char *)buf, "dwnldstate")) ||
4165 #endif /* BCMPCIE */
4166 #if defined(DHD_EFI) && defined (BT_OVER_PCIE)
4167 					!bcmstricmp((char *)buf, "btop_test") ||
4168 					!bcmstricmp((char *)buf, "control_signal") ||
4169 #endif /* DHD_EFI && BT_OVER_PCIE */
4170 				    FALSE)
4171 				{
4172 					DHD_BUS_BUSY_CLEAR_IN_DHD_IOVAR(dhd_pub);
4173 				}
4174 				DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
4175 
4176 				/* scan past the name to any arguments */
4177 				for (arg = buf, arglen = buflen; *arg && arglen; arg++, arglen--)
4178 					;
4179 
4180 				if (arglen == 0 || *arg) {
4181 					bcmerror = BCME_BUFTOOSHORT;
4182 					goto unlock_exit;
4183 				}
4184 
4185 				/* account for the NUL terminator */
4186 				arg++, arglen--;
4187 				/* call with the appropriate arguments */
4188 				if (ioc->cmd == DHD_GET_VAR) {
4189 					bcmerror = dhd_iovar_op(dhd_pub, buf, arg, arglen,
4190 							buf, buflen, IOV_GET);
4191 				} else {
4192 					bcmerror = dhd_iovar_op(dhd_pub, buf, NULL, 0,
4193 							arg, arglen, IOV_SET);
4194 				}
4195 				if (bcmerror != BCME_UNSUPPORTED) {
4196 					goto unlock_exit;
4197 				}
4198 
4199 				/* not in generic table, try protocol module */
4200 				if (ioc->cmd == DHD_GET_VAR) {
4201 					bcmerror = dhd_prot_iovar_op(dhd_pub, buf, arg,
4202 							arglen, buf, buflen, IOV_GET);
4203 				} else {
4204 					bcmerror = dhd_prot_iovar_op(dhd_pub, buf,
4205 							NULL, 0, arg, arglen, IOV_SET);
4206 				}
4207 				if (bcmerror != BCME_UNSUPPORTED) {
4208 					goto unlock_exit;
4209 				}
4210 
4211 				/* if still not found, try bus module */
4212 				if (ioc->cmd == DHD_GET_VAR) {
4213 					bcmerror = dhd_bus_iovar_op(dhd_pub, buf,
4214 							arg, arglen, buf, buflen, IOV_GET);
4215 				} else {
4216 					bcmerror = dhd_bus_iovar_op(dhd_pub, buf,
4217 							NULL, 0, arg, arglen, IOV_SET);
4218 				}
4219 				if (bcmerror != BCME_UNSUPPORTED) {
4220 					goto unlock_exit;
4221 				}
4222 
4223 #ifdef DHD_TIMESYNC
4224 				/* check TS module */
4225 				if (ioc->cmd == DHD_GET_VAR)
4226 					bcmerror = dhd_timesync_iovar_op(dhd_pub->ts, buf, arg,
4227 						arglen, buf, buflen, IOV_GET);
4228 				else
4229 					bcmerror = dhd_timesync_iovar_op(dhd_pub->ts, buf,
4230 						NULL, 0, arg, arglen, IOV_SET);
4231 #endif /* DHD_TIMESYNC */
4232 			}
4233 			goto unlock_exit;
4234 
4235 		default:
4236 			bcmerror = BCME_UNSUPPORTED;
4237 	}
4238 	dhd_os_dhdiovar_unlock(dhd_pub);
4239 	return bcmerror;
4240 
4241 unlock_exit:
4242 	DHD_LINUX_GENERAL_LOCK(dhd_pub, flags);
4243 	DHD_BUS_BUSY_CLEAR_IN_DHD_IOVAR(dhd_pub);
4244 	dhd_os_busbusy_wake(dhd_pub);
4245 	DHD_LINUX_GENERAL_UNLOCK(dhd_pub, flags);
4246 	dhd_os_dhdiovar_unlock(dhd_pub);
4247 	return bcmerror;
4248 }
4249 
4250 #ifdef SHOW_EVENTS
4251 
4252 #if defined(DHD_AWDL) && defined(AWDL_SLOT_STATS)
4253 static void
dhd_update_awdl_stats(dhd_pub_t * dhd_pub,const awdl_aws_event_data_t * aw)4254 dhd_update_awdl_stats(dhd_pub_t *dhd_pub, const awdl_aws_event_data_t *aw)
4255 {
4256 	dhd_awdl_stats_t *awdl_stats;
4257 	unsigned long lock_flags;
4258 
4259 	/* since AWDL stats are read on clear to protect against clear,
4260 	 * lock before update
4261 	 */
4262 	DHD_AWDL_STATS_LOCK(dhd_pub->awdl_stats_lock, lock_flags);
4263 	/* Start of AWDL slot */
4264 	if (!(aw->flags & AWDL_AW_LAST_EXT)) {
4265 		dhd_pub->awdl_tx_status_slot =
4266 			((aw->aw_counter/AWDL_SLOT_MULT) % AWDL_NUM_SLOTS);
4267 		awdl_stats = &dhd_pub->awdl_stats[dhd_pub->awdl_tx_status_slot];
4268 		awdl_stats->slot_start_time = OSL_SYSUPTIME_US();
4269 		awdl_stats->fw_slot_start_time = ntoh32_ua(&aw->fw_time);
4270 		awdl_stats->num_slots++;
4271 	} else {
4272 		/* End of AWDL slot */
4273 		awdl_stats = &dhd_pub->awdl_stats[dhd_pub->awdl_tx_status_slot];
4274 		if (awdl_stats->slot_start_time) {
4275 			awdl_stats->cum_slot_time +=
4276 				OSL_SYSUPTIME_US() - awdl_stats->slot_start_time;
4277 			/* FW reports time in us in a 32bit number.
4278 			 * This 32bit number wrap-arround in ~90 minutes.
4279 			 * Below logic considers wrap-arround too
4280 			 */
4281 			awdl_stats->fw_cum_slot_time +=
4282 				((ntoh32_ua(&aw->fw_time) - awdl_stats->fw_slot_start_time) &
4283 					(UINT_MAX));
4284 
4285 		}
4286 	}
4287 	DHD_AWDL_STATS_UNLOCK(dhd_pub->awdl_stats_lock, lock_flags);
4288 }
4289 #endif /* DHD_AWDL && AWDL_SLOT_STATS */
4290 
4291 static void
wl_show_roam_event(dhd_pub_t * dhd_pub,uint status,uint datalen,const char * event_name,char * eabuf,void * event_data)4292 wl_show_roam_event(dhd_pub_t *dhd_pub, uint status, uint datalen,
4293 	const char *event_name, char *eabuf, void *event_data)
4294 {
4295 #ifdef REPORT_FATAL_TIMEOUTS
4296 	OSL_ATOMIC_SET(dhd_pub->osh, &dhd_pub->set_ssid_rcvd, TRUE);
4297 	dhd_clear_join_error(dhd_pub, WLC_SSID_MASK);
4298 #endif /* REPORT_FATAL_TIMEOUTS */
4299 	if (status == WLC_E_STATUS_SUCCESS) {
4300 		DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
4301 	} else {
4302 #ifdef REPORT_FATAL_TIMEOUTS
4303 		/*
4304 		 * For secure join if WLC_E_SET_SSID returns with any failure case,
4305 		 * donot expect WLC_E_PSK_SUP. So clear the mask.
4306 		 */
4307 		dhd_clear_join_error(dhd_pub, WLC_WPA_MASK);
4308 #endif /* REPORT_FATAL_TIMEOUTS */
4309 		if (status == WLC_E_STATUS_FAIL) {
4310 			DHD_EVENT(("MACEVENT: %s, failed status %d\n", event_name, status));
4311 		} else if (status == WLC_E_STATUS_NO_NETWORKS) {
4312 			if (datalen) {
4313 				uint8 id = *((uint8 *)event_data);
4314 				if (id != DOT11_MNG_PROPR_ID) {
4315 					wl_roam_event_t *roam_data =
4316 						(wl_roam_event_t *)event_data;
4317 					bcm_xtlv_t *tlv = (bcm_xtlv_t *)roam_data->xtlvs;
4318 					if (tlv->id == WLC_ROAM_NO_NETWORKS_TLV_ID) {
4319 						uint32 *fail_reason = (uint32 *)tlv->data;
4320 						switch (*fail_reason) {
4321 							case WLC_E_REASON_NO_NETWORKS:
4322 								DHD_EVENT(("MACEVENT: %s,"
4323 									" no networks found\n",
4324 									event_name));
4325 								break;
4326 							case WLC_E_REASON_NO_NETWORKS_BY_SCORE:
4327 								DHD_EVENT(("MACEVENT: %s,"
4328 								" no networks found by score\n",
4329 									event_name));
4330 								break;
4331 							default:
4332 								DHD_ERROR(("MACEVENT: %s,"
4333 								" unknown fail reason 0x%x\n",
4334 									event_name,
4335 									*fail_reason));
4336 								ASSERT(0);
4337 						}
4338 					} else {
4339 						DHD_EVENT(("MACEVENT: %s,"
4340 							" no networks found\n",
4341 							event_name));
4342 					}
4343 				} else {
4344 					DHD_EVENT(("MACEVENT: %s,"
4345 						" no networks found\n",
4346 						event_name));
4347 				}
4348 			} else {
4349 				DHD_EVENT(("MACEVENT: %s, no networks found\n",
4350 					event_name));
4351 			}
4352 		} else {
4353 			DHD_EVENT(("MACEVENT: %s, unexpected status %d\n",
4354 				event_name, (int)status));
4355 		}
4356 	}
4357 }
4358 
4359 static void
wl_show_roam_cache_update_event(const char * name,uint status,uint reason,uint datalen,void * event_data)4360 wl_show_roam_cache_update_event(const char *name, uint status,
4361 	uint reason, uint datalen, void *event_data)
4362 {
4363 	wlc_roam_cache_update_event_t *cache_update;
4364 	uint16 len_of_tlvs;
4365 	void *val_tlv_ptr;
4366 	bcm_xtlv_t *val_xtlv;
4367 	char ntoa_buf[ETHER_ADDR_STR_LEN];
4368 	uint idx;
4369 	const char* reason_name = NULL;
4370 	const char* status_name = NULL;
4371 	static struct {
4372 		uint event;
4373 		const char *event_name;
4374 	} reason_names[] = {
4375 		{WLC_E_REASON_INITIAL_ASSOC, "INITIAL ASSOCIATION"},
4376 		{WLC_E_REASON_LOW_RSSI, "LOW_RSSI"},
4377 		{WLC_E_REASON_DEAUTH, "RECEIVED DEAUTHENTICATION"},
4378 		{WLC_E_REASON_DISASSOC, "RECEIVED DISASSOCATION"},
4379 		{WLC_E_REASON_BCNS_LOST, "BEACONS LOST"},
4380 		{WLC_E_REASON_BETTER_AP, "BETTER AP FOUND"},
4381 		{WLC_E_REASON_MINTXRATE, "STUCK AT MIN TX RATE"},
4382 		{WLC_E_REASON_BSSTRANS_REQ, "REQUESTED ROAM"},
4383 		{WLC_E_REASON_TXFAIL, "TOO MANY TXFAILURES"}
4384 	};
4385 
4386 	static struct {
4387 		uint event;
4388 		const char *event_name;
4389 	} status_names[] = {
4390 		{WLC_E_STATUS_SUCCESS, "operation was successful"},
4391 		{WLC_E_STATUS_FAIL, "operation failed"},
4392 		{WLC_E_STATUS_TIMEOUT, "operation timed out"},
4393 		{WLC_E_STATUS_NO_NETWORKS, "failed due to no matching network found"},
4394 		{WLC_E_STATUS_ABORT, "operation was aborted"},
4395 		{WLC_E_STATUS_NO_ACK, "protocol failure: packet not ack'd"},
4396 		{WLC_E_STATUS_UNSOLICITED, "AUTH or ASSOC packet was unsolicited"},
4397 		{WLC_E_STATUS_ATTEMPT, "attempt to assoc to an auto auth configuration"},
4398 		{WLC_E_STATUS_PARTIAL, "scan results are incomplete"},
4399 		{WLC_E_STATUS_NEWSCAN, "scan aborted by another scan"},
4400 		{WLC_E_STATUS_NEWASSOC, "scan aborted due to assoc in progress"},
4401 		{WLC_E_STATUS_11HQUIET, "802.11h quiet period started"},
4402 		{WLC_E_STATUS_SUPPRESS, "user disabled scanning"},
4403 		{WLC_E_STATUS_NOCHANS, "no allowable channels to scan"},
4404 		{WLC_E_STATUS_CS_ABORT, "abort channel select"},
4405 		{WLC_E_STATUS_ERROR, "request failed due to error"},
4406 		{WLC_E_STATUS_INVALID, "Invalid status code"}
4407 	};
4408 
4409 	switch (reason) {
4410 	case WLC_ROAM_CACHE_UPDATE_NEW_ROAM_CACHE:
4411 		DHD_EVENT(("Current roam cache status %d, "
4412 			"reason for cache update is new roam cache\n", status));
4413 		break;
4414 	case WLC_ROAM_CACHE_UPDATE_JOIN:
4415 		DHD_EVENT(("Current roam cache status %d, "
4416 			"reason for cache update is start of join\n", status));
4417 		break;
4418 	case WLC_ROAM_CACHE_UPDATE_RSSI_DELTA:
4419 		DHD_EVENT(("Current roam cache status %d, "
4420 			"reason for cache update is delta in rssi\n", status));
4421 		break;
4422 	case WLC_ROAM_CACHE_UPDATE_MOTION_RSSI_DELTA:
4423 		DHD_EVENT(("Current roam cache status %d, "
4424 			"reason for cache update is motion delta in rssi\n", status));
4425 		break;
4426 	case WLC_ROAM_CACHE_UPDATE_CHANNEL_MISS:
4427 		DHD_EVENT(("Current roam cache status %d, "
4428 			"reason for cache update is missed channel\n", status));
4429 		break;
4430 	case WLC_ROAM_CACHE_UPDATE_START_SPLIT_SCAN:
4431 		DHD_EVENT(("Current roam cache status %d, "
4432 			"reason for cache update is start of split scan\n", status));
4433 		break;
4434 	case WLC_ROAM_CACHE_UPDATE_START_FULL_SCAN:
4435 		DHD_EVENT(("Current roam cache status %d, "
4436 			"reason for cache update is start of full scan\n", status));
4437 		break;
4438 	case WLC_ROAM_CACHE_UPDATE_INIT_ASSOC:
4439 		DHD_EVENT(("Current roam cache status %d, "
4440 			"reason for cache update is init association\n", status));
4441 		break;
4442 	case WLC_ROAM_CACHE_UPDATE_FULL_SCAN_FAILED:
4443 		DHD_EVENT(("Current roam cache status %d, "
4444 			"reason for cache update is failure in full scan\n", status));
4445 		break;
4446 	case WLC_ROAM_CACHE_UPDATE_NO_AP_FOUND:
4447 		DHD_EVENT(("Current roam cache status %d, "
4448 			"reason for cache update is empty scan result\n", status));
4449 		break;
4450 	case WLC_ROAM_CACHE_UPDATE_MISSING_AP:
4451 		DHD_EVENT(("Current roam cache status %d, "
4452 			"reason for cache update is missed ap\n", status));
4453 		break;
4454 	default:
4455 		DHD_EVENT(("Current roam cache status %d, "
4456 			"reason for cache update is unknown %d\n", status, reason));
4457 		break;
4458 	}
4459 
4460 	if (datalen < sizeof(wlc_roam_cache_update_event_t)) {
4461 		DHD_ERROR(("MACEVENT: %s, missing event data\n", name));
4462 		return;
4463 	}
4464 
4465 	cache_update = (wlc_roam_cache_update_event_t *)event_data;
4466 	val_tlv_ptr = (void *)cache_update->xtlvs;
4467 	len_of_tlvs = datalen - sizeof(wlc_roam_cache_update_event_t);
4468 	val_xtlv = (bcm_xtlv_t *)val_tlv_ptr;
4469 	if (val_xtlv->id != WL_RMC_RPT_CMD_DATA) {
4470 		DHD_ERROR(("MACEVENT: %s, unexpected xtlv id %d\n",
4471 			name, val_xtlv->id));
4472 		return;
4473 	}
4474 	val_tlv_ptr = (uint8 *)val_tlv_ptr + BCM_XTLV_HDR_SIZE;
4475 	len_of_tlvs = val_xtlv->len;
4476 
4477 	while (len_of_tlvs && len_of_tlvs > BCM_XTLV_HDR_SIZE) {
4478 		val_xtlv = (bcm_xtlv_t *)val_tlv_ptr;
4479 		switch (val_xtlv->id) {
4480 			case WL_RMC_RPT_XTLV_BSS_INFO:
4481 			{
4482 				rmc_bss_info_v1_t *bss_info = (rmc_bss_info_v1_t *)(val_xtlv->data);
4483 				DHD_EVENT(("\t Current BSS INFO:\n"));
4484 				DHD_EVENT(("\t\tRSSI: %d\n", bss_info->rssi));
4485 				DHD_EVENT(("\t\tNumber of full scans performed "
4486 					"on current BSS: %d\n", bss_info->fullscan_count));
4487 				for (idx = 0; idx < ARRAYSIZE(reason_names); idx++) {
4488 					if (reason_names[idx].event == bss_info->reason) {
4489 						reason_name = reason_names[idx].event_name;
4490 					}
4491 				}
4492 				DHD_EVENT(("\t\tReason code for last full scan: %s(%d)\n",
4493 					reason_name, bss_info->reason));
4494 				DHD_EVENT(("\t\tDelta between current time and "
4495 					"last full scan: %d\n", bss_info->time_full_scan));
4496 				for (idx = 0; idx < ARRAYSIZE(status_names); idx++) {
4497 					if (status_names[idx].event == bss_info->status)
4498 						status_name = status_names[idx].event_name;
4499 				}
4500 				DHD_EVENT(("\t\tLast status code for not roaming: %s(%d)\n",
4501 					status_name, bss_info->status));
4502 
4503 			}
4504 				break;
4505 			case WL_RMC_RPT_XTLV_CANDIDATE_INFO:
4506 			case WL_RMC_RPT_XTLV_USER_CACHE_INFO:
4507 			{
4508 				rmc_candidate_info_v1_t *candidate_info =
4509 					(rmc_candidate_info_v1_t *)(val_xtlv->data);
4510 				if (val_xtlv->id == WL_RMC_RPT_XTLV_CANDIDATE_INFO) {
4511 					DHD_EVENT(("\t Candidate INFO:\n"));
4512 				} else {
4513 					DHD_EVENT(("\t User Candidate INFO:\n"));
4514 				}
4515 				DHD_EVENT(("\t\tBSSID: %s\n",
4516 					bcm_ether_ntoa((const struct ether_addr *)
4517 					&candidate_info->bssid, ntoa_buf)));
4518 				DHD_EVENT(("\t\tRSSI: %d\n", candidate_info->rssi));
4519 				DHD_EVENT(("\t\tChannel: %d\n", candidate_info->ctl_channel));
4520 				DHD_EVENT(("\t\tDelta between current time and last "
4521 					"seen time: %d\n", candidate_info->time_last_seen));
4522 				DHD_EVENT(("\t\tBSS load: %d\n", candidate_info->bss_load));
4523 			}
4524 				break;
4525 			default:
4526 				DHD_ERROR(("MACEVENT: %s, unexpected xtlv id %d\n",
4527 					name, val_xtlv->id));
4528 				return;
4529 		}
4530 		val_tlv_ptr = (uint8 *)val_tlv_ptr + bcm_xtlv_size(val_xtlv,
4531 			BCM_XTLV_OPTION_NONE);
4532 		len_of_tlvs -= (uint16)bcm_xtlv_size(val_xtlv, BCM_XTLV_OPTION_NONE);
4533 	}
4534 }
4535 
4536 static void
wl_show_host_event(dhd_pub_t * dhd_pub,wl_event_msg_t * event,void * event_data,void * raw_event_ptr,char * eventmask)4537 wl_show_host_event(dhd_pub_t *dhd_pub, wl_event_msg_t *event, void *event_data,
4538 	void *raw_event_ptr, char *eventmask)
4539 {
4540 	uint i, status, reason;
4541 	bool group = FALSE, flush_txq = FALSE, link = FALSE;
4542 	bool host_data = FALSE; /* prints  event data after the case  when set */
4543 	const char *auth_str;
4544 	const char *event_name;
4545 	const uchar *buf;
4546 	char err_msg[256], eabuf[ETHER_ADDR_STR_LEN];
4547 	uint event_type, flags, auth_type, datalen;
4548 
4549 	event_type = ntoh32(event->event_type);
4550 	flags = ntoh16(event->flags);
4551 	status = ntoh32(event->status);
4552 	reason = ntoh32(event->reason);
4553 	BCM_REFERENCE(reason);
4554 	auth_type = ntoh32(event->auth_type);
4555 	datalen = (event_data != NULL) ? ntoh32(event->datalen) : 0;
4556 
4557 	/* debug dump of event messages */
4558 	snprintf(eabuf, sizeof(eabuf), MACDBG, MAC2STRDBG(event->addr.octet));
4559 
4560 	event_name = bcmevent_get_name(event_type);
4561 	BCM_REFERENCE(event_name);
4562 
4563 	if (flags & WLC_EVENT_MSG_LINK)
4564 		link = TRUE;
4565 	if (flags & WLC_EVENT_MSG_GROUP)
4566 		group = TRUE;
4567 	if (flags & WLC_EVENT_MSG_FLUSHTXQ)
4568 		flush_txq = TRUE;
4569 
4570 	switch (event_type) {
4571 	case WLC_E_START:
4572 		DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
4573 		break;
4574 	case WLC_E_DEAUTH:
4575 	case WLC_E_DISASSOC:
4576 		DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
4577 #ifdef REPORT_FATAL_TIMEOUTS
4578 		dhd_clear_join_error(dhd_pub, WLC_SSID_MASK | WLC_WPA_MASK);
4579 #endif /* REPORT_FATAL_TIMEOUTS */
4580 		break;
4581 
4582 	case WLC_E_ASSOC_IND:
4583 	case WLC_E_REASSOC_IND:
4584 
4585 		DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
4586 #ifdef REPORT_FATAL_TIMEOUTS
4587 		if (status != WLC_E_STATUS_SUCCESS) {
4588 			dhd_clear_join_error(dhd_pub, WLC_SSID_MASK | WLC_WPA_MASK);
4589 		}
4590 #endif /* REPORT_FATAL_TIMEOUTS */
4591 
4592 		break;
4593 
4594 	case WLC_E_ASSOC:
4595 	case WLC_E_REASSOC:
4596 		if (status == WLC_E_STATUS_SUCCESS) {
4597 			DHD_EVENT(("MACEVENT: %s, MAC %s, SUCCESS\n", event_name, eabuf));
4598 		} else if (status == WLC_E_STATUS_TIMEOUT) {
4599 			DHD_EVENT(("MACEVENT: %s, MAC %s, TIMEOUT\n", event_name, eabuf));
4600 		} else if (status == WLC_E_STATUS_FAIL) {
4601 			DHD_EVENT(("MACEVENT: %s, MAC %s, FAILURE, status %d reason %d\n",
4602 			       event_name, eabuf, (int)status, (int)reason));
4603 		} else if (status == WLC_E_STATUS_SUPPRESS) {
4604 			DHD_EVENT(("MACEVENT: %s, MAC %s, SUPPRESS\n", event_name, eabuf));
4605 		} else if (status == WLC_E_STATUS_NO_ACK) {
4606 			DHD_EVENT(("MACEVENT: %s, MAC %s, NOACK\n", event_name, eabuf));
4607 		} else {
4608 			DHD_EVENT(("MACEVENT: %s, MAC %s, unexpected status %d\n",
4609 			       event_name, eabuf, (int)status));
4610 		}
4611 #ifdef REPORT_FATAL_TIMEOUTS
4612 		if (status != WLC_E_STATUS_SUCCESS) {
4613 			dhd_clear_join_error(dhd_pub, WLC_SSID_MASK | WLC_WPA_MASK);
4614 		}
4615 #endif /* REPORT_FATAL_TIMEOUTS */
4616 
4617 		break;
4618 
4619 	case WLC_E_DEAUTH_IND:
4620 	case WLC_E_DISASSOC_IND:
4621 #ifdef REPORT_FATAL_TIMEOUTS
4622 		dhd_clear_join_error(dhd_pub, WLC_SSID_MASK | WLC_WPA_MASK);
4623 #endif /* REPORT_FATAL_TIMEOUTS */
4624 		DHD_EVENT(("MACEVENT: %s, MAC %s, reason %d\n", event_name, eabuf, (int)reason));
4625 		break;
4626 
4627 	case WLC_E_AUTH:
4628 	case WLC_E_AUTH_IND:
4629 		if (auth_type == DOT11_OPEN_SYSTEM)
4630 			auth_str = "Open System";
4631 		else if (auth_type == DOT11_SHARED_KEY)
4632 			auth_str = "Shared Key";
4633 		else if (auth_type == DOT11_SAE)
4634 			auth_str = "SAE";
4635 		else {
4636 			snprintf(err_msg, sizeof(err_msg), "AUTH unknown: %d", (int)auth_type);
4637 			auth_str = err_msg;
4638 		}
4639 
4640 		if (event_type == WLC_E_AUTH_IND) {
4641 			DHD_EVENT(("MACEVENT: %s, MAC %s, %s\n", event_name, eabuf, auth_str));
4642 		} else if (status == WLC_E_STATUS_SUCCESS) {
4643 			DHD_EVENT(("MACEVENT: %s, MAC %s, %s, SUCCESS\n",
4644 				event_name, eabuf, auth_str));
4645 		} else if (status == WLC_E_STATUS_TIMEOUT) {
4646 			DHD_EVENT(("MACEVENT: %s, MAC %s, %s, TIMEOUT\n",
4647 				event_name, eabuf, auth_str));
4648 		} else if (status == WLC_E_STATUS_FAIL) {
4649 			DHD_EVENT(("MACEVENT: %s, MAC %s, %s, FAILURE, status %d reason %d\n",
4650 			       event_name, eabuf, auth_str, (int)status, (int)reason));
4651 		} else if (status == WLC_E_STATUS_SUPPRESS) {
4652 			DHD_EVENT(("MACEVENT: %s, MAC %s, %s, SUPPRESS\n",
4653 			       event_name, eabuf, auth_str));
4654 		} else if (status == WLC_E_STATUS_NO_ACK) {
4655 			DHD_EVENT(("MACEVENT: %s, MAC %s, %s, NOACK\n",
4656 			       event_name, eabuf, auth_str));
4657 		} else {
4658 			DHD_EVENT(("MACEVENT: %s, MAC %s, %s, status %d reason %d\n",
4659 				event_name, eabuf, auth_str, (int)status, (int)reason));
4660 		}
4661 		BCM_REFERENCE(auth_str);
4662 #ifdef REPORT_FATAL_TIMEOUTS
4663 		if (status != WLC_E_STATUS_SUCCESS) {
4664 			dhd_clear_join_error(dhd_pub, WLC_SSID_MASK | WLC_WPA_MASK);
4665 		}
4666 #endif /* REPORT_FATAL_TIMEOUTS */
4667 
4668 		break;
4669 
4670 	case WLC_E_ROAM:
4671 		wl_show_roam_event(dhd_pub, status, datalen,
4672 			event_name, eabuf, event_data);
4673 		break;
4674 	case WLC_E_ROAM_START:
4675 		if (datalen >= sizeof(wlc_roam_start_event_t)) {
4676 			const wlc_roam_start_event_t *roam_start =
4677 				(wlc_roam_start_event_t *)event_data;
4678 			DHD_EVENT(("MACEVENT: %s %d, MAC %s, status %d,"
4679 				" reason %d, auth %d, current bss rssi %d\n",
4680 				event_name, event_type, eabuf, (int)status, (int)reason,
4681 				(int)auth_type, (int)roam_start->rssi));
4682 		} else {
4683 			DHD_EVENT(("MACEVENT: %s %d, MAC %s, status %d, reason %d, auth %d\n",
4684 				event_name, event_type, eabuf, (int)status, (int)reason,
4685 				(int)auth_type));
4686 		}
4687 		break;
4688 	case WLC_E_ROAM_PREP:
4689 		if (datalen >= sizeof(wlc_roam_prep_event_t)) {
4690 			const wlc_roam_prep_event_t *roam_prep =
4691 				(wlc_roam_prep_event_t *)event_data;
4692 			DHD_EVENT(("MACEVENT: %s %d, MAC %s, status %d,"
4693 				" reason %d, auth %d, target bss rssi %d\n",
4694 				event_name, event_type, eabuf, (int)status, (int)reason,
4695 				(int)auth_type, (int)roam_prep->rssi));
4696 		} else {
4697 			DHD_EVENT(("MACEVENT: %s %d, MAC %s, status %d, reason %d, auth %d\n",
4698 				event_name, event_type, eabuf, (int)status, (int)reason,
4699 				(int)auth_type));
4700 		}
4701 		break;
4702 	case WLC_E_ROAM_CACHE_UPDATE:
4703 		DHD_EVENT(("MACEVENT: %s\n", event_name));
4704 		wl_show_roam_cache_update_event(event_name, status,
4705 			reason, datalen, event_data);
4706 		break;
4707 	case WLC_E_JOIN:
4708 	case WLC_E_SET_SSID:
4709 #ifdef REPORT_FATAL_TIMEOUTS
4710 		OSL_ATOMIC_SET(dhd_pub->osh, &dhd_pub->set_ssid_rcvd, TRUE);
4711 		dhd_clear_join_error(dhd_pub, WLC_SSID_MASK);
4712 #endif /* REPORT_FATAL_TIMEOUTS */
4713 		if (status == WLC_E_STATUS_SUCCESS) {
4714 			DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
4715 		} else {
4716 #ifdef REPORT_FATAL_TIMEOUTS
4717 			/*
4718 			 * For secure join if WLC_E_SET_SSID returns with any failure case,
4719 			 * donot expect WLC_E_PSK_SUP. So clear the mask.
4720 			 */
4721 			dhd_clear_join_error(dhd_pub, WLC_WPA_MASK);
4722 #endif /* REPORT_FATAL_TIMEOUTS */
4723 			if (status == WLC_E_STATUS_FAIL) {
4724 				DHD_EVENT(("MACEVENT: %s, failed status %d\n", event_name, status));
4725 			} else if (status == WLC_E_STATUS_NO_NETWORKS) {
4726 				DHD_EVENT(("MACEVENT: %s, no networks found\n", event_name));
4727 			} else {
4728 				DHD_EVENT(("MACEVENT: %s, unexpected status %d\n",
4729 					event_name, (int)status));
4730 			}
4731 		}
4732 		break;
4733 
4734 	case WLC_E_BEACON_RX:
4735 		if (status == WLC_E_STATUS_SUCCESS) {
4736 			DHD_EVENT(("MACEVENT: %s, SUCCESS\n", event_name));
4737 		} else if (status == WLC_E_STATUS_FAIL) {
4738 			DHD_EVENT(("MACEVENT: %s, FAIL\n", event_name));
4739 		} else {
4740 			DHD_EVENT(("MACEVENT: %s, status %d\n", event_name, status));
4741 		}
4742 		break;
4743 
4744 	case WLC_E_LINK:
4745 		DHD_EVENT(("MACEVENT: %s %s flags:0x%x status:%d reason:%d\n",
4746 			event_name, link?"UP":"DOWN", flags, status, reason));
4747 #ifdef PCIE_FULL_DONGLE
4748 #ifdef REPORT_FATAL_TIMEOUTS
4749 		{
4750 			uint8 ifindex = (uint8)dhd_ifname2idx(dhd_pub->info, event->ifname);
4751 			uint8 role = dhd_flow_rings_ifindex2role(dhd_pub, ifindex);
4752 			if ((role == WLC_E_IF_ROLE_STA) && (!link)) {
4753 				dhd_clear_join_error(dhd_pub, WLC_SSID_MASK | WLC_WPA_MASK);
4754 			}
4755 		}
4756 #endif /* PCIE_FULL_DONGLE */
4757 #endif /* REPORT_FATAL_TIMEOUTS */
4758 		BCM_REFERENCE(link);
4759 		break;
4760 
4761 	case WLC_E_MIC_ERROR:
4762 		DHD_EVENT(("MACEVENT: %s, MAC %s, Group %d, Flush %d\n",
4763 		       event_name, eabuf, group, flush_txq));
4764 		BCM_REFERENCE(group);
4765 		BCM_REFERENCE(flush_txq);
4766 		break;
4767 
4768 	case WLC_E_ICV_ERROR:
4769 	case WLC_E_UNICAST_DECODE_ERROR:
4770 	case WLC_E_MULTICAST_DECODE_ERROR:
4771 		DHD_EVENT(("MACEVENT: %s, MAC %s\n",
4772 		       event_name, eabuf));
4773 		break;
4774 
4775 	case WLC_E_TXFAIL:
4776 		DHD_EVENT(("MACEVENT: %s, RA %s status %d\n", event_name, eabuf, status));
4777 		break;
4778 
4779 	case WLC_E_ASSOC_REQ_IE:
4780 	case WLC_E_ASSOC_RESP_IE:
4781 	case WLC_E_PMKID_CACHE:
4782 		DHD_EVENT(("MACEVENT: %s\n", event_name));
4783 		break;
4784 
4785 	case WLC_E_SCAN_COMPLETE:
4786 		DHD_EVENT(("MACEVENT: %s\n", event_name));
4787 #ifdef REPORT_FATAL_TIMEOUTS
4788 		dhd_stop_scan_timer(dhd_pub, FALSE, 0);
4789 #endif /* REPORT_FATAL_TIMEOUTS */
4790 		break;
4791 	case WLC_E_RSSI_LQM:
4792 	case WLC_E_PFN_NET_FOUND:
4793 	case WLC_E_PFN_NET_LOST:
4794 	case WLC_E_PFN_SCAN_COMPLETE:
4795 	case WLC_E_PFN_SCAN_NONE:
4796 	case WLC_E_PFN_SCAN_ALLGONE:
4797 	case WLC_E_PFN_GSCAN_FULL_RESULT:
4798 	case WLC_E_PFN_SSID_EXT:
4799 		DHD_EVENT(("PNOEVENT: %s\n", event_name));
4800 		break;
4801 
4802 	case WLC_E_PFN_SCAN_BACKOFF:
4803 	case WLC_E_PFN_BSSID_SCAN_BACKOFF:
4804 		DHD_EVENT(("PNOEVENT: %s, status %d, reason %d\n",
4805 		           event_name, (int)status, (int)reason));
4806 		break;
4807 
4808 	case WLC_E_PSK_SUP:
4809 	case WLC_E_PRUNE:
4810 		DHD_EVENT(("MACEVENT: %s, status %d, reason %d\n",
4811 		           event_name, (int)status, (int)reason));
4812 #ifdef REPORT_FATAL_TIMEOUTS
4813 		dhd_clear_join_error(dhd_pub, WLC_WPA_MASK);
4814 #endif /* REPORT_FATAL_TIMEOUTS */
4815 		break;
4816 
4817 #ifdef WIFI_ACT_FRAME
4818 	case WLC_E_ACTION_FRAME:
4819 		DHD_TRACE(("MACEVENT: %s Bssid %s\n", event_name, eabuf));
4820 		break;
4821 	case WLC_E_ACTION_FRAME_COMPLETE:
4822 		if (datalen >= sizeof(uint32)) {
4823 			const uint32 *pktid = event_data;
4824 			BCM_REFERENCE(pktid);
4825 			DHD_EVENT(("MACEVENT: %s status %d, reason %d, pktid 0x%x\n",
4826 				event_name, (int)status, (int)reason, *pktid));
4827 		}
4828 		break;
4829 #endif /* WIFI_ACT_FRAME */
4830 
4831 #ifdef SHOW_LOGTRACE
4832 	case WLC_E_TRACE:
4833 	{
4834 		dhd_dbg_trace_evnt_handler(dhd_pub, event_data, raw_event_ptr, datalen);
4835 		break;
4836 	}
4837 #endif /* SHOW_LOGTRACE */
4838 
4839 	case WLC_E_RSSI:
4840 		if (datalen >= sizeof(int)) {
4841 			DHD_EVENT(("MACEVENT: %s %d\n", event_name, ntoh32(*((int *)event_data))));
4842 		}
4843 		break;
4844 
4845 	case WLC_E_SERVICE_FOUND:
4846 	case WLC_E_P2PO_ADD_DEVICE:
4847 	case WLC_E_P2PO_DEL_DEVICE:
4848 		DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
4849 		break;
4850 
4851 #ifdef BT_WIFI_HANDOBER
4852 	case WLC_E_BT_WIFI_HANDOVER_REQ:
4853 		DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
4854 		break;
4855 #endif
4856 #ifdef DHD_AWDL
4857 	case WLC_E_AWDL_AW:
4858 		if (datalen >= sizeof(awdl_aws_event_data_t)) {
4859 			const awdl_aws_event_data_t *aw =
4860 				(awdl_aws_event_data_t *)event_data;
4861 			BCM_REFERENCE(aw);
4862 			DHD_EVENT(("MACEVENT: %s, MAC %s aw_cnt %u ext_cnt %u flags %u "
4863 					"aw_ch %u\n", event_name, eabuf, aw->aw_counter,
4864 					aw->aw_ext_count, aw->flags, CHSPEC_CHANNEL(aw->aw_chan)));
4865 			host_data = TRUE;
4866 
4867 #if defined(DHD_AWDL) && defined(AWDL_SLOT_STATS)
4868 			dhd_update_awdl_stats(dhd_pub, aw);
4869 			/* Store last received aw counter */
4870 			dhd_pub->awdl_aw_counter = aw->aw_counter;
4871 #endif /* DHD_AWDL */
4872 		}
4873 		break;
4874 	case WLC_E_AWDL_ROLE:
4875 		DHD_EVENT(("MACEVENT: %s, MAC %s ROLE %d\n", event_name, eabuf, (int)status));
4876 		break;
4877 	case WLC_E_AWDL_EVENT:
4878 		DHD_EVENT(("MACEVENT: %s, MAC %s status %d reason %d\n",
4879 			event_name, eabuf, (int)status, (int)reason));
4880 		if (datalen >= OFFSETOF(awdl_scan_event_data_t, chan_list)) {
4881 			const awdl_scan_event_data_t *scan_evt =
4882 				(awdl_scan_event_data_t *)event_data;
4883 			BCM_REFERENCE(scan_evt);
4884 			DHD_EVENT(("scan_usage %d, nscan_chans %d, ncached_chans %d, "
4885 				"iscan_flags 0x%x\n", scan_evt->scan_usage,
4886 				scan_evt->nscan_chans, scan_evt->ncached_chans,
4887 				scan_evt->flags));
4888 			host_data = TRUE;
4889 		}
4890 		break;
4891 #endif /* DHD_AWDL  */
4892 
4893 	case WLC_E_CCA_CHAN_QUAL:
4894 		/* I would like to check here that datalen >= sizeof(cca_chan_qual_event_t)
4895 		 * but since definition of cca_chan_qual_event_t is different
4896 		 * between blazar and legacy firmware, I will
4897 		 * check only that datalen is bigger than 0.
4898 		 */
4899 		if (datalen > 0) {
4900 			const cca_chan_qual_event_t *cca_event =
4901 				(cca_chan_qual_event_t *)event_data;
4902 			if ((cca_event->id == WL_CHAN_QUAL_FULLPM_CCA) ||
4903 			    (cca_event->id == WL_CHAN_QUAL_FULLPM_CCA_OFDM_DESENSE)) {
4904 				const cca_only_chan_qual_event_t *cca_only_event =
4905 					(const cca_only_chan_qual_event_t *)cca_event;
4906 				BCM_REFERENCE(cca_only_event);
4907 				DHD_EVENT((
4908 					"MACEVENT: %s %d, MAC %s, status %d, reason %d, auth %d,"
4909 					" channel 0x%02x\n",
4910 					event_name, event_type, eabuf, (int)status,
4911 					(int)reason, (int)auth_type, cca_event->chanspec));
4912 				DHD_EVENT((
4913 					"\tTOTAL (dur %dms me %dms notme %dms interf %dms"
4914 					" ts 0x%08x)\n",
4915 					cca_only_event->cca_busy_ext.duration,
4916 					cca_only_event->cca_busy_ext.congest_ibss,
4917 					cca_only_event->cca_busy_ext.congest_obss,
4918 					cca_only_event->cca_busy_ext.interference,
4919 					cca_only_event->cca_busy_ext.timestamp));
4920 				DHD_EVENT((
4921 					"\t  !PM (dur %dms me %dms notme %dms interf %dms)\n",
4922 					cca_only_event->cca_busy_nopm.duration,
4923 					cca_only_event->cca_busy_nopm.congest_ibss,
4924 					cca_only_event->cca_busy_nopm.congest_obss,
4925 					cca_only_event->cca_busy_nopm.interference));
4926 				DHD_EVENT((
4927 					"\t   PM (dur %dms me %dms notme %dms interf %dms)\n",
4928 					cca_only_event->cca_busy_pm.duration,
4929 					cca_only_event->cca_busy_pm.congest_ibss,
4930 					cca_only_event->cca_busy_pm.congest_obss,
4931 					cca_only_event->cca_busy_pm.interference));
4932 				if (cca_event->id == WL_CHAN_QUAL_FULLPM_CCA_OFDM_DESENSE) {
4933 					DHD_EVENT(("\t OFDM desense %d\n",
4934 						((const cca_only_chan_qual_event_v2_t *)
4935 						cca_only_event)->ofdm_desense));
4936 				}
4937 			} else if (cca_event->id == WL_CHAN_QUAL_FULL_CCA) {
4938 				DHD_EVENT((
4939 					"MACEVENT: %s %d, MAC %s, status %d, reason %d, auth %d,"
4940 					" channel 0x%02x (dur %dms ibss %dms obss %dms interf %dms"
4941 					" ts 0x%08x)\n",
4942 					event_name, event_type, eabuf, (int)status,
4943 					(int)reason, (int)auth_type, cca_event->chanspec,
4944 					cca_event->cca_busy_ext.duration,
4945 					cca_event->cca_busy_ext.congest_ibss,
4946 					cca_event->cca_busy_ext.congest_obss,
4947 					cca_event->cca_busy_ext.interference,
4948 					cca_event->cca_busy_ext.timestamp));
4949 			} else if (cca_event->id == WL_CHAN_QUAL_CCA) {
4950 				DHD_EVENT((
4951 					"MACEVENT: %s %d, MAC %s, status %d, reason %d, auth %d,"
4952 					" channel 0x%02x (dur %dms busy %dms ts 0x%08x)\n",
4953 					event_name, event_type, eabuf, (int)status,
4954 					(int)reason, (int)auth_type, cca_event->chanspec,
4955 					cca_event->cca_busy.duration,
4956 					cca_event->cca_busy.congest,
4957 					cca_event->cca_busy.timestamp));
4958 			} else if ((cca_event->id == WL_CHAN_QUAL_NF) ||
4959 			           (cca_event->id == WL_CHAN_QUAL_NF_LTE)) {
4960 				DHD_EVENT((
4961 					"MACEVENT: %s %d, MAC %s, status %d, reason %d, auth %d,"
4962 					" channel 0x%02x (NF[%d] %ddB)\n",
4963 					event_name, event_type, eabuf, (int)status,
4964 					(int)reason, (int)auth_type, cca_event->chanspec,
4965 					cca_event->id, cca_event->noise));
4966 			} else {
4967 				DHD_EVENT((
4968 					"MACEVENT: %s %d, MAC %s, status %d, reason %d, auth %d,"
4969 					" channel 0x%02x (unknown ID %d)\n",
4970 					event_name, event_type, eabuf, (int)status,
4971 					(int)reason, (int)auth_type, cca_event->chanspec,
4972 					cca_event->id));
4973 			}
4974 		}
4975 		break;
4976 	case WLC_E_ESCAN_RESULT:
4977 		if (datalen >= sizeof(wl_escan_result_v2_t)) {
4978 			const wl_escan_result_v2_t *escan_result =
4979 				(wl_escan_result_v2_t *)event_data;
4980 			BCM_REFERENCE(escan_result);
4981 #ifdef OEM_ANDROID
4982 			/* Because WLC_E_ESCAN_RESULT event log are being print too many.
4983 			* So, DHD_EVENT() changes to be used DHD_TRACE() in HW4 platform.
4984 			*/
4985 			DHD_EVENT(("MACEVENT: %s %d, MAC %s, status %d \n",
4986 				event_name, event_type, eabuf, (int)status));
4987 #else
4988 			DHD_EVENT(("MACEVENT: %s %d, MAC %s, status %d sync-id %u\n",
4989 				event_name, event_type, eabuf,
4990 				(int)status, dtoh16(escan_result->sync_id)));
4991 #endif /* CUSTOMER_HW4 */
4992 #ifdef REPORT_FATAL_TIMEOUTS
4993 			/* a 'partial' status means the escan is still in progress
4994 			* any other status implies the escan has either finished or aborted
4995 			*/
4996 			if (status != WLC_E_STATUS_PARTIAL) {
4997 				unsigned long timeout_flags = 0;
4998 				uint16 syncid = dtoh16(escan_result->sync_id);
4999 				/* this is to take care of the specific case where
5000 				* escan event returns abort and is processed immediately
5001 				* by dhd before the escan iovar has returned. In that case
5002 				* if the iovar returns success, then we will be starting a
5003 				* timeout even though the escan has already been aborted !
5004 				* So the flag below is checked before starting the escan timeout
5005 				*/
5006 				if (dhd_pub->timeout_info) {
5007 					DHD_TIMER_LOCK(dhd_pub->timeout_info->scan_timer_lock,
5008 						timeout_flags);
5009 					if (!dhd_pub->timeout_info->scan_timer_active &&
5010 						syncid == dhd_pub->esync_id) {
5011 						dhd_pub->timeout_info->escan_aborted = TRUE;
5012 						dhd_pub->timeout_info->abort_syncid = syncid;
5013 						DHD_TIMER_UNLOCK(
5014 							dhd_pub->timeout_info->scan_timer_lock,
5015 							timeout_flags);
5016 						break;
5017 					} else {
5018 						dhd_pub->timeout_info->escan_aborted = FALSE;
5019 					}
5020 					DHD_TIMER_UNLOCK(dhd_pub->timeout_info->scan_timer_lock,
5021 						timeout_flags);
5022 				}
5023 				dhd_stop_scan_timer(dhd_pub, TRUE, dtoh16(escan_result->sync_id));
5024 			}
5025 #endif /* REPORT_FATAL_TIMEOUTS */
5026 		}
5027 		break;
5028 	case WLC_E_IF:
5029 		if (datalen >= sizeof(struct wl_event_data_if)) {
5030 			const struct wl_event_data_if *ifevent =
5031 				(struct wl_event_data_if *)event_data;
5032 			BCM_REFERENCE(ifevent);
5033 
5034 			DHD_EVENT(("MACEVENT: %s, opcode:0x%d  ifidx:%d role:%d\n",
5035 				event_name, ifevent->opcode, ifevent->ifidx, ifevent->role));
5036 		}
5037 		break;
5038 #ifdef SHOW_LOGTRACE
5039 	case WLC_E_MSCH:
5040 	{
5041 		wl_mschdbg_event_handler(dhd_pub, raw_event_ptr, reason, event_data, datalen);
5042 		break;
5043 	}
5044 #endif /* SHOW_LOGTRACE */
5045 
5046 	case WLC_E_PSK_AUTH:
5047 		DHD_EVENT(("MACEVENT: %s, RA %s status %d Reason:%d\n",
5048 			event_name, eabuf, status, reason));
5049 		break;
5050 	case WLC_E_AGGR_EVENT:
5051 		if (datalen >= sizeof(event_aggr_data_t)) {
5052 			const event_aggr_data_t *aggrbuf = event_data;
5053 			int j = 0, len = 0;
5054 			const uint8 *data = aggrbuf->data;
5055 			DHD_EVENT(("MACEVENT: %s, num of events %d total len %d sub events: ",
5056 				event_name, aggrbuf->num_events, aggrbuf->len));
5057 			for (j = 0; j < aggrbuf->num_events; j++)
5058 			{
5059 				const wl_event_msg_t * sub_event = (const wl_event_msg_t *)data;
5060 				if (len > aggrbuf->len) {
5061 					DHD_ERROR(("%s: Aggr events corrupted!",
5062 						__FUNCTION__));
5063 					break;
5064 				}
5065 				DHD_EVENT(("\n Event type: %d ", ntoh32(sub_event->event_type)));
5066 				len += ALIGN_SIZE((ntoh32(sub_event->datalen) +
5067 					sizeof(wl_event_msg_t)), sizeof(uint64));
5068 				buf = (const uchar *)(data + sizeof(wl_event_msg_t));
5069 				BCM_REFERENCE(buf);
5070 				DHD_EVENT((" data (%d) : ", ntoh32(sub_event->datalen)));
5071 				for (i = 0; i < ntoh32(sub_event->datalen); i++) {
5072 					DHD_EVENT((" 0x%02x ", buf[i]));
5073 				}
5074 				data = aggrbuf->data + len;
5075 			}
5076 			DHD_EVENT(("\n"));
5077 		}
5078 		break;
5079 	case WLC_E_PHY_CAL:
5080 		{
5081 			DHD_EVENT(("MACEVENT: %s, reason:%d\n", event_name, reason));
5082 			break;
5083 		}
5084 	case WLC_E_NAN_CRITICAL:
5085 		{
5086 			DHD_EVENT(("MACEVENT: %s, type:%d\n", event_name, reason));
5087 			break;
5088 		}
5089 	case WLC_E_NAN_NON_CRITICAL:
5090 		{
5091 			DHD_TRACE(("MACEVENT: %s, type:%d\n", event_name, reason));
5092 			break;
5093 		}
5094 	case WLC_E_PROXD:
5095 		if (datalen >= sizeof(wl_proxd_event_t)) {
5096 			const wl_proxd_event_t *proxd =
5097 				(wl_proxd_event_t*)event_data;
5098 			DHD_LOG_MEM(("MACEVENT: %s, event:%d, status:%d\n",
5099 				event_name, proxd->type, reason));
5100 		}
5101 		break;
5102 	case WLC_E_RPSNOA:
5103 		if (datalen >= sizeof(rpsnoa_stats_t)) {
5104 			const rpsnoa_stats_t *stat = event_data;
5105 			if (datalen == sizeof(*stat)) {
5106 				DHD_EVENT(("MACEVENT: %s, band %s, status %d, pps %d\n", event_name,
5107 					(stat->band == WLC_BAND_2G) ? "2G":"5G",
5108 					stat->state, stat->last_pps));
5109 			}
5110 		}
5111 		break;
5112 	case WLC_E_WA_LQM:
5113 		if (datalen >= sizeof(wl_event_wa_lqm_t)) {
5114 			const wl_event_wa_lqm_t *event_wa_lqm =
5115 				(wl_event_wa_lqm_t *)event_data;
5116 			const bcm_xtlv_t *subevent;
5117 			const wl_event_wa_lqm_basic_t *elqm_basic;
5118 
5119 			if ((event_wa_lqm->ver != WL_EVENT_WA_LQM_VER) ||
5120 			    (event_wa_lqm->len < sizeof(wl_event_wa_lqm_t) + BCM_XTLV_HDR_SIZE)) {
5121 				DHD_ERROR(("MACEVENT: %s invalid (ver=%d len=%d)\n",
5122 					event_name, event_wa_lqm->ver, event_wa_lqm->len));
5123 				break;
5124 			}
5125 
5126 			subevent = (const bcm_xtlv_t *)event_wa_lqm->subevent;
5127 			 if ((subevent->id != WL_EVENT_WA_LQM_BASIC) ||
5128 			     (subevent->len < sizeof(wl_event_wa_lqm_basic_t))) {
5129 				DHD_ERROR(("MACEVENT: %s invalid sub-type (id=%d len=%d)\n",
5130 					event_name, subevent->id, subevent->len));
5131 				break;
5132 			}
5133 
5134 			elqm_basic = (const wl_event_wa_lqm_basic_t *)subevent->data;
5135 			BCM_REFERENCE(elqm_basic);
5136 			DHD_EVENT(("MACEVENT: %s (RSSI=%d SNR=%d TxRate=%d RxRate=%d)\n",
5137 				event_name, elqm_basic->rssi, elqm_basic->snr,
5138 				elqm_basic->tx_rate, elqm_basic->rx_rate));
5139 		}
5140 		break;
5141 
5142 	case WLC_E_OBSS_DETECTION:
5143 		{
5144 			DHD_EVENT(("MACEVENT: %s, type:%d\n", event_name, reason));
5145 			break;
5146 		}
5147 
5148 	case WLC_E_AP_BCN_MUTE:
5149 		if (datalen >= sizeof(wlc_bcn_mute_miti_event_data_v1_t)) {
5150 			const wlc_bcn_mute_miti_event_data_v1_t
5151 				*bcn_mute_miti_evnt_data = event_data;
5152 			DHD_EVENT(("MACEVENT: %s, reason :%d uatbtt_count: %d\n",
5153 				event_name, reason, bcn_mute_miti_evnt_data->uatbtt_count));
5154 		}
5155 		break;
5156 
5157 	case WLC_E_TWT_SETUP:
5158 		DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
5159 		break;
5160 	case WLC_E_TWT_TEARDOWN:
5161 		DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
5162 		break;
5163 	case WLC_E_TWT_INFO_FRM:
5164 		DHD_EVENT(("MACEVENT: %s, MAC %s\n", event_name, eabuf));
5165 		break;
5166 	default:
5167 		DHD_EVENT(("MACEVENT: %s %d, MAC %s, status %d, reason %d, auth %d\n",
5168 		       event_name, event_type, eabuf, (int)status, (int)reason,
5169 		       (int)auth_type));
5170 		break;
5171 	}
5172 
5173 	/* show any appended data if message level is set to bytes or host_data is set */
5174 	if ((DHD_BYTES_ON() || (host_data == TRUE)) && DHD_EVENT_ON() && datalen) {
5175 		buf = (uchar *) event_data;
5176 		BCM_REFERENCE(buf);
5177 		DHD_EVENT((" data (%d) : ", datalen));
5178 		for (i = 0; i < datalen; i++) {
5179 			DHD_EVENT((" 0x%02x ", buf[i]));
5180 		}
5181 		DHD_EVENT(("\n"));
5182 	}
5183 } /* wl_show_host_event */
5184 #endif /* SHOW_EVENTS */
5185 
5186 #ifdef DNGL_EVENT_SUPPORT
5187 /* Check whether packet is a BRCM dngl event pkt. If it is, process event data. */
5188 int
dngl_host_event(dhd_pub_t * dhdp,void * pktdata,bcm_dngl_event_msg_t * dngl_event,size_t pktlen)5189 dngl_host_event(dhd_pub_t *dhdp, void *pktdata, bcm_dngl_event_msg_t *dngl_event, size_t pktlen)
5190 {
5191 	bcm_dngl_event_t *pvt_data = (bcm_dngl_event_t *)pktdata;
5192 
5193 	dngl_host_event_process(dhdp, pvt_data, dngl_event, pktlen);
5194 	return BCME_OK;
5195 }
5196 
5197 #ifdef PARSE_DONGLE_HOST_EVENT
5198 typedef struct hck_id_to_str_s {
5199 	uint32 id;
5200 	char *name;
5201 } hck_id_to_str_t;
5202 
5203 hck_id_to_str_t hck_sw_id_to_str[] = {
5204 	{WL_HC_DD_PCIE, "WL_HC_DD_PCIE"},
5205 	{WL_HC_DD_RX_DMA_STALL, "WL_HC_DD_RX_DMA_STALL"},
5206 	{WL_HC_DD_RX_STALL, "WL_HC_DD_RX_STALL"},
5207 	{WL_HC_DD_TX_STALL, "WL_HC_DD_TX_STALL"},
5208 	{WL_HC_DD_SCAN_STALL, "WL_HC_DD_SCAN_STALL"},
5209 	{WL_HC_DD_PHY, "WL_HC_DD_PHY"},
5210 	{WL_HC_DD_REINIT, "WL_HC_DD_REINIT"},
5211 	{WL_HC_DD_TXQ_STALL, "WL_HC_DD_TXQ_STALL"},
5212 	{0, NULL}
5213 };
5214 
5215 hck_id_to_str_t hck_pcie_module_to_str[] = {
5216 	{HEALTH_CHECK_PCIEDEV_INDUCED_IND, "PCIEDEV_INDUCED_IND"},
5217 	{HEALTH_CHECK_PCIEDEV_H2D_DMA_IND, "PCIEDEV_H2D_DMA_IND"},
5218 	{HEALTH_CHECK_PCIEDEV_D2H_DMA_IND, "PCIEDEV_D2H_DMA_IND"},
5219 	{HEALTH_CHECK_PCIEDEV_IOCTL_STALL_IND, "PCIEDEV_IOCTL_STALL_IND"},
5220 	{HEALTH_CHECK_PCIEDEV_D3ACK_STALL_IND, "PCIEDEV_D3ACK_STALL_IND"},
5221 	{HEALTH_CHECK_PCIEDEV_NODS_IND, "PCIEDEV_NODS_IND"},
5222 	{HEALTH_CHECK_PCIEDEV_LINKSPEED_FALLBACK_IND, "PCIEDEV_LINKSPEED_FALLBACK_IND"},
5223 	{HEALTH_CHECK_PCIEDEV_DSACK_STALL_IND, "PCIEDEV_DSACK_STALL_IND"},
5224 	{0, NULL}
5225 };
5226 
5227 hck_id_to_str_t hck_rx_stall_v2_to_str[] = {
5228 	{BCM_RX_HC_RESERVED, "BCM_RX_HC_RESERVED"},
5229 	{BCM_RX_HC_UNSPECIFIED, "BCM_RX_HC_UNSPECIFIED"},
5230 	{BCM_RX_HC_UNICAST_DECRYPT_FAIL, "BCM_RX_HC_UNICAST_DECRYPT_FAIL"},
5231 	{BCM_RX_HC_BCMC_DECRYPT_FAIL, "BCM_RX_HC_BCMC_DECRYPT_FAIL"},
5232 	{BCM_RX_HC_UNICAST_REPLAY, "BCM_RX_HC_UNICAST_REPLAY"},
5233 	{BCM_RX_HC_BCMC_REPLAY, "BCM_RX_HC_BCMC_REPLAY"},
5234 	{BCM_RX_HC_AMPDU_DUP, "BCM_RX_HC_AMPDU_DUP"},
5235 	{0, NULL}
5236 };
5237 
5238 static void
dhd_print_dongle_hck_id(uint32 id,hck_id_to_str_t * hck)5239 dhd_print_dongle_hck_id(uint32 id, hck_id_to_str_t *hck)
5240 {
5241 	while (hck->name != NULL) {
5242 		if (hck->id == id) {
5243 			DHD_ERROR(("DONGLE_HCK_EVENT: %s\n", hck->name));
5244 			return;
5245 		}
5246 		hck++;
5247 	}
5248 }
5249 
5250 void
dhd_parse_hck_common_sw_event(bcm_xtlv_t * wl_hc)5251 dhd_parse_hck_common_sw_event(bcm_xtlv_t *wl_hc)
5252 {
5253 
5254 	wl_rx_hc_info_v2_t *hck_rx_stall_v2;
5255 	uint16 id;
5256 
5257 	id = ltoh16(wl_hc->id);
5258 
5259 	if (id == WL_HC_DD_RX_STALL_V2) {
5260 		/*  map the hck_rx_stall_v2 structure to the value of the XTLV */
5261 		hck_rx_stall_v2 =
5262 			(wl_rx_hc_info_v2_t*)wl_hc;
5263 		DHD_ERROR(("type:%d len:%d if_idx:%d ac:%d pkts:%d"
5264 			" drop:%d alert_th:%d reason:%d peer_ea:"MACF"\n",
5265 			hck_rx_stall_v2->type,
5266 			hck_rx_stall_v2->length,
5267 			hck_rx_stall_v2->if_idx,
5268 			hck_rx_stall_v2->ac,
5269 			hck_rx_stall_v2->rx_hc_pkts,
5270 			hck_rx_stall_v2->rx_hc_dropped_all,
5271 			hck_rx_stall_v2->rx_hc_alert_th,
5272 			hck_rx_stall_v2->reason,
5273 			ETHER_TO_MACF(hck_rx_stall_v2->peer_ea)));
5274 		dhd_print_dongle_hck_id(
5275 				ltoh32(hck_rx_stall_v2->reason),
5276 				hck_rx_stall_v2_to_str);
5277 	} else {
5278 		dhd_print_dongle_hck_id(ltoh16(wl_hc->id),
5279 				hck_sw_id_to_str);
5280 	}
5281 
5282 }
5283 
5284 #endif /* PARSE_DONGLE_HOST_EVENT */
5285 
5286 void
dngl_host_event_process(dhd_pub_t * dhdp,bcm_dngl_event_t * event,bcm_dngl_event_msg_t * dngl_event,size_t pktlen)5287 dngl_host_event_process(dhd_pub_t *dhdp, bcm_dngl_event_t *event,
5288 	bcm_dngl_event_msg_t *dngl_event, size_t pktlen)
5289 {
5290 	uint8 *p = (uint8 *)(event + 1);
5291 	uint16 type = ntoh16_ua((void *)&dngl_event->event_type);
5292 	uint16 datalen = ntoh16_ua((void *)&dngl_event->datalen);
5293 	uint16 version = ntoh16_ua((void *)&dngl_event->version);
5294 
5295 	DHD_EVENT(("VERSION:%d, EVENT TYPE:%d, DATALEN:%d\n", version, type, datalen));
5296 	if (datalen > (pktlen - sizeof(bcm_dngl_event_t) + ETHER_TYPE_LEN)) {
5297 		return;
5298 	}
5299 	if (version != BCM_DNGL_EVENT_MSG_VERSION) {
5300 		DHD_ERROR(("%s:version mismatch:%d:%d\n", __FUNCTION__,
5301 			version, BCM_DNGL_EVENT_MSG_VERSION));
5302 		return;
5303 	}
5304 	switch (type) {
5305 	   case DNGL_E_SOCRAM_IND:
5306 		{
5307 		   bcm_dngl_socramind_t *socramind_ptr = (bcm_dngl_socramind_t *)p;
5308 		   uint16 tag = ltoh32(socramind_ptr->tag);
5309 		   uint16 taglen = ltoh32(socramind_ptr->length);
5310 		   p = (uint8 *)socramind_ptr->value;
5311 		   DHD_EVENT(("Tag:%d Len:%d Datalen:%d\n", tag, taglen, datalen));
5312 		   switch (tag) {
5313 			case SOCRAM_IND_ASSERT_TAG:
5314 			    {
5315 				/*
5316 				* The payload consists of -
5317 				* null terminated function name padded till 32 bit boundary +
5318 				* Line number - (32 bits)
5319 				* Caller address (32 bits)
5320 				*/
5321 				char *fnname = (char *)p;
5322 				if (datalen < (ROUNDUP(strlen(fnname) + 1, sizeof(uint32)) +
5323 					sizeof(uint32) * 2)) {
5324 					DHD_ERROR(("Wrong length:%d\n", datalen));
5325 					return;
5326 				}
5327 				DHD_EVENT(("ASSRT Function:%s ", p));
5328 				p += ROUNDUP(strlen(p) + 1, sizeof(uint32));
5329 				DHD_EVENT(("Line:%d ", *(uint32 *)p));
5330 				p += sizeof(uint32);
5331 				DHD_EVENT(("Caller Addr:0x%x\n", *(uint32 *)p));
5332 #ifdef PARSE_DONGLE_HOST_EVENT
5333 				DHD_ERROR(("DONGLE_HCK_EVENT: SOCRAM_IND_ASSERT_TAG\n"));
5334 #endif /* PARSE_DONGLE_HOST_EVENT */
5335 				break;
5336 			    }
5337 			case SOCRAM_IND_TAG_HEALTH_CHECK:
5338 			   {
5339 				bcm_dngl_healthcheck_t *dngl_hc = (bcm_dngl_healthcheck_t *)p;
5340 				DHD_EVENT(("SOCRAM_IND_HEALTHCHECK_TAG:%d Len:%d datalen:%d\n",
5341 					ltoh32(dngl_hc->top_module_tag),
5342 					ltoh32(dngl_hc->top_module_len),
5343 					datalen));
5344 				if (DHD_EVENT_ON()) {
5345 					prhex("HEALTHCHECK", p, MIN(ltoh32(dngl_hc->top_module_len)
5346 						+ BCM_XTLV_HDR_SIZE, datalen));
5347 				}
5348 #ifdef DHD_LOG_DUMP
5349 				memset(dhdp->health_chk_event_data, 0, HEALTH_CHK_BUF_SIZE);
5350 				memcpy(dhdp->health_chk_event_data, p,
5351 						MIN(ltoh32(dngl_hc->top_module_len),
5352 						HEALTH_CHK_BUF_SIZE));
5353 #endif /* DHD_LOG_DUMP */
5354 				p = (uint8 *)dngl_hc->value;
5355 
5356 				switch (ltoh32(dngl_hc->top_module_tag)) {
5357 					case HEALTH_CHECK_TOP_LEVEL_MODULE_PCIEDEV_RTE:
5358 					   {
5359 						bcm_dngl_pcie_hc_t *pcie_hc;
5360 						pcie_hc = (bcm_dngl_pcie_hc_t *)p;
5361 						BCM_REFERENCE(pcie_hc);
5362 						if (ltoh32(dngl_hc->top_module_len) <
5363 								sizeof(bcm_dngl_pcie_hc_t)) {
5364 							DHD_ERROR(("Wrong length:%d\n",
5365 								ltoh32(dngl_hc->top_module_len)));
5366 							return;
5367 						}
5368 						DHD_EVENT(("%d:PCIE HC error:%d flag:0x%x,"
5369 							" control:0x%x\n",
5370 							ltoh32(pcie_hc->version),
5371 							ltoh32(pcie_hc->pcie_err_ind_type),
5372 							ltoh32(pcie_hc->pcie_flag),
5373 							ltoh32(pcie_hc->pcie_control_reg)));
5374 #ifdef PARSE_DONGLE_HOST_EVENT
5375 						dhd_print_dongle_hck_id(
5376 							ltoh32(pcie_hc->pcie_err_ind_type),
5377 								hck_pcie_module_to_str);
5378 #endif /* PARSE_DONGLE_HOST_EVENT */
5379 						break;
5380 					   }
5381 #ifdef HCHK_COMMON_SW_EVENT
5382 					case HCHK_SW_ENTITY_WL_PRIMARY:
5383 					case HCHK_SW_ENTITY_WL_SECONDARY:
5384 					{
5385 						bcm_xtlv_t *wl_hc = (bcm_xtlv_t*)p;
5386 
5387 						if (ltoh32(dngl_hc->top_module_len) <
5388 								sizeof(bcm_xtlv_t)) {
5389 							DHD_ERROR(("WL SW HC Wrong length:%d\n",
5390 								ltoh32(dngl_hc->top_module_len)));
5391 							return;
5392 						}
5393 						BCM_REFERENCE(wl_hc);
5394 						DHD_EVENT(("WL SW HC type %d len %d\n",
5395 						ltoh16(wl_hc->id), ltoh16(wl_hc->len)));
5396 
5397 #ifdef PARSE_DONGLE_HOST_EVENT
5398 						dhd_parse_hck_common_sw_event(wl_hc);
5399 #endif /* PARSE_DONGLE_HOST_EVENT */
5400 						break;
5401 
5402 					}
5403 #endif /* HCHK_COMMON_SW_EVENT */
5404 					default:
5405 					{
5406 						DHD_ERROR(("%s:Unknown module TAG:%d\n",
5407 						  __FUNCTION__,
5408 						  ltoh32(dngl_hc->top_module_tag)));
5409 						break;
5410 					}
5411 				}
5412 				break;
5413 			   }
5414 			default:
5415 			   DHD_ERROR(("%s:Unknown TAG\n", __FUNCTION__));
5416 			   if (p && DHD_EVENT_ON()) {
5417 				   prhex("SOCRAMIND", p, taglen);
5418 			   }
5419 			   break;
5420 		   }
5421 		   break;
5422 		}
5423 	   default:
5424 		DHD_ERROR(("%s:Unknown DNGL Event Type:%d\n", __FUNCTION__, type));
5425 		if (p && DHD_EVENT_ON()) {
5426 			prhex("SOCRAMIND", p, datalen);
5427 		}
5428 		break;
5429 	}
5430 #ifndef BCMDBUS
5431 #ifdef DHD_FW_COREDUMP
5432 	if (dhdp->memdump_enabled) {
5433 		dhdp->memdump_type = DUMP_TYPE_DONGLE_HOST_EVENT;
5434 		if (
5435 #ifdef GDB_PROXY
5436 			!dhdp->gdb_proxy_active &&
5437 #endif /* GDB_PROXY */
5438 			dhd_schedule_socram_dump(dhdp)) {
5439 				DHD_ERROR(("%s: socram dump failed\n", __FUNCTION__));
5440 		}
5441 	}
5442 #else
5443 	dhd_dbg_send_urgent_evt(dhdp, p, datalen);
5444 #endif /* DHD_FW_COREDUMP */
5445 #endif /* !BCMDBUS */
5446 }
5447 
5448 #endif /* DNGL_EVENT_SUPPORT */
5449 
5450 /* Stub for now. Will become real function as soon as shim
5451  * is being integrated to Android, Linux etc.
5452  */
5453 #if !defined(NDIS)
5454 int
wl_event_process_default(wl_event_msg_t * event,struct wl_evt_pport * evt_pport)5455 wl_event_process_default(wl_event_msg_t *event, struct wl_evt_pport *evt_pport)
5456 {
5457 	return BCME_OK;
5458 }
5459 #endif
5460 
5461 int
wl_event_process(dhd_pub_t * dhd_pub,int * ifidx,void * pktdata,uint pktlen,void ** data_ptr,void * raw_event)5462 wl_event_process(dhd_pub_t *dhd_pub, int *ifidx, void *pktdata,
5463 	uint pktlen, void **data_ptr, void *raw_event)
5464 {
5465 	wl_evt_pport_t evt_pport;
5466 	wl_event_msg_t event;
5467 	bcm_event_msg_u_t evu;
5468 	int ret;
5469 
5470 	/* make sure it is a BRCM event pkt and record event data */
5471 	ret = wl_host_event_get_data(pktdata, pktlen, &evu);
5472 	if (ret != BCME_OK) {
5473 		return ret;
5474 	}
5475 
5476 	memcpy(&event, &evu.event, sizeof(wl_event_msg_t));
5477 
5478 	/* convert event from network order to host order */
5479 	wl_event_to_host_order(&event);
5480 
5481 	/* record event params to evt_pport */
5482 	evt_pport.dhd_pub = dhd_pub;
5483 	evt_pport.ifidx = ifidx;
5484 	evt_pport.pktdata = pktdata;
5485 	evt_pport.data_ptr = data_ptr;
5486 	evt_pport.raw_event = raw_event;
5487 	evt_pport.data_len = pktlen;
5488 
5489 #if defined(WL_WLC_SHIM) && defined(WL_WLC_SHIM_EVENTS)
5490 	{
5491 		struct wl_shim_node *shim = dhd_pub_shim(dhd_pub);
5492 		if (shim) {
5493 			ret = wl_shim_event_process(shim, &event, &evt_pport);
5494 		} else {
5495 			/* events can come even before shim is initialized
5496 			 (when waiting for "wlc_ver" response)
5497 			 * handle them in a non-shim way.
5498 			 */
5499 			DHD_ERROR(("%s: Events coming before shim initialization!\n",
5500 				__FUNCTION__));
5501 			ret = wl_event_process_default(&event, &evt_pport);
5502 		}
5503 	}
5504 #else
5505 	ret = wl_event_process_default(&event, &evt_pport);
5506 #endif /* WL_WLC_SHIM && WL_WLC_SHIM_EVENTS */
5507 
5508 	return ret;
5509 } /* wl_event_process */
5510 
5511 /* Check whether packet is a BRCM event pkt. If it is, record event data. */
5512 int
wl_host_event_get_data(void * pktdata,uint pktlen,bcm_event_msg_u_t * evu)5513 wl_host_event_get_data(void *pktdata, uint pktlen, bcm_event_msg_u_t *evu)
5514 {
5515 	int ret;
5516 
5517 	ret = is_wlc_event_frame(pktdata, pktlen, 0, evu);
5518 	if (ret != BCME_OK) {
5519 		DHD_ERROR(("%s: Invalid event frame, err = %d\n",
5520 			__FUNCTION__, ret));
5521 	}
5522 
5523 	return ret;
5524 }
5525 
5526 int
wl_process_host_event(dhd_pub_t * dhd_pub,int * ifidx,void * pktdata,uint pktlen,wl_event_msg_t * event,void ** data_ptr,void * raw_event)5527 wl_process_host_event(dhd_pub_t *dhd_pub, int *ifidx, void *pktdata, uint pktlen,
5528 	wl_event_msg_t *event, void **data_ptr, void *raw_event)
5529 {
5530 	bcm_event_t *pvt_data = (bcm_event_t *)pktdata;
5531 	bcm_event_msg_u_t evu;
5532 	uint8 *event_data;
5533 	uint32 type, status, datalen, reason;
5534 	uint16 flags;
5535 	uint evlen;
5536 	int ret;
5537 	uint16 usr_subtype;
5538 #if defined(__linux__)
5539 	dhd_if_t *ifp = NULL;
5540 	BCM_REFERENCE(ifp);
5541 #endif /* DHD_POST_EAPOL_M1_AFTER_ROAM_EVT */
5542 
5543 	ret = wl_host_event_get_data(pktdata, pktlen, &evu);
5544 	if (ret != BCME_OK) {
5545 		return ret;
5546 	}
5547 
5548 	usr_subtype = ntoh16_ua((void *)&pvt_data->bcm_hdr.usr_subtype);
5549 	switch (usr_subtype) {
5550 	case BCMILCP_BCM_SUBTYPE_EVENT:
5551 		memcpy(event, &evu.event, sizeof(wl_event_msg_t));
5552 		*data_ptr = &pvt_data[1];
5553 		break;
5554 	case BCMILCP_BCM_SUBTYPE_DNGLEVENT:
5555 #ifdef DNGL_EVENT_SUPPORT
5556 		/* If it is a DNGL event process it first */
5557 		if (dngl_host_event(dhd_pub, pktdata, &evu.dngl_event, pktlen) == BCME_OK) {
5558 			/*
5559 			 * Return error purposely to prevent DNGL event being processed
5560 			 * as BRCM event
5561 			 */
5562 			return BCME_ERROR;
5563 		}
5564 #endif /* DNGL_EVENT_SUPPORT */
5565 		return BCME_NOTFOUND;
5566 	default:
5567 		return BCME_NOTFOUND;
5568 	}
5569 
5570 	/* start wl_event_msg process */
5571 	event_data = *data_ptr;
5572 	type = ntoh32_ua((void *)&event->event_type);
5573 	flags = ntoh16_ua((void *)&event->flags);
5574 	status = ntoh32_ua((void *)&event->status);
5575 	reason = ntoh32_ua((void *)&event->reason);
5576 	datalen = ntoh32_ua((void *)&event->datalen);
5577 	evlen = datalen + sizeof(bcm_event_t);
5578 
5579 	switch (type) {
5580 #ifdef PROP_TXSTATUS
5581 	case WLC_E_FIFO_CREDIT_MAP:
5582 		dhd_wlfc_enable(dhd_pub);
5583 		dhd_wlfc_FIFOcreditmap_event(dhd_pub, event_data);
5584 		WLFC_DBGMESG(("WLC_E_FIFO_CREDIT_MAP:(AC0,AC1,AC2,AC3),(BC_MC),(OTHER): "
5585 			"(%d,%d,%d,%d),(%d),(%d)\n", event_data[0], event_data[1],
5586 			event_data[2],
5587 			event_data[3], event_data[4], event_data[5]));
5588 		break;
5589 
5590 	case WLC_E_BCMC_CREDIT_SUPPORT:
5591 		dhd_wlfc_BCMCCredit_support_event(dhd_pub);
5592 		break;
5593 #ifdef LIMIT_BORROW
5594 	case WLC_E_ALLOW_CREDIT_BORROW:
5595 		dhd_wlfc_disable_credit_borrow_event(dhd_pub, event_data);
5596 		break;
5597 #endif /* LIMIT_BORROW */
5598 #endif /* PROP_TXSTATUS */
5599 
5600 	case WLC_E_ULP:
5601 		break;
5602 	case WLC_E_TDLS_PEER_EVENT:
5603 #if defined(WLTDLS) && defined(PCIE_FULL_DONGLE)
5604 		{
5605 			dhd_tdls_event_handler(dhd_pub, event);
5606 		}
5607 #endif
5608 		break;
5609 
5610 	case WLC_E_IF:
5611 		{
5612 		struct wl_event_data_if *ifevent = (struct wl_event_data_if *)event_data;
5613 
5614 		/* Ignore the event if NOIF is set */
5615 		if (ifevent->reserved & WLC_E_IF_FLAGS_BSSCFG_NOIF) {
5616 			DHD_ERROR(("WLC_E_IF: NO_IF set, event Ignored\r\n"));
5617 			return (BCME_UNSUPPORTED);
5618 		}
5619 #ifdef PCIE_FULL_DONGLE
5620 		dhd_update_interface_flow_info(dhd_pub, ifevent->ifidx,
5621 			ifevent->opcode, ifevent->role);
5622 #endif
5623 #ifdef PROP_TXSTATUS
5624 		{
5625 			uint8* ea = pvt_data->eth.ether_dhost;
5626 			WLFC_DBGMESG(("WLC_E_IF: idx:%d, action:%s, iftype:%s, ["MACDBG"]\n"
5627 						  ifevent->ifidx,
5628 						  ((ifevent->opcode == WLC_E_IF_ADD) ? "ADD":"DEL"),
5629 						  ((ifevent->role == 0) ? "STA":"AP "),
5630 						  MAC2STRDBG(ea)));
5631 			(void)ea;
5632 
5633 			if (ifevent->opcode == WLC_E_IF_CHANGE)
5634 				dhd_wlfc_interface_event(dhd_pub,
5635 					eWLFC_MAC_ENTRY_ACTION_UPDATE,
5636 					ifevent->ifidx, ifevent->role, ea);
5637 			else
5638 				dhd_wlfc_interface_event(dhd_pub,
5639 					((ifevent->opcode == WLC_E_IF_ADD) ?
5640 					eWLFC_MAC_ENTRY_ACTION_ADD : eWLFC_MAC_ENTRY_ACTION_DEL),
5641 					ifevent->ifidx, ifevent->role, ea);
5642 
5643 			/* dhd already has created an interface by default, for 0 */
5644 			if (ifevent->ifidx == 0)
5645 				break;
5646 		}
5647 #endif /* PROP_TXSTATUS */
5648 
5649 		if (ifevent->ifidx > 0 && ifevent->ifidx < DHD_MAX_IFS) {
5650 			if (ifevent->opcode == WLC_E_IF_ADD) {
5651 				if (dhd_event_ifadd(dhd_pub->info, ifevent, event->ifname,
5652 					event->addr.octet)) {
5653 
5654 					DHD_ERROR(("%s: dhd_event_ifadd failed ifidx: %d  %s\n",
5655 						__FUNCTION__, ifevent->ifidx, event->ifname));
5656 					return (BCME_ERROR);
5657 				}
5658 			} else if (ifevent->opcode == WLC_E_IF_DEL) {
5659 #ifdef PCIE_FULL_DONGLE
5660 				dhd_flow_rings_delete(dhd_pub,
5661 					(uint8)dhd_ifname2idx(dhd_pub->info, event->ifname));
5662 #endif /* PCIE_FULL_DONGLE */
5663 				dhd_event_ifdel(dhd_pub->info, ifevent, event->ifname,
5664 					event->addr.octet);
5665 			} else if (ifevent->opcode == WLC_E_IF_CHANGE) {
5666 #ifdef WL_CFG80211
5667 				dhd_event_ifchange(dhd_pub->info, ifevent, event->ifname,
5668 					event->addr.octet);
5669 #endif /* WL_CFG80211 */
5670 			}
5671 		} else {
5672 #if !defined(PROP_TXSTATUS) && !defined(PCIE_FULL_DONGLE) && defined(WL_CFG80211)
5673 			DHD_INFO(("%s: Invalid ifidx %d for %s\n",
5674 			   __FUNCTION__, ifevent->ifidx, event->ifname));
5675 #endif /* !PROP_TXSTATUS && !PCIE_FULL_DONGLE && WL_CFG80211 */
5676 		}
5677 			/* send up the if event: btamp user needs it */
5678 			*ifidx = dhd_ifname2idx(dhd_pub->info, event->ifname);
5679 			/* push up to external supp/auth */
5680 			dhd_event(dhd_pub->info, (char *)pvt_data, evlen, *ifidx);
5681 		break;
5682 	}
5683 
5684 	case WLC_E_NDIS_LINK:
5685 		break;
5686 	case WLC_E_PFN_NET_FOUND:
5687 	case WLC_E_PFN_SCAN_ALLGONE: /* share with WLC_E_PFN_BSSID_NET_LOST */
5688 	case WLC_E_PFN_NET_LOST:
5689 		break;
5690 #if defined(OEM_ANDROID) && defined(PNO_SUPPORT)
5691 	case WLC_E_PFN_BSSID_NET_FOUND:
5692 	case WLC_E_PFN_BEST_BATCHING:
5693 		dhd_pno_event_handler(dhd_pub, event, (void *)event_data);
5694 		break;
5695 #endif /* #if defined(OEM_ANDROID) && defined(PNO_SUPPORT) */
5696 #if defined(RTT_SUPPORT)
5697 	case WLC_E_PROXD:
5698 #ifndef WL_CFG80211
5699 		dhd_rtt_event_handler(dhd_pub, event, (void *)event_data);
5700 #endif /* WL_CFG80211 */
5701 		break;
5702 #endif /* RTT_SUPPORT */
5703 		/* These are what external supplicant/authenticator wants */
5704 	case WLC_E_ASSOC_IND:
5705 	case WLC_E_AUTH_IND:
5706 	case WLC_E_REASSOC_IND:
5707 		dhd_findadd_sta(dhd_pub,
5708 			dhd_ifname2idx(dhd_pub->info, event->ifname),
5709 			&event->addr.octet);
5710 		break;
5711 #if !defined(BCMDBUS) && defined(DHD_FW_COREDUMP)
5712 	case WLC_E_PSM_WATCHDOG:
5713 		DHD_ERROR(("%s: WLC_E_PSM_WATCHDOG event received : \n", __FUNCTION__));
5714 		if (dhd_socram_dump(dhd_pub->bus) != BCME_OK) {
5715 			DHD_ERROR(("%s: socram dump ERROR : \n", __FUNCTION__));
5716 		}
5717 	break;
5718 #endif
5719 #ifdef DHD_WMF
5720 	case WLC_E_PSTA_PRIMARY_INTF_IND:
5721 		dhd_update_psta_interface_for_sta(dhd_pub, event->ifname,
5722 			(void *)(event->addr.octet), (void*) event_data);
5723 		break;
5724 #endif
5725 #ifdef BCM_ROUTER_DHD
5726 	case WLC_E_DPSTA_INTF_IND:
5727 		dhd_update_dpsta_interface_for_sta(dhd_pub, (uint8)dhd_ifname2idx(dhd_pub->info,
5728 			event->ifname), (void*) event_data);
5729 		break;
5730 #endif /* BCM_ROUTER_DHD */
5731 #ifdef BCMDBG
5732 	case WLC_E_MACDBG:
5733 		dhd_macdbg_event_handler(dhd_pub, reason, event_data, datalen);
5734 		break;
5735 #endif /* BCMDBG */
5736 	case WLC_E_NATOE_NFCT:
5737 #ifdef WL_NATOE
5738 		DHD_EVENT(("%s: WLC_E_NATOE_NFCT event received \n", __FUNCTION__));
5739 		dhd_natoe_ct_event(dhd_pub, event_data);
5740 #endif /* WL_NATOE */
5741 	break;
5742 	case WLC_E_SLOTTED_BSS_PEER_OP:
5743 		DHD_EVENT(("%s: WLC_E_SLOTTED_BSS_PEER_OP event received for peer: "
5744 			"" MACDBG ", status = %d\n",
5745 			__FUNCTION__, MAC2STRDBG(event->addr.octet), status));
5746 		if (status == WLC_E_STATUS_SLOTTED_PEER_ADD) {
5747 			dhd_findadd_sta(dhd_pub, dhd_ifname2idx(dhd_pub->info,
5748 				event->ifname), &event->addr.octet);
5749 		} else if (status == WLC_E_STATUS_SLOTTED_PEER_DEL) {
5750 			uint8 ifindex = (uint8)dhd_ifname2idx(dhd_pub->info, event->ifname);
5751 			BCM_REFERENCE(ifindex);
5752 			dhd_del_sta(dhd_pub, dhd_ifname2idx(dhd_pub->info,
5753 				event->ifname), &event->addr.octet);
5754 #ifdef PCIE_FULL_DONGLE
5755 			dhd_flow_rings_delete_for_peer(dhd_pub, ifindex,
5756 				(char *)&event->addr.octet[0]);
5757 #endif
5758 		} else {
5759 			DHD_ERROR(("%s: WLC_E_SLOTTED_BSS_PEER_OP: Status is not expected = %d\n",
5760 				__FUNCTION__, status));
5761 		}
5762 		break;
5763 #ifdef DHD_POST_EAPOL_M1_AFTER_ROAM_EVT
5764 	case WLC_E_REASSOC:
5765 		ifp = dhd_get_ifp(dhd_pub, event->ifidx);
5766 
5767 		if (!ifp)
5768 			break;
5769 
5770 		/* Consider STA role only since roam is disabled on P2P GC.
5771 		 * Drop EAPOL M1 frame only if roam is done to same BSS.
5772 		 */
5773 		if ((status == WLC_E_STATUS_SUCCESS) &&
5774 			IS_STA_IFACE(ndev_to_wdev(ifp->net)) &&
5775 			wl_cfg80211_is_event_from_connected_bssid(ifp->net, event, event->ifidx)) {
5776 			ifp->recv_reassoc_evt = TRUE;
5777 		}
5778 		break;
5779 #endif /* DHD_POST_EAPOL_M1_AFTER_ROAM_EVT */
5780 #if defined(CSI_SUPPORT)
5781 	case WLC_E_CSI:
5782 		dhd_csi_event_handler(dhd_pub, event, (void *)event_data);
5783 		break;
5784 #endif /* CSI_SUPPORT */
5785 	case WLC_E_LINK:
5786 #ifdef PCIE_FULL_DONGLE
5787 		if (dhd_update_interface_link_status(dhd_pub, (uint8)dhd_ifname2idx(dhd_pub->info,
5788 			event->ifname), (uint8)flags) != BCME_OK) {
5789 			DHD_ERROR(("%s: dhd_update_interface_link_status Failed.\n",
5790 				__FUNCTION__));
5791 			break;
5792 		}
5793 		if (!flags) {
5794 			DHD_ERROR(("%s: Deleting all STA from assoc list and flowrings.\n",
5795 				__FUNCTION__));
5796 			/* Delete all sta and flowrings */
5797 			dhd_del_all_sta(dhd_pub, dhd_ifname2idx(dhd_pub->info, event->ifname));
5798 			dhd_flow_rings_delete(dhd_pub, (uint8)dhd_ifname2idx(dhd_pub->info,
5799 				event->ifname));
5800 		}
5801 		/* fall through */
5802 #endif /* PCIE_FULL_DONGLE */
5803 	case WLC_E_DEAUTH:
5804 	case WLC_E_DEAUTH_IND:
5805 	case WLC_E_DISASSOC:
5806 	case WLC_E_DISASSOC_IND:
5807 #ifdef PCIE_FULL_DONGLE
5808 		if (type != WLC_E_LINK) {
5809 			uint8 ifindex = (uint8)dhd_ifname2idx(dhd_pub->info, event->ifname);
5810 			uint8 role = dhd_flow_rings_ifindex2role(dhd_pub, ifindex);
5811 			uint8 del_sta = TRUE;
5812 #ifdef WL_CFG80211
5813 			if (role == WLC_E_IF_ROLE_STA &&
5814 				!wl_cfg80211_is_roam_offload(dhd_idx2net(dhd_pub, ifindex)) &&
5815 					!wl_cfg80211_is_event_from_connected_bssid(
5816 						dhd_idx2net(dhd_pub, ifindex), event, *ifidx)) {
5817 				del_sta = FALSE;
5818 			}
5819 #endif /* WL_CFG80211 */
5820 			DHD_EVENT(("%s: Link event %d, flags %x, status %x, role %d, del_sta %d\n",
5821 				__FUNCTION__, type, flags, status, role, del_sta));
5822 
5823 			if (del_sta) {
5824 				DHD_EVENT(("%s: Deleting STA " MACDBG "\n",
5825 					__FUNCTION__, MAC2STRDBG(event->addr.octet)));
5826 
5827 				dhd_del_sta(dhd_pub, dhd_ifname2idx(dhd_pub->info,
5828 					event->ifname), &event->addr.octet);
5829 				/* Delete all flowrings for STA and P2P Client */
5830 				if (role == WLC_E_IF_ROLE_STA || role == WLC_E_IF_ROLE_P2P_CLIENT) {
5831 					dhd_flow_rings_delete(dhd_pub, ifindex);
5832 				} else {
5833 					dhd_flow_rings_delete_for_peer(dhd_pub, ifindex,
5834 						(char *)&event->addr.octet[0]);
5835 				}
5836 			}
5837 		}
5838 #endif /* PCIE_FULL_DONGLE */
5839 #ifdef DHD_POST_EAPOL_M1_AFTER_ROAM_EVT
5840 		/* fall through */
5841 		ifp = dhd_get_ifp(dhd_pub, event->ifidx);
5842 		if (ifp) {
5843 			ifp->recv_reassoc_evt = FALSE;
5844 			ifp->post_roam_evt = FALSE;
5845 		}
5846 #endif /* DHD_POST_EAPOL_M1_AFTER_ROAM_EVT */
5847 		/* fall through */
5848 	default:
5849 		*ifidx = dhd_ifname2idx(dhd_pub->info, event->ifname);
5850 #ifdef DHD_UPDATE_INTF_MAC
5851 		if ((WLC_E_LINK==type)&&(WLC_EVENT_MSG_LINK&flags)) {
5852 			dhd_event_ifchange(dhd_pub->info,
5853 			(struct wl_event_data_if *)event,
5854 			event->ifname,
5855 			event->addr.octet);
5856 		}
5857 #endif /* DHD_UPDATE_INTF_MAC */
5858 		/* push up to external supp/auth */
5859 		dhd_event(dhd_pub->info, (char *)pvt_data, evlen, *ifidx);
5860 		DHD_TRACE(("%s: MAC event %d, flags %x, status %x\n",
5861 			__FUNCTION__, type, flags, status));
5862 		BCM_REFERENCE(flags);
5863 		BCM_REFERENCE(status);
5864 		BCM_REFERENCE(reason);
5865 
5866 		break;
5867 	}
5868 #if defined(BCM_ROUTER_DHD) || defined(STBAP)
5869 	/* For routers, EAPD will be working on these events.
5870 	 * Overwrite interface name to that event is pushed
5871 	 * to host with its registered interface name
5872 	 */
5873 	memcpy(pvt_data->event.ifname, dhd_ifname(dhd_pub, *ifidx), IFNAMSIZ);
5874 #endif
5875 
5876 #ifdef DHD_STATUS_LOGGING
5877 	if (dhd_pub->statlog) {
5878 		dhd_statlog_process_event(dhd_pub, type, *ifidx,
5879 			status, reason, flags);
5880 	}
5881 #endif /* DHD_STATUS_LOGGING */
5882 
5883 #ifdef SHOW_EVENTS
5884 	if (DHD_FWLOG_ON() || DHD_EVENT_ON()) {
5885 		wl_show_host_event(dhd_pub, event,
5886 			(void *)event_data, raw_event, dhd_pub->enable_log);
5887 	}
5888 #endif /* SHOW_EVENTS */
5889 
5890 	return (BCME_OK);
5891 } /* wl_process_host_event */
5892 
5893 int
wl_host_event(dhd_pub_t * dhd_pub,int * ifidx,void * pktdata,uint pktlen,wl_event_msg_t * event,void ** data_ptr,void * raw_event)5894 wl_host_event(dhd_pub_t *dhd_pub, int *ifidx, void *pktdata, uint pktlen,
5895 	wl_event_msg_t *event, void **data_ptr, void *raw_event)
5896 {
5897 	return wl_process_host_event(dhd_pub, ifidx, pktdata, pktlen, event, data_ptr,
5898 			raw_event);
5899 }
5900 
5901 void
dhd_print_buf(void * pbuf,int len,int bytes_per_line)5902 dhd_print_buf(void *pbuf, int len, int bytes_per_line)
5903 {
5904 #ifdef DHD_DEBUG
5905 	int i, j = 0;
5906 	unsigned char *buf = pbuf;
5907 
5908 	if (bytes_per_line == 0) {
5909 		bytes_per_line = len;
5910 	}
5911 
5912 	for (i = 0; i < len; i++) {
5913 		printf("%2.2x", *buf++);
5914 		j++;
5915 		if (j == bytes_per_line) {
5916 			printf("\n");
5917 			j = 0;
5918 		} else {
5919 			printf(":");
5920 		}
5921 	}
5922 	printf("\n");
5923 #endif /* DHD_DEBUG */
5924 }
5925 #ifndef strtoul
5926 #define strtoul(nptr, endptr, base) bcm_strtoul((nptr), (endptr), (base))
5927 #endif
5928 
5929 /* Convert user's input in hex pattern to byte-size mask */
5930 int
wl_pattern_atoh(char * src,char * dst)5931 wl_pattern_atoh(char *src, char *dst)
5932 {
5933 	int i;
5934 	if (strncmp(src, "0x", 2) != 0 &&
5935 	    strncmp(src, "0X", 2) != 0) {
5936 		DHD_ERROR(("Mask invalid format. Needs to start with 0x\n"));
5937 		return -1;
5938 	}
5939 	src = src + 2; /* Skip past 0x */
5940 	if (strlen(src) % 2 != 0) {
5941 		DHD_ERROR(("Mask invalid format. Needs to be of even length\n"));
5942 		return -1;
5943 	}
5944 	for (i = 0; *src != '\0'; i++) {
5945 		char num[3];
5946 		bcm_strncpy_s(num, sizeof(num), src, 2);
5947 		num[2] = '\0';
5948 		dst[i] = (uint8)strtoul(num, NULL, 16);
5949 		src += 2;
5950 	}
5951 	return i;
5952 }
5953 
5954 #if defined(PKT_FILTER_SUPPORT) || defined(DHD_PKT_LOGGING)
5955 int
pattern_atoh_len(char * src,char * dst,int len)5956 pattern_atoh_len(char *src, char *dst, int len)
5957 {
5958 	int i;
5959 	if (strncmp(src, "0x", HD_PREFIX_SIZE) != 0 &&
5960 			strncmp(src, "0X", HD_PREFIX_SIZE) != 0) {
5961 		DHD_ERROR(("Mask invalid format. Needs to start with 0x\n"));
5962 		return -1;
5963 	}
5964 	src = src + HD_PREFIX_SIZE; /* Skip past 0x */
5965 	if (strlen(src) % HD_BYTE_SIZE != 0) {
5966 		DHD_ERROR(("Mask invalid format. Needs to be of even length\n"));
5967 		return -1;
5968 	}
5969 	for (i = 0; *src != '\0'; i++) {
5970 		char num[HD_BYTE_SIZE + 1];
5971 
5972 		if (i > len - 1) {
5973 			DHD_ERROR(("pattern not in range, idx: %d len: %d\n", i, len));
5974 			return -1;
5975 		}
5976 		bcm_strncpy_s(num, sizeof(num), src, HD_BYTE_SIZE);
5977 		num[HD_BYTE_SIZE] = '\0';
5978 		dst[i] = (uint8)strtoul(num, NULL, 16);
5979 		src += HD_BYTE_SIZE;
5980 	}
5981 	return i;
5982 }
5983 #endif /* PKT_FILTER_SUPPORT || DHD_PKT_LOGGING */
5984 
5985 #ifdef PKT_FILTER_SUPPORT
5986 void
dhd_pktfilter_offload_enable(dhd_pub_t * dhd,char * arg,int enable,int master_mode)5987 dhd_pktfilter_offload_enable(dhd_pub_t * dhd, char *arg, int enable, int master_mode)
5988 {
5989 	char				*argv[8];
5990 	int					i = 0;
5991 	const char			*str;
5992 	int					buf_len;
5993 	int					str_len;
5994 	char				*arg_save = 0, *arg_org = 0;
5995 	int					rc;
5996 	char				buf[32] = {0};
5997 	wl_pkt_filter_enable_t	enable_parm;
5998 	wl_pkt_filter_enable_t	* pkt_filterp;
5999 
6000 	if (!arg)
6001 		return;
6002 
6003 	if (!(arg_save = MALLOC(dhd->osh, strlen(arg) + 1))) {
6004 		DHD_ERROR(("%s: malloc failed\n", __FUNCTION__));
6005 		goto fail;
6006 	}
6007 	arg_org = arg_save;
6008 	memcpy(arg_save, arg, strlen(arg) + 1);
6009 
6010 	argv[i] = bcmstrtok(&arg_save, " ", 0);
6011 
6012 	i = 0;
6013 	if (argv[i] == NULL) {
6014 		DHD_ERROR(("No args provided\n"));
6015 		goto fail;
6016 	}
6017 
6018 	str = "pkt_filter_enable";
6019 	str_len = strlen(str);
6020 	bcm_strncpy_s(buf, sizeof(buf) - 1, str, sizeof(buf) - 1);
6021 	buf[ sizeof(buf) - 1 ] = '\0';
6022 	buf_len = str_len + 1;
6023 
6024 	pkt_filterp = (wl_pkt_filter_enable_t *)(buf + str_len + 1);
6025 
6026 	/* Parse packet filter id. */
6027 	enable_parm.id = htod32(strtoul(argv[i], NULL, 0));
6028 	if (dhd_conf_del_pkt_filter(dhd, enable_parm.id))
6029 		goto fail;
6030 
6031 	/* Parse enable/disable value. */
6032 	enable_parm.enable = htod32(enable);
6033 
6034 	buf_len += sizeof(enable_parm);
6035 	memcpy((char *)pkt_filterp,
6036 	       &enable_parm,
6037 	       sizeof(enable_parm));
6038 
6039 	/* Enable/disable the specified filter. */
6040 	rc = dhd_wl_ioctl_cmd(dhd, WLC_SET_VAR, buf, buf_len, TRUE, 0);
6041 	rc = rc >= 0 ? 0 : rc;
6042 	if (rc) {
6043 		DHD_ERROR(("%s: failed to %s pktfilter %s, retcode = %d\n",
6044 		__FUNCTION__, enable?"enable":"disable", arg, rc));
6045 		dhd_set_packet_filter(dhd);
6046 		rc = dhd_wl_ioctl_cmd(dhd, WLC_SET_VAR, buf, buf_len, TRUE, 0);
6047 		rc = rc >= 0 ? 0 : rc;
6048 		if (rc) {
6049 			DHD_TRACE_HW4(("%s: 2nd retry failed to add pktfilter %s, retcode = %d\n",
6050 			__FUNCTION__, arg, rc));
6051 		} else {
6052 			DHD_TRACE_HW4(("%s: 2nd retry successfully added pktfilter %s\n",
6053 			__FUNCTION__, arg));
6054 		}
6055 	}
6056 	else
6057 		DHD_TRACE(("%s: successfully %s pktfilter %s\n",
6058 		__FUNCTION__, enable?"enable":"disable", arg));
6059 
6060 	/* Contorl the master mode */
6061 	rc = dhd_wl_ioctl_set_intiovar(dhd, "pkt_filter_mode",
6062 		master_mode, WLC_SET_VAR, TRUE, 0);
6063 	rc = rc >= 0 ? 0 : rc;
6064 	if (rc)
6065 		DHD_TRACE(("%s: failed to set pkt_filter_mode %d, retcode = %d\n",
6066 			__FUNCTION__, master_mode, rc));
6067 
6068 fail:
6069 	if (arg_org)
6070 		MFREE(dhd->osh, arg_org, strlen(arg) + 1);
6071 }
6072 
6073 /* Packet filter section: extended filters have named offsets, add table here */
6074 typedef struct {
6075 	char *name;
6076 	uint16 base;
6077 } wl_pfbase_t;
6078 
6079 static wl_pfbase_t basenames[] = { WL_PKT_FILTER_BASE_NAMES };
6080 
6081 static int
wl_pkt_filter_base_parse(char * name)6082 wl_pkt_filter_base_parse(char *name)
6083 {
6084 	uint i;
6085 	char *bname, *uname;
6086 
6087 	for (i = 0; i < ARRAYSIZE(basenames); i++) {
6088 		bname = basenames[i].name;
6089 		for (uname = name; *uname; bname++, uname++) {
6090 			if (*bname != bcm_toupper(*uname)) {
6091 				break;
6092 			}
6093 		}
6094 		if (!*uname && !*bname) {
6095 			break;
6096 		}
6097 	}
6098 
6099 	if (i < ARRAYSIZE(basenames)) {
6100 		return basenames[i].base;
6101 	} else {
6102 		return -1;
6103 	}
6104 }
6105 
6106 void
dhd_pktfilter_offload_set(dhd_pub_t * dhd,char * arg)6107 dhd_pktfilter_offload_set(dhd_pub_t * dhd, char *arg)
6108 {
6109 	const char			*str;
6110 	wl_pkt_filter_t			pkt_filter;
6111 	wl_pkt_filter_t			*pkt_filterp;
6112 	int				buf_len;
6113 	int				str_len;
6114 	int				rc = -1;
6115 	uint32				mask_size;
6116 	uint32				pattern_size;
6117 	char				*argv[MAXPKT_ARG] = {0}, * buf = 0;
6118 	int				i = 0;
6119 	char				*arg_save = 0, *arg_org = 0;
6120 
6121 	if (!arg)
6122 		return;
6123 
6124 	if (!(arg_save = MALLOC(dhd->osh, strlen(arg) + 1))) {
6125 		DHD_ERROR(("%s: malloc failed\n", __FUNCTION__));
6126 		goto fail;
6127 	}
6128 
6129 	arg_org = arg_save;
6130 
6131 	if (!(buf = MALLOC(dhd->osh, MAX_PKTFLT_BUF_SIZE))) {
6132 		DHD_ERROR(("%s: malloc failed\n", __FUNCTION__));
6133 		goto fail;
6134 	}
6135 
6136 	memset(buf, 0, MAX_PKTFLT_BUF_SIZE);
6137 	memcpy(arg_save, arg, strlen(arg) + 1);
6138 
6139 	if (strlen(arg) > MAX_PKTFLT_BUF_SIZE) {
6140 		DHD_ERROR(("Not enough buffer %d < %d\n", (int)strlen(arg), (int)sizeof(buf)));
6141 		goto fail;
6142 	}
6143 
6144 	argv[i] = bcmstrtok(&arg_save, " ", 0);
6145 	while (argv[i++]) {
6146 		if (i >= MAXPKT_ARG) {
6147 			DHD_ERROR(("Invalid args provided\n"));
6148 			goto fail;
6149 		}
6150 		argv[i] = bcmstrtok(&arg_save, " ", 0);
6151 	}
6152 
6153 	i = 0;
6154 	if (argv[i] == NULL) {
6155 		DHD_ERROR(("No args provided\n"));
6156 		goto fail;
6157 	}
6158 
6159 	str = "pkt_filter_add";
6160 	str_len = strlen(str);
6161 	bcm_strncpy_s(buf, MAX_PKTFLT_BUF_SIZE, str, str_len);
6162 	buf[ str_len ] = '\0';
6163 	buf_len = str_len + 1;
6164 
6165 	pkt_filterp = (wl_pkt_filter_t *) (buf + str_len + 1);
6166 
6167 	/* Parse packet filter id. */
6168 	pkt_filter.id = htod32(strtoul(argv[i], NULL, 0));
6169 
6170 	if (argv[++i] == NULL) {
6171 		DHD_ERROR(("Polarity not provided\n"));
6172 		goto fail;
6173 	}
6174 
6175 	/* Parse filter polarity. */
6176 	pkt_filter.negate_match = htod32(strtoul(argv[i], NULL, 0));
6177 
6178 	if (argv[++i] == NULL) {
6179 		DHD_ERROR(("Filter type not provided\n"));
6180 		goto fail;
6181 	}
6182 
6183 	/* Parse filter type. */
6184 	pkt_filter.type = htod32(strtoul(argv[i], NULL, 0));
6185 
6186 	if ((pkt_filter.type == 0) || (pkt_filter.type == 1)) {
6187 		if (argv[++i] == NULL) {
6188 			DHD_ERROR(("Offset not provided\n"));
6189 			goto fail;
6190 		}
6191 
6192 		/* Parse pattern filter offset. */
6193 		pkt_filter.u.pattern.offset = htod32(strtoul(argv[i], NULL, 0));
6194 
6195 		if (argv[++i] == NULL) {
6196 			DHD_ERROR(("Bitmask not provided\n"));
6197 			goto fail;
6198 		}
6199 
6200 		/* Parse pattern filter mask. */
6201 		rc  = wl_pattern_atoh(argv[i],
6202 			(char *) pkt_filterp->u.pattern.mask_and_pattern);
6203 
6204 		if (rc == -1) {
6205 			DHD_ERROR(("Rejecting: %s\n", argv[i]));
6206 			goto fail;
6207 		}
6208 		mask_size = htod32(rc);
6209 		if (argv[++i] == NULL) {
6210 			DHD_ERROR(("Pattern not provided\n"));
6211 			goto fail;
6212 		}
6213 
6214 		/* Parse pattern filter pattern. */
6215 		rc = wl_pattern_atoh(argv[i],
6216 			(char *) &pkt_filterp->u.pattern.mask_and_pattern[rc]);
6217 
6218 		if (rc == -1) {
6219 			DHD_ERROR(("Rejecting: %s\n", argv[i]));
6220 			goto fail;
6221 		}
6222 		pattern_size = htod32(rc);
6223 		if (mask_size != pattern_size) {
6224 			DHD_ERROR(("Mask and pattern not the same size\n"));
6225 			goto fail;
6226 		}
6227 
6228 		pkt_filter.u.pattern.size_bytes = mask_size;
6229 		buf_len += WL_PKT_FILTER_FIXED_LEN;
6230 		buf_len += (WL_PKT_FILTER_PATTERN_FIXED_LEN + 2 * rc);
6231 
6232 		/* Keep-alive attributes are set in local	variable (keep_alive_pkt), and
6233 		 * then memcpy'ed into buffer (keep_alive_pktp) since there is no
6234 		 * guarantee that the buffer is properly aligned.
6235 		 */
6236 		memcpy((char *)pkt_filterp,
6237 			&pkt_filter,
6238 			WL_PKT_FILTER_FIXED_LEN + WL_PKT_FILTER_PATTERN_FIXED_LEN);
6239 	} else if ((pkt_filter.type == 2) || (pkt_filter.type == 6)) {
6240 		int list_cnt = 0;
6241 		char *endptr = NULL;
6242 		wl_pkt_filter_pattern_listel_t *pf_el =
6243 			(wl_pkt_filter_pattern_listel_t *)&pkt_filterp->u.patlist.patterns[0];
6244 
6245 		while (argv[++i] != NULL) {
6246 			/* Check valid buffer size. */
6247 			if ((buf_len + MAX_PKTFLT_FIXED_BUF_SIZE) > MAX_PKTFLT_BUF_SIZE) {
6248 				DHD_ERROR(("buffer over length MAX_PKTFLT_FIXED_BUF_SIZE\n"));
6249 				goto fail;
6250 			}
6251 
6252 			/* Parse pattern filter base and offset. */
6253 			if (bcm_isdigit(*argv[i])) {
6254 				/* Numeric base */
6255 				rc = strtoul(argv[i], &endptr, 0);
6256 			} else {
6257 				endptr = strchr(argv[i], ':');
6258 				if (endptr) {
6259 					*endptr = '\0';
6260 					rc = wl_pkt_filter_base_parse(argv[i]);
6261 					if (rc == -1) {
6262 						 printf("Invalid base %s\n", argv[i]);
6263 						goto fail;
6264 					}
6265 					*endptr = ':';
6266 				}
6267 			}
6268 
6269 			if (endptr == NULL) {
6270 				printf("Invalid [base:]offset format: %s\n", argv[i]);
6271 				goto fail;
6272 			}
6273 
6274 			if (*endptr == ':') {
6275 				pf_el->base_offs = htod16(rc);
6276 				rc = strtoul(endptr + 1, &endptr, 0);
6277 			} else {
6278 				/* Must have had a numeric offset only */
6279 				pf_el->base_offs = htod16(0);
6280 			}
6281 
6282 			if (*endptr) {
6283 				printf("Invalid [base:]offset format: %s\n", argv[i]);
6284 				goto fail;
6285 			}
6286 			if (rc > 0x0000FFFF) {
6287 				printf("Offset too large\n");
6288 				goto fail;
6289 			}
6290 			pf_el->rel_offs = htod16(rc);
6291 
6292 			/* Clear match_flag (may be set in parsing which follows) */
6293 			pf_el->match_flags = htod16(0);
6294 
6295 			/* Parse pattern filter mask and pattern directly into ioctl buffer */
6296 			if (argv[++i] == NULL) {
6297 				printf("Bitmask not provided\n");
6298 				goto fail;
6299 			}
6300 			rc = wl_pattern_atoh(argv[i], (char*)pf_el->mask_and_data);
6301 			if ((rc == -1) || (rc > MAX_PKTFLT_FIXED_PATTERN_SIZE)) {
6302 				printf("Rejecting: %s\n", argv[i]);
6303 				goto fail;
6304 			}
6305 			mask_size = htod16(rc);
6306 
6307 			if (argv[++i] == NULL) {
6308 				printf("Pattern not provided\n");
6309 				goto fail;
6310 			}
6311 
6312 			endptr = argv[i];
6313 			if (*endptr == '!') {
6314 				pf_el->match_flags =
6315 					htod16(WL_PKT_FILTER_MFLAG_NEG);
6316 				if (*(++endptr) == '\0') {
6317 					printf("Pattern not provided\n");
6318 					goto fail;
6319 				}
6320 			}
6321 			rc = wl_pattern_atoh(endptr, (char*)&pf_el->mask_and_data[rc]);
6322 			if ((rc == -1) || (rc > MAX_PKTFLT_FIXED_PATTERN_SIZE)) {
6323 				printf("Rejecting: %s\n", argv[i]);
6324 				goto fail;
6325 			}
6326 			pattern_size = htod16(rc);
6327 
6328 			if (mask_size != pattern_size) {
6329 				printf("Mask and pattern not the same size\n");
6330 				goto fail;
6331 			}
6332 
6333 			pf_el->size_bytes = mask_size;
6334 
6335 			/* Account for the size of this pattern element */
6336 			buf_len += WL_PKT_FILTER_PATTERN_LISTEL_FIXED_LEN + 2 * rc;
6337 
6338 			/* Move to next element location in ioctl buffer */
6339 			pf_el = (wl_pkt_filter_pattern_listel_t*)
6340 				((uint8*)pf_el + WL_PKT_FILTER_PATTERN_LISTEL_FIXED_LEN + 2 * rc);
6341 
6342 			/* Count list element */
6343 			list_cnt++;
6344 		}
6345 
6346 		/* Account for initial fixed size, and copy initial fixed fields */
6347 		buf_len += WL_PKT_FILTER_FIXED_LEN + WL_PKT_FILTER_PATTERN_LIST_FIXED_LEN;
6348 
6349 		if (buf_len > MAX_PKTFLT_BUF_SIZE) {
6350 			DHD_ERROR(("buffer over length MAX_PKTFLT_BUF_SIZE\n"));
6351 			goto fail;
6352 		}
6353 
6354 		/* Update list count and total size */
6355 		pkt_filter.u.patlist.list_cnt = list_cnt;
6356 		pkt_filter.u.patlist.PAD1[0] = 0;
6357 		pkt_filter.u.patlist.totsize = buf + buf_len - (char*)pkt_filterp;
6358 		pkt_filter.u.patlist.totsize -= WL_PKT_FILTER_FIXED_LEN;
6359 
6360 		memcpy((char *)pkt_filterp, &pkt_filter,
6361 			WL_PKT_FILTER_FIXED_LEN + WL_PKT_FILTER_PATTERN_LIST_FIXED_LEN);
6362 	} else {
6363 		DHD_ERROR(("Invalid filter type %d\n", pkt_filter.type));
6364 		goto fail;
6365 	}
6366 
6367 	rc = dhd_wl_ioctl_cmd(dhd, WLC_SET_VAR, buf, buf_len, TRUE, 0);
6368 	rc = rc >= 0 ? 0 : rc;
6369 
6370 	if (rc)
6371 		DHD_ERROR(("%s: failed to add pktfilter %s, retcode = %d\n",
6372 		__FUNCTION__, arg, rc));
6373 	else
6374 		DHD_TRACE(("%s: successfully added pktfilter %s\n",
6375 		__FUNCTION__, arg));
6376 
6377 fail:
6378 	if (arg_org)
6379 		MFREE(dhd->osh, arg_org, strlen(arg) + 1);
6380 
6381 	if (buf)
6382 		MFREE(dhd->osh, buf, MAX_PKTFLT_BUF_SIZE);
6383 }
6384 
6385 void
dhd_pktfilter_offload_delete(dhd_pub_t * dhd,int id)6386 dhd_pktfilter_offload_delete(dhd_pub_t *dhd, int id)
6387 {
6388 	int ret;
6389 
6390 	ret = dhd_wl_ioctl_set_intiovar(dhd, "pkt_filter_delete",
6391 		id, WLC_SET_VAR, TRUE, 0);
6392 	if (ret < 0) {
6393 		DHD_ERROR(("%s: Failed to delete filter ID:%d, ret=%d\n",
6394 			__FUNCTION__, id, ret));
6395 	}
6396 	else
6397 		DHD_TRACE(("%s: successfully deleted pktfilter %d\n",
6398 		__FUNCTION__, id));
6399 }
6400 #endif /* PKT_FILTER_SUPPORT */
6401 
6402 /* ========================== */
6403 /* ==== ARP OFFLOAD SUPPORT = */
6404 /* ========================== */
6405 #ifdef ARP_OFFLOAD_SUPPORT
6406 void
dhd_arp_offload_set(dhd_pub_t * dhd,int arp_mode)6407 dhd_arp_offload_set(dhd_pub_t * dhd, int arp_mode)
6408 {
6409 	int retcode;
6410 
6411 	retcode = dhd_wl_ioctl_set_intiovar(dhd, "arp_ol",
6412 		arp_mode, WLC_SET_VAR, TRUE, 0);
6413 
6414 	retcode = retcode >= 0 ? 0 : retcode;
6415 	if (retcode) {
6416 		DHD_ERROR(("%s: failed to set ARP offload mode to 0x%x, retcode = %d\n",
6417 			__FUNCTION__, arp_mode, retcode));
6418 	} else {
6419 		DHD_ARPOE(("%s: successfully set ARP offload mode to 0x%x\n",
6420 			__FUNCTION__, arp_mode));
6421 		dhd->arpol_configured = TRUE;
6422 	}
6423 }
6424 
6425 void
dhd_arp_offload_enable(dhd_pub_t * dhd,int arp_enable)6426 dhd_arp_offload_enable(dhd_pub_t * dhd, int arp_enable)
6427 {
6428 	int retcode;
6429 
6430 	if (!dhd->arpol_configured) {
6431 		/* If arpol is not applied, apply it */
6432 		dhd_arp_offload_set(dhd, dhd_arp_mode);
6433 	}
6434 
6435 	retcode = dhd_wl_ioctl_set_intiovar(dhd, "arpoe",
6436 		arp_enable, WLC_SET_VAR, TRUE, 0);
6437 	retcode = retcode >= 0 ? 0 : retcode;
6438 	if (retcode)
6439 		DHD_ERROR(("%s: failed to enabe ARP offload to %d, retcode = %d\n",
6440 			__FUNCTION__, arp_enable, retcode));
6441 	else
6442 #ifdef DHD_LOG_DUMP
6443 		DHD_LOG_MEM(("%s: successfully enabed ARP offload to %d\n",
6444 			__FUNCTION__, arp_enable));
6445 #else
6446 		DHD_ARPOE(("%s: successfully enabed ARP offload to %d\n",
6447 			__FUNCTION__, arp_enable));
6448 #endif /* DHD_LOG_DUMP */
6449 	if (arp_enable) {
6450 		uint32 version;
6451 		retcode = dhd_wl_ioctl_get_intiovar(dhd, "arp_version",
6452 			&version, WLC_GET_VAR, FALSE, 0);
6453 		if (retcode) {
6454 			DHD_INFO(("%s: fail to get version (maybe version 1:retcode = %d\n",
6455 				__FUNCTION__, retcode));
6456 			dhd->arp_version = 1;
6457 		}
6458 		else {
6459 			DHD_INFO(("%s: ARP Version= %x\n", __FUNCTION__, version));
6460 			dhd->arp_version = version;
6461 		}
6462 	}
6463 }
6464 
6465 /* XXX ANDREY: clear AOE arp_table  */
6466 void
dhd_aoe_arp_clr(dhd_pub_t * dhd,int idx)6467 dhd_aoe_arp_clr(dhd_pub_t *dhd, int idx)
6468 {
6469 	int ret = 0;
6470 
6471 	if (dhd == NULL) return;
6472 	if (dhd->arp_version == 1)
6473 		idx = 0;
6474 
6475 	ret = dhd_iovar(dhd, idx, "arp_table_clear", NULL, 0, NULL, 0, TRUE);
6476 	if (ret < 0)
6477 		DHD_ERROR(("%s failed code %d\n", __FUNCTION__, ret));
6478 	else {
6479 #ifdef DHD_LOG_DUMP
6480 		DHD_LOG_MEM(("%s: ARP table clear\n", __FUNCTION__));
6481 #else
6482 		DHD_TRACE(("%s: ARP table clear\n", __FUNCTION__));
6483 #endif /* DHD_LOG_DUMP */
6484 	}
6485 	/* mac address isn't cleared here but it will be cleared after dongle off */
6486 	dhd->hmac_updated = 0;
6487 }
6488 
6489 /* XXX ANDREY: clear hostip table  */
6490 void
dhd_aoe_hostip_clr(dhd_pub_t * dhd,int idx)6491 dhd_aoe_hostip_clr(dhd_pub_t *dhd, int idx)
6492 {
6493 	int ret = 0;
6494 
6495 	if (dhd == NULL) return;
6496 	if (dhd->arp_version == 1)
6497 		idx = 0;
6498 
6499 	ret = dhd_iovar(dhd, idx, "arp_hostip_clear", NULL, 0, NULL, 0, TRUE);
6500 	if (ret < 0)
6501 		DHD_ERROR(("%s failed code %d\n", __FUNCTION__, ret));
6502 	else {
6503 #ifdef DHD_LOG_DUMP
6504 		DHD_LOG_MEM(("%s: ARP host ip clear\n", __FUNCTION__));
6505 #else
6506 		DHD_TRACE(("%s: ARP host ip clear\n", __FUNCTION__));
6507 #endif /* DHD_LOG_DUMP */
6508 	}
6509 }
6510 
6511 void
dhd_arp_offload_add_ip(dhd_pub_t * dhd,uint32 ipaddr,int idx)6512 dhd_arp_offload_add_ip(dhd_pub_t *dhd, uint32 ipaddr, int idx)
6513 {
6514 	int ret;
6515 
6516 	if (dhd == NULL) return;
6517 	if (dhd->arp_version == 1)
6518 		idx = 0;
6519 
6520 	ret = dhd_iovar(dhd, idx, "arp_hostip", (char *)&ipaddr, sizeof(ipaddr),
6521 			NULL, 0, TRUE);
6522 	if (ret < 0)
6523 		DHD_ERROR(("%s: ARP ip addr add failed, ret = %d\n", __FUNCTION__, ret));
6524 	else {
6525 		/* mac address is updated in the dongle */
6526 		dhd->hmac_updated = 1;
6527 #ifdef DHD_LOG_DUMP
6528 		DHD_LOG_MEM(("%s: ARP ip addr entry added \n", __FUNCTION__));
6529 #else
6530 		DHD_ARPOE(("%s: ARP ip addr entry added \n", __FUNCTION__));
6531 #endif /* DHD_LOG_DUMP */
6532 	}
6533 }
6534 
6535 int
dhd_arp_get_arp_hostip_table(dhd_pub_t * dhd,void * buf,int buflen,int idx)6536 dhd_arp_get_arp_hostip_table(dhd_pub_t *dhd, void *buf, int buflen, int idx)
6537 {
6538 	int ret, i;
6539 	uint32 *ptr32 = buf;
6540 	bool clr_bottom = FALSE;
6541 
6542 	if (!buf)
6543 		return -1;
6544 	if (dhd == NULL) return -1;
6545 	if (dhd->arp_version == 1)
6546 		idx = 0;
6547 
6548 	ret = dhd_iovar(dhd, idx, "arp_hostip", NULL, 0, (char *)buf, buflen,
6549 			FALSE);
6550 	if (ret) {
6551 		DHD_ERROR(("%s: ioctl WLC_GET_VAR error %d\n",
6552 		__FUNCTION__, ret));
6553 
6554 		return -1;
6555 	}
6556 
6557 	/* clean up the buf, ascii reminder */
6558 	for (i = 0; i < MAX_IPV4_ENTRIES; i++) {
6559 		if (!clr_bottom) {
6560 			if (*ptr32 == 0)
6561 				clr_bottom = TRUE;
6562 		} else {
6563 			*ptr32 = 0;
6564 		}
6565 		ptr32++;
6566 	}
6567 
6568 	return 0;
6569 }
6570 #endif /* ARP_OFFLOAD_SUPPORT  */
6571 
6572 /*
6573  * Neighbor Discovery Offload: enable NDO feature
6574  * Called  by ipv6 event handler when interface comes up/goes down
6575  */
6576 int
dhd_ndo_enable(dhd_pub_t * dhd,int ndo_enable)6577 dhd_ndo_enable(dhd_pub_t * dhd, int ndo_enable)
6578 {
6579 	int retcode;
6580 
6581 	if (dhd == NULL)
6582 		return -1;
6583 
6584 #if defined(WL_CFG80211) && defined(WL_NAN)
6585 	if (wl_cfgnan_is_dp_active(dhd_linux_get_primary_netdev(dhd))) {
6586 		/* If nan dp is active, skip NDO */
6587 		DHD_INFO(("Active NAN DP, skip NDO\n"));
6588 		return 0;
6589 	}
6590 #endif /* WL_CFG80211 && WL_NAN */
6591 #ifdef WL_CFG80211
6592 	if (dhd->op_mode & DHD_FLAG_HOSTAP_MODE) {
6593 		/* NDO disable on STA+SOFTAP mode */
6594 		ndo_enable = FALSE;
6595 	}
6596 #endif /* WL_CFG80211 */
6597 	retcode = dhd_wl_ioctl_set_intiovar(dhd, "ndoe",
6598 		ndo_enable, WLC_SET_VAR, TRUE, 0);
6599 	if (retcode)
6600 		DHD_ERROR(("%s: failed to enabe ndo to %d, retcode = %d\n",
6601 			__FUNCTION__, ndo_enable, retcode));
6602 	else
6603 		DHD_TRACE(("%s: successfully enabed ndo offload to %d\n",
6604 			__FUNCTION__, ndo_enable));
6605 
6606 	return retcode;
6607 }
6608 
6609 /*
6610  * Neighbor Discover Offload: enable NDO feature
6611  * Called  by ipv6 event handler when interface comes up
6612  */
6613 int
dhd_ndo_add_ip(dhd_pub_t * dhd,char * ipv6addr,int idx)6614 dhd_ndo_add_ip(dhd_pub_t *dhd, char* ipv6addr, int idx)
6615 {
6616 	int iov_len = 0;
6617 	char iovbuf[DHD_IOVAR_BUF_SIZE];
6618 	int retcode;
6619 
6620 	if (dhd == NULL)
6621 		return -1;
6622 
6623 	iov_len = bcm_mkiovar("nd_hostip", (char *)ipv6addr,
6624 		IPV6_ADDR_LEN, iovbuf, sizeof(iovbuf));
6625 	if (!iov_len) {
6626 		DHD_ERROR(("%s: Insufficient iovar buffer size %zu \n",
6627 			__FUNCTION__, sizeof(iovbuf)));
6628 		return -1;
6629 	}
6630 	retcode = dhd_wl_ioctl_cmd(dhd, WLC_SET_VAR, iovbuf, iov_len, TRUE, idx);
6631 
6632 	if (retcode)
6633 		DHD_ERROR(("%s: ndo ip addr add failed, retcode = %d\n",
6634 		__FUNCTION__, retcode));
6635 	else
6636 		DHD_TRACE(("%s: ndo ipaddr entry added \n",
6637 		__FUNCTION__));
6638 
6639 	return retcode;
6640 }
6641 
6642 /*
6643  * Neighbor Discover Offload: enable NDO feature
6644  * Called  by ipv6 event handler when interface goes down
6645  */
6646 int
dhd_ndo_remove_ip(dhd_pub_t * dhd,int idx)6647 dhd_ndo_remove_ip(dhd_pub_t *dhd, int idx)
6648 {
6649 	int iov_len = 0;
6650 	char iovbuf[DHD_IOVAR_BUF_SIZE];
6651 	int retcode;
6652 
6653 	if (dhd == NULL)
6654 		return -1;
6655 
6656 	iov_len = bcm_mkiovar("nd_hostip_clear", NULL,
6657 		0, iovbuf, sizeof(iovbuf));
6658 	if (!iov_len) {
6659 		DHD_ERROR(("%s: Insufficient iovar buffer size %zu \n",
6660 			__FUNCTION__, sizeof(iovbuf)));
6661 		return -1;
6662 	}
6663 	retcode = dhd_wl_ioctl_cmd(dhd, WLC_SET_VAR, iovbuf, iov_len, TRUE, idx);
6664 
6665 	if (retcode)
6666 		DHD_ERROR(("%s: ndo ip addr remove failed, retcode = %d\n",
6667 		__FUNCTION__, retcode));
6668 	else
6669 		DHD_TRACE(("%s: ndo ipaddr entry removed \n",
6670 		__FUNCTION__));
6671 
6672 	return retcode;
6673 }
6674 /* Enhanced ND offload */
6675 uint16
dhd_ndo_get_version(dhd_pub_t * dhdp)6676 dhd_ndo_get_version(dhd_pub_t *dhdp)
6677 {
6678 	char iovbuf[DHD_IOVAR_BUF_SIZE];
6679 	wl_nd_hostip_t ndo_get_ver;
6680 	int iov_len;
6681 	int retcode;
6682 	uint16 ver = 0;
6683 
6684 	if (dhdp == NULL) {
6685 		return BCME_ERROR;
6686 	}
6687 
6688 	memset(&iovbuf, 0, sizeof(iovbuf));
6689 	ndo_get_ver.version = htod16(WL_ND_HOSTIP_IOV_VER);
6690 	ndo_get_ver.op_type = htod16(WL_ND_HOSTIP_OP_VER);
6691 	ndo_get_ver.length = htod32(WL_ND_HOSTIP_FIXED_LEN + sizeof(uint16));
6692 	ndo_get_ver.u.version = 0;
6693 	iov_len = bcm_mkiovar("nd_hostip", (char *)&ndo_get_ver,
6694 		WL_ND_HOSTIP_FIXED_LEN + sizeof(uint16), iovbuf, sizeof(iovbuf));
6695 
6696 	if (!iov_len) {
6697 		DHD_ERROR(("%s: Insufficient iovar buffer size %zu \n",
6698 			__FUNCTION__, sizeof(iovbuf)));
6699 		return BCME_ERROR;
6700 	}
6701 
6702 	retcode = dhd_wl_ioctl_cmd(dhdp, WLC_GET_VAR, iovbuf, iov_len, FALSE, 0);
6703 
6704 	if (retcode) {
6705 		DHD_ERROR(("%s: failed, retcode = %d\n", __FUNCTION__, retcode));
6706 		/* ver iovar not supported. NDO version is 0 */
6707 		ver = 0;
6708 	} else {
6709 		wl_nd_hostip_t *ndo_ver_ret = (wl_nd_hostip_t *)iovbuf;
6710 
6711 		if ((dtoh16(ndo_ver_ret->version) == WL_ND_HOSTIP_IOV_VER) &&
6712 				(dtoh16(ndo_ver_ret->op_type) == WL_ND_HOSTIP_OP_VER) &&
6713 				(dtoh32(ndo_ver_ret->length) == WL_ND_HOSTIP_FIXED_LEN
6714 					+ sizeof(uint16))) {
6715 			/* nd_hostip iovar version */
6716 			ver = dtoh16(ndo_ver_ret->u.version);
6717 		}
6718 
6719 		DHD_TRACE(("%s: successfully get version: %d\n", __FUNCTION__, ver));
6720 	}
6721 
6722 	return ver;
6723 }
6724 
6725 int
dhd_ndo_add_ip_with_type(dhd_pub_t * dhdp,char * ipv6addr,uint8 type,int idx)6726 dhd_ndo_add_ip_with_type(dhd_pub_t *dhdp, char *ipv6addr, uint8 type, int idx)
6727 {
6728 	char iovbuf[DHD_IOVAR_BUF_SIZE];
6729 	wl_nd_hostip_t ndo_add_addr;
6730 	int iov_len;
6731 	int retcode;
6732 
6733 	if (dhdp == NULL || ipv6addr == 0) {
6734 		return BCME_ERROR;
6735 	}
6736 
6737 	/* wl_nd_hostip_t fixed param */
6738 	ndo_add_addr.version = htod16(WL_ND_HOSTIP_IOV_VER);
6739 	ndo_add_addr.op_type = htod16(WL_ND_HOSTIP_OP_ADD);
6740 	ndo_add_addr.length = htod32(WL_ND_HOSTIP_WITH_ADDR_LEN);
6741 	/* wl_nd_host_ip_addr_t param for add */
6742 	memcpy(&ndo_add_addr.u.host_ip.ip_addr, ipv6addr, IPV6_ADDR_LEN);
6743 	ndo_add_addr.u.host_ip.type = type;
6744 
6745 	iov_len = bcm_mkiovar("nd_hostip", (char *)&ndo_add_addr,
6746 		WL_ND_HOSTIP_WITH_ADDR_LEN, iovbuf, sizeof(iovbuf));
6747 	if (!iov_len) {
6748 		DHD_ERROR(("%s: Insufficient iovar buffer size %zu \n",
6749 			__FUNCTION__, sizeof(iovbuf)));
6750 		return BCME_ERROR;
6751 	}
6752 
6753 	retcode = dhd_wl_ioctl_cmd(dhdp, WLC_SET_VAR, iovbuf, iov_len, TRUE, idx);
6754 	if (retcode) {
6755 		DHD_ERROR(("%s: failed, retcode = %d\n", __FUNCTION__, retcode));
6756 #ifdef NDO_CONFIG_SUPPORT
6757 		if (retcode == BCME_NORESOURCE) {
6758 			/* number of host ip addr exceeds FW capacity, Deactivate ND offload */
6759 			DHD_INFO(("%s: Host IP count exceed device capacity,"
6760 				"ND offload deactivated\n", __FUNCTION__));
6761 			dhdp->ndo_host_ip_overflow = TRUE;
6762 			dhd_ndo_enable(dhdp, FALSE);
6763 		}
6764 #endif /* NDO_CONFIG_SUPPORT */
6765 	} else {
6766 		DHD_TRACE(("%s: successfully added: %d\n", __FUNCTION__, retcode));
6767 	}
6768 
6769 	return retcode;
6770 }
6771 
6772 int
dhd_ndo_remove_ip_by_addr(dhd_pub_t * dhdp,char * ipv6addr,int idx)6773 dhd_ndo_remove_ip_by_addr(dhd_pub_t *dhdp, char *ipv6addr, int idx)
6774 {
6775 	char iovbuf[DHD_IOVAR_BUF_SIZE];
6776 	wl_nd_hostip_t ndo_del_addr;
6777 	int iov_len;
6778 	int retcode;
6779 
6780 	if (dhdp == NULL || ipv6addr == 0) {
6781 		return BCME_ERROR;
6782 	}
6783 
6784 	/* wl_nd_hostip_t fixed param */
6785 	ndo_del_addr.version = htod16(WL_ND_HOSTIP_IOV_VER);
6786 	ndo_del_addr.op_type = htod16(WL_ND_HOSTIP_OP_DEL);
6787 	ndo_del_addr.length = htod32(WL_ND_HOSTIP_WITH_ADDR_LEN);
6788 	/* wl_nd_host_ip_addr_t param for del */
6789 	memcpy(&ndo_del_addr.u.host_ip.ip_addr, ipv6addr, IPV6_ADDR_LEN);
6790 	ndo_del_addr.u.host_ip.type = 0;	/* don't care */
6791 
6792 	iov_len = bcm_mkiovar("nd_hostip", (char *)&ndo_del_addr,
6793 		WL_ND_HOSTIP_WITH_ADDR_LEN, iovbuf, sizeof(iovbuf));
6794 
6795 	if (!iov_len) {
6796 		DHD_ERROR(("%s: Insufficient iovar buffer size %zu \n",
6797 			__FUNCTION__, sizeof(iovbuf)));
6798 		return BCME_ERROR;
6799 	}
6800 
6801 	retcode = dhd_wl_ioctl_cmd(dhdp, WLC_SET_VAR, iovbuf, iov_len, TRUE, idx);
6802 	if (retcode) {
6803 		DHD_ERROR(("%s: failed, retcode = %d\n", __FUNCTION__, retcode));
6804 	} else {
6805 		DHD_TRACE(("%s: successfully removed: %d\n", __FUNCTION__, retcode));
6806 	}
6807 
6808 	return retcode;
6809 }
6810 
6811 int
dhd_ndo_remove_ip_by_type(dhd_pub_t * dhdp,uint8 type,int idx)6812 dhd_ndo_remove_ip_by_type(dhd_pub_t *dhdp, uint8 type, int idx)
6813 {
6814 	char iovbuf[DHD_IOVAR_BUF_SIZE];
6815 	wl_nd_hostip_t ndo_del_addr;
6816 	int iov_len;
6817 	int retcode;
6818 
6819 	if (dhdp == NULL) {
6820 		return BCME_ERROR;
6821 	}
6822 
6823 	/* wl_nd_hostip_t fixed param */
6824 	ndo_del_addr.version = htod16(WL_ND_HOSTIP_IOV_VER);
6825 	if (type == WL_ND_IPV6_ADDR_TYPE_UNICAST) {
6826 		ndo_del_addr.op_type = htod16(WL_ND_HOSTIP_OP_DEL_UC);
6827 	} else if (type == WL_ND_IPV6_ADDR_TYPE_ANYCAST) {
6828 		ndo_del_addr.op_type = htod16(WL_ND_HOSTIP_OP_DEL_AC);
6829 	} else {
6830 		return BCME_BADARG;
6831 	}
6832 	ndo_del_addr.length = htod32(WL_ND_HOSTIP_FIXED_LEN);
6833 
6834 	iov_len = bcm_mkiovar("nd_hostip", (char *)&ndo_del_addr, WL_ND_HOSTIP_FIXED_LEN,
6835 			iovbuf, sizeof(iovbuf));
6836 
6837 	if (!iov_len) {
6838 		DHD_ERROR(("%s: Insufficient iovar buffer size %zu \n",
6839 			__FUNCTION__, sizeof(iovbuf)));
6840 		return BCME_ERROR;
6841 	}
6842 
6843 	retcode = dhd_wl_ioctl_cmd(dhdp, WLC_SET_VAR, iovbuf, iov_len, TRUE, idx);
6844 	if (retcode) {
6845 		DHD_ERROR(("%s: failed, retcode = %d\n", __FUNCTION__, retcode));
6846 	} else {
6847 		DHD_TRACE(("%s: successfully removed: %d\n", __FUNCTION__, retcode));
6848 	}
6849 
6850 	return retcode;
6851 }
6852 
6853 int
dhd_ndo_unsolicited_na_filter_enable(dhd_pub_t * dhdp,int enable)6854 dhd_ndo_unsolicited_na_filter_enable(dhd_pub_t *dhdp, int enable)
6855 {
6856 	char iovbuf[DHD_IOVAR_BUF_SIZE];
6857 	int iov_len;
6858 	int retcode;
6859 
6860 	if (dhdp == NULL) {
6861 		return BCME_ERROR;
6862 	}
6863 
6864 	iov_len = bcm_mkiovar("nd_unsolicited_na_filter", (char *)&enable, sizeof(int),
6865 			iovbuf, sizeof(iovbuf));
6866 
6867 	if (!iov_len) {
6868 		DHD_ERROR(("%s: Insufficient iovar buffer size %zu \n",
6869 			__FUNCTION__, sizeof(iovbuf)));
6870 		return BCME_ERROR;
6871 	}
6872 
6873 	retcode = dhd_wl_ioctl_cmd(dhdp, WLC_SET_VAR, iovbuf, iov_len, TRUE, 0);
6874 	if (retcode)
6875 		DHD_ERROR(("%s: failed to enable Unsolicited NA filter to %d, retcode = %d\n",
6876 			__FUNCTION__, enable, retcode));
6877 	else {
6878 		DHD_TRACE(("%s: successfully enabled Unsolicited NA filter to %d\n",
6879 			__FUNCTION__, enable));
6880 	}
6881 
6882 	return retcode;
6883 }
6884 #ifdef SIMPLE_ISCAN
6885 
6886 uint iscan_thread_id = 0;
6887 iscan_buf_t * iscan_chain = 0;
6888 
6889 iscan_buf_t *
dhd_iscan_allocate_buf(dhd_pub_t * dhd,iscan_buf_t ** iscanbuf)6890 dhd_iscan_allocate_buf(dhd_pub_t *dhd, iscan_buf_t **iscanbuf)
6891 {
6892 	iscan_buf_t *iscanbuf_alloc = 0;
6893 	iscan_buf_t *iscanbuf_head;
6894 
6895 	DHD_ISCAN(("%s: Entered\n", __FUNCTION__));
6896 	dhd_iscan_lock();
6897 
6898 	iscanbuf_alloc = (iscan_buf_t*)MALLOC(dhd->osh, sizeof(iscan_buf_t));
6899 	if (iscanbuf_alloc == NULL)
6900 		goto fail;
6901 
6902 	iscanbuf_alloc->next = NULL;
6903 	iscanbuf_head = *iscanbuf;
6904 
6905 	DHD_ISCAN(("%s: addr of allocated node = 0x%X"
6906 		   "addr of iscanbuf_head = 0x%X dhd = 0x%X\n",
6907 		   __FUNCTION__, iscanbuf_alloc, iscanbuf_head, dhd));
6908 
6909 	if (iscanbuf_head == NULL) {
6910 		*iscanbuf = iscanbuf_alloc;
6911 		DHD_ISCAN(("%s: Head is allocated\n", __FUNCTION__));
6912 		goto fail;
6913 	}
6914 
6915 	while (iscanbuf_head->next)
6916 		iscanbuf_head = iscanbuf_head->next;
6917 
6918 	iscanbuf_head->next = iscanbuf_alloc;
6919 
6920 fail:
6921 	dhd_iscan_unlock();
6922 	return iscanbuf_alloc;
6923 }
6924 
6925 void
dhd_iscan_free_buf(void * dhdp,iscan_buf_t * iscan_delete)6926 dhd_iscan_free_buf(void *dhdp, iscan_buf_t *iscan_delete)
6927 {
6928 	iscan_buf_t *iscanbuf_free = 0;
6929 	iscan_buf_t *iscanbuf_prv = 0;
6930 	iscan_buf_t *iscanbuf_cur;
6931 	dhd_pub_t *dhd = dhd_bus_pub(dhdp);
6932 	DHD_ISCAN(("%s: Entered\n", __FUNCTION__));
6933 
6934 	dhd_iscan_lock();
6935 
6936 	iscanbuf_cur = iscan_chain;
6937 
6938 	/* If iscan_delete is null then delete the entire
6939 	 * chain or else delete specific one provided
6940 	 */
6941 	if (!iscan_delete) {
6942 		while (iscanbuf_cur) {
6943 			iscanbuf_free = iscanbuf_cur;
6944 			iscanbuf_cur = iscanbuf_cur->next;
6945 			iscanbuf_free->next = 0;
6946 			MFREE(dhd->osh, iscanbuf_free, sizeof(iscan_buf_t));
6947 		}
6948 		iscan_chain = 0;
6949 	} else {
6950 		while (iscanbuf_cur) {
6951 			if (iscanbuf_cur == iscan_delete)
6952 				break;
6953 			iscanbuf_prv = iscanbuf_cur;
6954 			iscanbuf_cur = iscanbuf_cur->next;
6955 		}
6956 		if (iscanbuf_prv)
6957 			iscanbuf_prv->next = iscan_delete->next;
6958 
6959 		iscan_delete->next = 0;
6960 		MFREE(dhd->osh, iscan_delete, sizeof(iscan_buf_t));
6961 
6962 		if (!iscanbuf_prv)
6963 			iscan_chain = 0;
6964 	}
6965 	dhd_iscan_unlock();
6966 }
6967 
6968 iscan_buf_t *
dhd_iscan_result_buf(void)6969 dhd_iscan_result_buf(void)
6970 {
6971 	return iscan_chain;
6972 }
6973 
6974 int
dhd_iscan_issue_request(void * dhdp,wl_iscan_params_t * pParams,uint32 size)6975 dhd_iscan_issue_request(void * dhdp, wl_iscan_params_t *pParams, uint32 size)
6976 {
6977 	int rc = -1;
6978 	dhd_pub_t *dhd = dhd_bus_pub(dhdp);
6979 	char *buf;
6980 	char iovar[] = "iscan";
6981 	uint32 allocSize = 0;
6982 	wl_ioctl_t ioctl;
6983 	int len;
6984 
6985 	if (pParams) {
6986 		allocSize = (size + strlen(iovar) + 1);
6987 		if ((allocSize < size) || (allocSize < strlen(iovar)))
6988 		{
6989 			DHD_ERROR(("%s: overflow - allocation size too large %d < %d + %d!\n",
6990 				__FUNCTION__, allocSize, size, strlen(iovar)));
6991 			goto cleanUp;
6992 		}
6993 		buf = MALLOC(dhd->osh, allocSize);
6994 
6995 		if (buf == NULL)
6996 			{
6997 			DHD_ERROR(("%s: malloc of size %d failed!\n", __FUNCTION__, allocSize));
6998 			goto cleanUp;
6999 			}
7000 		ioctl.cmd = WLC_SET_VAR;
7001 		len = bcm_mkiovar(iovar, (char *)pParams, size, buf, allocSize);
7002 		if (len == 0) {
7003 			rc = BCME_BUFTOOSHORT;
7004 			goto cleanUp;
7005 		}
7006 		rc = dhd_wl_ioctl(dhd, 0, &ioctl, buf, len);
7007 	}
7008 
7009 cleanUp:
7010 	if (buf) {
7011 		MFREE(dhd->osh, buf, allocSize);
7012 	}
7013 
7014 	return rc;
7015 }
7016 
7017 static int
dhd_iscan_get_partial_result(void * dhdp,uint * scan_count)7018 dhd_iscan_get_partial_result(void *dhdp, uint *scan_count)
7019 {
7020 	wl_iscan_results_t *list_buf;
7021 	wl_iscan_results_t list;
7022 	wl_scan_results_t *results;
7023 	iscan_buf_t *iscan_cur;
7024 	int status = -1;
7025 	dhd_pub_t *dhd = dhd_bus_pub(dhdp);
7026 	int rc;
7027 	wl_ioctl_t ioctl;
7028 	int len;
7029 
7030 	DHD_ISCAN(("%s: Enter\n", __FUNCTION__));
7031 
7032 	iscan_cur = dhd_iscan_allocate_buf(dhd, &iscan_chain);
7033 	if (!iscan_cur) {
7034 		DHD_ERROR(("%s: Failed to allocate node\n", __FUNCTION__));
7035 		dhd_iscan_free_buf(dhdp, 0);
7036 		dhd_iscan_request(dhdp, WL_SCAN_ACTION_ABORT);
7037 		dhd_ind_scan_confirm(dhdp, FALSE);
7038 		goto fail;
7039 	}
7040 
7041 	dhd_iscan_lock();
7042 
7043 	memset(iscan_cur->iscan_buf, 0, WLC_IW_ISCAN_MAXLEN);
7044 	list_buf = (wl_iscan_results_t*)iscan_cur->iscan_buf;
7045 	results = &list_buf->results;
7046 	results->buflen = WL_ISCAN_RESULTS_FIXED_SIZE;
7047 	results->version = 0;
7048 	results->count = 0;
7049 
7050 	memset(&list, 0, sizeof(list));
7051 	list.results.buflen = htod32(WLC_IW_ISCAN_MAXLEN);
7052 	len = bcm_mkiovar("iscanresults", (char *)&list, WL_ISCAN_RESULTS_FIXED_SIZE,
7053 		iscan_cur->iscan_buf, WLC_IW_ISCAN_MAXLEN);
7054 	if (len == 0) {
7055 		dhd_iscan_free_buf(dhdp, 0);
7056 		dhd_iscan_request(dhdp, WL_SCAN_ACTION_ABORT);
7057 		dhd_ind_scan_confirm(dhdp, FALSE);
7058 		status = BCME_BUFTOOSHORT;
7059 		goto fail;
7060 	}
7061 	ioctl.cmd = WLC_GET_VAR;
7062 	ioctl.set = FALSE;
7063 	rc = dhd_wl_ioctl(dhd, 0, &ioctl, iscan_cur->iscan_buf, WLC_IW_ISCAN_MAXLEN);
7064 
7065 	results->buflen = dtoh32(results->buflen);
7066 	results->version = dtoh32(results->version);
7067 	*scan_count = results->count = dtoh32(results->count);
7068 	status = dtoh32(list_buf->status);
7069 	DHD_ISCAN(("%s: Got %d resuls status = (%x)\n", __FUNCTION__, results->count, status));
7070 
7071 	dhd_iscan_unlock();
7072 
7073 	if (!(*scan_count)) {
7074 		 /* TODO: race condition when FLUSH already called */
7075 		dhd_iscan_free_buf(dhdp, 0);
7076 	}
7077 fail:
7078 	return status;
7079 }
7080 
7081 #ifdef NDIS
7082 /* XXXX Following code had bit of OS dependency.
7083  * Cleanup to move the OS dependency to other
7084  * per port code so that iscan logic here can be
7085  * leveraged across all OS's
7086  */
7087 NDIS_EVENT iscan_event;
7088 HANDLE tHandle;
7089 NDIS_SPIN_LOCK	dhd_iscan_queue_lock;
7090 
7091 void
dhd_iscan_lock(void)7092 dhd_iscan_lock(void)
7093 {
7094 	NdisAcquireSpinLock(&dhd_iscan_queue_lock);
7095 }
7096 
7097 void
dhd_iscan_unlock(void)7098 dhd_iscan_unlock(void)
7099 {
7100 	NdisReleaseSpinLock(&dhd_iscan_queue_lock);
7101 }
7102 
7103 void
dhd_iscan_notify(void)7104 dhd_iscan_notify(void)
7105 {
7106 	DHD_ISCAN(("%s: Entered\n", __FUNCTION__));
7107 	NdisSetEvent(&iscan_event);
7108 }
7109 
7110 static void
dhd_iscan_func(void * h)7111 dhd_iscan_func(void *h)
7112 {
7113 	int status;
7114 	uint scan_count;
7115 	dhd_pub_t *dhd = dhd_bus_pub(h);
7116 
7117 	/* Read the priority from registry */
7118 	CeSetThreadPriority(GetCurrentThread(), 128);
7119 	DHD_ISCAN(("%s: thread created\n", __FUNCTION__));
7120 
7121 	while (TRUE) {
7122 		NdisWaitEvent(&iscan_event, 0);		/* wait forever */
7123 		NdisResetEvent(&iscan_event);		/* reset the event */
7124 		DHD_ISCAN(("%s: thread scheduled\n", __FUNCTION__));
7125 
7126 		status = dhd_iscan_get_partial_result(h, &scan_count);
7127 
7128 		if (status == WL_SCAN_RESULTS_PARTIAL) {
7129 			dhd_iscan_request(h, WL_SCAN_ACTION_CONTINUE);
7130 		} else if (status == WL_SCAN_RESULTS_SUCCESS) {
7131 			if (dhd_iscan_in_progress(h)) {
7132 				dhd_ind_scan_confirm(h, TRUE);
7133 			}
7134 		} else if (status == WL_SCAN_RESULTS_ABORTED ||
7135 			status == WL_SCAN_RESULTS_NO_MEM) {
7136 			dhd_iscan_request(h, WL_SCAN_ACTION_ABORT);
7137 			dhd_ind_scan_confirm(h, FALSE);
7138 		} else {
7139 			dhd_iscan_request(h, WL_SCAN_ACTION_ABORT);
7140 			dhd_ind_scan_confirm(h, FALSE);
7141 		}
7142 	}
7143 }
7144 
7145 int
dhd_iscan_attach(void * dhdp)7146 dhd_iscan_attach(void *dhdp)
7147 {
7148 	DHD_ISCAN(("%s: dhdp = 0x%x\n", __FUNCTION__, dhdp));
7149 
7150 	NdisInitializeEvent(&iscan_event);
7151 	NdisResetEvent(&iscan_event);
7152 	NdisAllocateSpinLock(&dhd_iscan_queue_lock);
7153 
7154 	/* XXX - should move to ndishared sublayer */
7155 	tHandle = CreateThread(NULL,
7156 		0,
7157 		(LPTHREAD_START_ROUTINE)dhd_iscan_func,
7158 		(void *)dhdp,
7159 		0,
7160 		&iscan_thread_id);
7161 
7162 	if (!iscan_thread_id)
7163 		return NDIS_STATUS_FAILURE;
7164 
7165 	return NDIS_STATUS_SUCCESS;
7166 }
7167 
7168 void
dhd_iscan_deattach(void * dhdp)7169 dhd_iscan_deattach(void *dhdp)
7170 {
7171 	if (iscan_thread_id)
7172 	{
7173 	NdisFreeEvent(&iscan_event);
7174 	NdisFreeSpinLock(&dhd_iscan_queue_lock);
7175 	CloseHandle(tHandle);
7176 		iscan_thread_id = 0;
7177 	}
7178 }
7179 #endif /* NDIS */
7180 #endif /* SIMPLE_ISCAN */
7181 
7182 /*
7183  * returns = TRUE if associated, FALSE if not associated
7184  */
dhd_is_associated(dhd_pub_t * dhd,uint8 ifidx,int * retval)7185 bool dhd_is_associated(dhd_pub_t *dhd, uint8 ifidx, int *retval)
7186 {
7187 	char bssid[6], zbuf[6];
7188 	int ret = -1;
7189 
7190 	bzero(bssid, 6);
7191 	bzero(zbuf, 6);
7192 
7193 	ret  = dhd_wl_ioctl_cmd(dhd, WLC_GET_BSSID, (char *)&bssid,
7194 		ETHER_ADDR_LEN, FALSE, ifidx);
7195 	/* XXX:AS!!! res can be: -17(BCME_NOTASSOCIATED),-22(BCME_NORESOURCE), and 0(OK)
7196 	  OK - doesn't mean associated yet, the returned bssid
7197 	  still needs to be checked for non zero array
7198 	*/
7199 	DHD_TRACE((" %s WLC_GET_BSSID ioctl res = %d\n", __FUNCTION__, ret));
7200 
7201 	if (ret == BCME_NOTASSOCIATED) {
7202 		DHD_ERROR(("%s: WLC_GET_BSSID, NOT ASSOCIATED\n", __FUNCTION__));
7203 	}
7204 
7205 	if (retval)
7206 		*retval = ret;
7207 
7208 	if (ret < 0)
7209 		return FALSE;
7210 
7211 	if ((memcmp(bssid, zbuf, ETHER_ADDR_LEN) == 0)) {
7212 		DHD_TRACE(("%s: WLC_GET_BSSID ioctl returned zero bssid\n", __FUNCTION__));
7213 		return FALSE;
7214 	}
7215 	return TRUE;
7216 }
7217 
7218 /* Function to estimate possible DTIM_SKIP value */
7219 #if defined(OEM_ANDROID) && defined(BCMPCIE)
7220 int
dhd_get_suspend_bcn_li_dtim(dhd_pub_t * dhd,int * dtim_period,int * bcn_interval)7221 dhd_get_suspend_bcn_li_dtim(dhd_pub_t *dhd, int *dtim_period, int *bcn_interval)
7222 {
7223 	int bcn_li_dtim = 1; /* deafult no dtim skip setting */
7224 	int ret = -1;
7225 	int allowed_skip_dtim_cnt = 0;
7226 
7227 	if (dhd->disable_dtim_in_suspend) {
7228 		DHD_ERROR(("%s Disable bcn_li_dtim in suspend\n", __FUNCTION__));
7229 		bcn_li_dtim = 0;
7230 		return bcn_li_dtim;
7231 	}
7232 
7233 	/* Check if associated */
7234 	if (dhd_is_associated(dhd, 0, NULL) == FALSE) {
7235 		DHD_TRACE(("%s NOT assoc ret %d\n", __FUNCTION__, ret));
7236 		return bcn_li_dtim;
7237 	}
7238 
7239 	if (dtim_period == NULL || bcn_interval == NULL)
7240 		return bcn_li_dtim;
7241 
7242 	/* read associated AP beacon interval */
7243 	if ((ret = dhd_wl_ioctl_cmd(dhd, WLC_GET_BCNPRD,
7244 		bcn_interval, sizeof(*bcn_interval), FALSE, 0)) < 0) {
7245 		DHD_ERROR(("%s get beacon failed code %d\n", __FUNCTION__, ret));
7246 		return bcn_li_dtim;
7247 	}
7248 
7249 	/* read associated AP dtim setup */
7250 	if ((ret = dhd_wl_ioctl_cmd(dhd, WLC_GET_DTIMPRD,
7251 		dtim_period, sizeof(*dtim_period), FALSE, 0)) < 0) {
7252 		DHD_ERROR(("%s failed code %d\n", __FUNCTION__, ret));
7253 		return bcn_li_dtim;
7254 	}
7255 
7256 	/* if not assocated just return */
7257 	if (*dtim_period == 0) {
7258 		return bcn_li_dtim;
7259 	}
7260 
7261 	if (dhd->max_dtim_enable) {
7262 		bcn_li_dtim =
7263 			(int) (MAX_DTIM_ALLOWED_INTERVAL / ((*dtim_period) * (*bcn_interval)));
7264 		if (bcn_li_dtim == 0) {
7265 			bcn_li_dtim = 1;
7266 		}
7267 	} else {
7268 		/* attemp to use platform defined dtim skip interval */
7269 		bcn_li_dtim = dhd->suspend_bcn_li_dtim;
7270 
7271 		/* check if sta listen interval fits into AP dtim */
7272 		if (*dtim_period > CUSTOM_LISTEN_INTERVAL) {
7273 			/* AP DTIM to big for our Listen Interval : no dtim skiping */
7274 			bcn_li_dtim = NO_DTIM_SKIP;
7275 			DHD_ERROR(("%s DTIM=%d > Listen=%d : too big ...\n",
7276 				__FUNCTION__, *dtim_period, CUSTOM_LISTEN_INTERVAL));
7277 			return bcn_li_dtim;
7278 		}
7279 
7280 		if (((*dtim_period) * (*bcn_interval) * bcn_li_dtim) > MAX_DTIM_ALLOWED_INTERVAL) {
7281 			allowed_skip_dtim_cnt =
7282 				MAX_DTIM_ALLOWED_INTERVAL / ((*dtim_period) * (*bcn_interval));
7283 			bcn_li_dtim =
7284 				(allowed_skip_dtim_cnt != 0) ? allowed_skip_dtim_cnt : NO_DTIM_SKIP;
7285 		}
7286 
7287 		if ((bcn_li_dtim * (*dtim_period)) > CUSTOM_LISTEN_INTERVAL) {
7288 			/* Round up dtim_skip to fit into STAs Listen Interval */
7289 			bcn_li_dtim = (int)(CUSTOM_LISTEN_INTERVAL / *dtim_period);
7290 			DHD_TRACE(("%s agjust dtim_skip as %d\n", __FUNCTION__, bcn_li_dtim));
7291 		}
7292 	}
7293 
7294 	if (dhd->conf->suspend_bcn_li_dtim >= 0)
7295 		bcn_li_dtim = dhd->conf->suspend_bcn_li_dtim;
7296 	DHD_ERROR(("%s beacon=%d bcn_li_dtim=%d DTIM=%d Listen=%d\n",
7297 		__FUNCTION__, *bcn_interval, bcn_li_dtim, *dtim_period, CUSTOM_LISTEN_INTERVAL));
7298 
7299 	return bcn_li_dtim;
7300 }
7301 #else /* OEM_ANDROID && BCMPCIE */
7302 int
dhd_get_suspend_bcn_li_dtim(dhd_pub_t * dhd)7303 dhd_get_suspend_bcn_li_dtim(dhd_pub_t *dhd)
7304 {
7305 	int bcn_li_dtim = 1; /* deafult no dtim skip setting */
7306 	int ret = -1;
7307 	int dtim_period = 0;
7308 	int ap_beacon = 0;
7309 	int allowed_skip_dtim_cnt = 0;
7310 
7311 	if (dhd->disable_dtim_in_suspend) {
7312 		DHD_ERROR(("%s Disable bcn_li_dtim in suspend\n", __FUNCTION__));
7313 		bcn_li_dtim = 0;
7314 		goto exit;
7315 	}
7316 
7317 	/* Check if associated */
7318 	if (dhd_is_associated(dhd, 0, NULL) == FALSE) {
7319 		DHD_TRACE(("%s NOT assoc ret %d\n", __FUNCTION__, ret));
7320 		goto exit;
7321 	}
7322 
7323 	/* read associated AP beacon interval */
7324 	if ((ret = dhd_wl_ioctl_cmd(dhd, WLC_GET_BCNPRD,
7325 		&ap_beacon, sizeof(ap_beacon), FALSE, 0)) < 0) {
7326 		DHD_ERROR(("%s get beacon failed code %d\n", __FUNCTION__, ret));
7327 		goto exit;
7328 	}
7329 
7330 	/* read associated ap's dtim setup */
7331 	if ((ret = dhd_wl_ioctl_cmd(dhd, WLC_GET_DTIMPRD,
7332 		&dtim_period, sizeof(dtim_period), FALSE, 0)) < 0) {
7333 		DHD_ERROR(("%s failed code %d\n", __FUNCTION__, ret));
7334 		goto exit;
7335 	}
7336 
7337 	/* if not assocated just exit */
7338 	if (dtim_period == 0) {
7339 		goto exit;
7340 	}
7341 
7342 	if (dhd->max_dtim_enable) {
7343 		bcn_li_dtim =
7344 			(int) (MAX_DTIM_ALLOWED_INTERVAL / (ap_beacon * dtim_period));
7345 		if (bcn_li_dtim == 0) {
7346 			bcn_li_dtim = 1;
7347 		}
7348 	} else {
7349 		/* attemp to use platform defined dtim skip interval */
7350 		bcn_li_dtim = dhd->suspend_bcn_li_dtim;
7351 
7352 		/* check if sta listen interval fits into AP dtim */
7353 		if (dtim_period > CUSTOM_LISTEN_INTERVAL) {
7354 			/* AP DTIM to big for our Listen Interval : no dtim skiping */
7355 			bcn_li_dtim = NO_DTIM_SKIP;
7356 			DHD_ERROR(("%s DTIM=%d > Listen=%d : too big ...\n",
7357 				__FUNCTION__, dtim_period, CUSTOM_LISTEN_INTERVAL));
7358 			goto exit;
7359 		}
7360 
7361 		if ((dtim_period * ap_beacon * bcn_li_dtim) > MAX_DTIM_ALLOWED_INTERVAL) {
7362 			allowed_skip_dtim_cnt =
7363 				MAX_DTIM_ALLOWED_INTERVAL / (dtim_period * ap_beacon);
7364 			bcn_li_dtim =
7365 				(allowed_skip_dtim_cnt != 0) ? allowed_skip_dtim_cnt : NO_DTIM_SKIP;
7366 		}
7367 
7368 		if ((bcn_li_dtim * dtim_period) > CUSTOM_LISTEN_INTERVAL) {
7369 			/* Round up dtim_skip to fit into STAs Listen Interval */
7370 			bcn_li_dtim = (int)(CUSTOM_LISTEN_INTERVAL / dtim_period);
7371 			DHD_TRACE(("%s agjust dtim_skip as %d\n", __FUNCTION__, bcn_li_dtim));
7372 		}
7373 	}
7374 
7375 	if (dhd->conf->suspend_bcn_li_dtim >= 0)
7376 		bcn_li_dtim = dhd->conf->suspend_bcn_li_dtim;
7377 	DHD_ERROR(("%s beacon=%d bcn_li_dtim=%d DTIM=%d Listen=%d\n",
7378 		__FUNCTION__, ap_beacon, bcn_li_dtim, dtim_period, CUSTOM_LISTEN_INTERVAL));
7379 
7380 exit:
7381 	return bcn_li_dtim;
7382 }
7383 #endif /* OEM_ANDROID && BCMPCIE */
7384 
7385 #ifdef CONFIG_SILENT_ROAM
7386 int
dhd_sroam_set_mon(dhd_pub_t * dhd,bool set)7387 dhd_sroam_set_mon(dhd_pub_t *dhd, bool set)
7388 {
7389 	int ret = BCME_OK;
7390 	wlc_sroam_t *psroam;
7391 	wlc_sroam_info_t *sroam;
7392 	uint sroamlen = sizeof(*sroam) + SROAM_HDRLEN;
7393 
7394 	/* Check if associated */
7395 	if (dhd_is_associated(dhd, 0, NULL) == FALSE) {
7396 		DHD_TRACE(("%s NOT assoc\n", __FUNCTION__));
7397 		return ret;
7398 	}
7399 
7400 	if (set && (dhd->op_mode &
7401 		(DHD_FLAG_HOSTAP_MODE | DHD_FLAG_P2P_GC_MODE | DHD_FLAG_P2P_GO_MODE))) {
7402 		DHD_INFO((" Failed to set sroam %d, op_mode 0x%04x\n", set, dhd->op_mode));
7403 		return ret;
7404 	}
7405 
7406 	if (!dhd->sroam_turn_on) {
7407 		DHD_INFO((" Failed to set sroam %d, sroam turn %d\n", set, dhd->sroam_turn_on));
7408 		return ret;
7409 	}
7410 	psroam = (wlc_sroam_t *)MALLOCZ(dhd->osh, sroamlen);
7411 	if (!psroam) {
7412 		DHD_ERROR(("%s Fail to malloc buffer\n", __FUNCTION__));
7413 		return BCME_NOMEM;
7414 	}
7415 
7416 	ret = dhd_iovar(dhd, 0, "sroam", NULL, 0, (char *)psroam, sroamlen, FALSE);
7417 	if (ret < 0) {
7418 		DHD_ERROR(("%s Failed to Get sroam %d\n", __FUNCTION__, ret));
7419 		goto done;
7420 	}
7421 
7422 	if (psroam->ver != WLC_SILENT_ROAM_CUR_VER) {
7423 		ret = BCME_VERSION;
7424 		goto done;
7425 	}
7426 
7427 	sroam = (wlc_sroam_info_t *)psroam->data;
7428 	sroam->sroam_on = set;
7429 	DHD_INFO((" Silent roam monitor mode %s\n", set ? "On" : "Off"));
7430 
7431 	ret = dhd_iovar(dhd, 0, "sroam", (char *)psroam, sroamlen, NULL, 0, TRUE);
7432 	if (ret < 0) {
7433 		DHD_ERROR(("%s Failed to Set sroam %d\n", __FUNCTION__, ret));
7434 	}
7435 
7436 done:
7437 	if (psroam) {
7438 	    MFREE(dhd->osh, psroam, sroamlen);
7439 	}
7440 
7441 	return ret;
7442 }
7443 #endif /* CONFIG_SILENT_ROAM */
7444 
7445 /* Check if the mode supports STA MODE */
dhd_support_sta_mode(dhd_pub_t * dhd)7446 bool dhd_support_sta_mode(dhd_pub_t *dhd)
7447 {
7448 
7449 #ifdef  WL_CFG80211
7450 	if (!(dhd->op_mode & DHD_FLAG_STA_MODE))
7451 		return FALSE;
7452 	else
7453 #endif /* WL_CFG80211 */
7454 		return TRUE;
7455 }
7456 
7457 #if defined(KEEP_ALIVE)
dhd_keep_alive_onoff(dhd_pub_t * dhd)7458 int dhd_keep_alive_onoff(dhd_pub_t *dhd)
7459 {
7460 	char				buf[32] = {0};
7461 	const char			*str;
7462 	wl_mkeep_alive_pkt_t	mkeep_alive_pkt = {0, 0, 0, 0, 0, {0}};
7463 	wl_mkeep_alive_pkt_t	*mkeep_alive_pktp;
7464 	int					buf_len;
7465 	int					str_len;
7466 	int res					= -1;
7467 
7468 	if (!dhd_support_sta_mode(dhd))
7469 		return res;
7470 
7471 	DHD_TRACE(("%s execution\n", __FUNCTION__));
7472 
7473 	str = "mkeep_alive";
7474 	str_len = strlen(str);
7475 	strlcpy(buf, str, sizeof(buf));
7476 	mkeep_alive_pktp = (wl_mkeep_alive_pkt_t *) (buf + str_len + 1);
7477 	mkeep_alive_pkt.period_msec = dhd->conf->keep_alive_period;
7478 	buf_len = str_len + 1;
7479 	mkeep_alive_pkt.version = htod16(WL_MKEEP_ALIVE_VERSION);
7480 	mkeep_alive_pkt.length = htod16(WL_MKEEP_ALIVE_FIXED_LEN);
7481 	/* Setup keep alive zero for null packet generation */
7482 	mkeep_alive_pkt.keep_alive_id = 0;
7483 	mkeep_alive_pkt.len_bytes = 0;
7484 	buf_len += WL_MKEEP_ALIVE_FIXED_LEN;
7485 	bzero(mkeep_alive_pkt.data, sizeof(mkeep_alive_pkt.data));
7486 	/* Keep-alive attributes are set in local	variable (mkeep_alive_pkt), and
7487 	 * then memcpy'ed into buffer (mkeep_alive_pktp) since there is no
7488 	 * guarantee that the buffer is properly aligned.
7489 	 */
7490 	memcpy((char *)mkeep_alive_pktp, &mkeep_alive_pkt, WL_MKEEP_ALIVE_FIXED_LEN);
7491 
7492 	res = dhd_wl_ioctl_cmd(dhd, WLC_SET_VAR, buf, buf_len, TRUE, 0);
7493 
7494 	return res;
7495 }
7496 #endif /* defined(KEEP_ALIVE) */
7497 #if defined(OEM_ANDROID)
7498 #define	CSCAN_TLV_TYPE_SSID_IE	'S'
7499 /*
7500  *  SSIDs list parsing from cscan tlv list
7501  */
7502 int
wl_parse_ssid_list_tlv(char ** list_str,wlc_ssid_ext_t * ssid,int max,int * bytes_left)7503 wl_parse_ssid_list_tlv(char** list_str, wlc_ssid_ext_t* ssid, int max, int *bytes_left)
7504 {
7505 	char* str;
7506 	int idx = 0;
7507 	uint8 len;
7508 
7509 	if ((list_str == NULL) || (*list_str == NULL) || (*bytes_left < 0)) {
7510 		DHD_ERROR(("%s error paramters\n", __FUNCTION__));
7511 		return BCME_BADARG;
7512 	}
7513 	str = *list_str;
7514 	while (*bytes_left > 0) {
7515 		if (str[0] != CSCAN_TLV_TYPE_SSID_IE) {
7516 			*list_str = str;
7517 			DHD_TRACE(("nssid=%d left_parse=%d %d\n", idx, *bytes_left, str[0]));
7518 			return idx;
7519 		}
7520 
7521 		if (idx >= max) {
7522 			DHD_ERROR(("%s number of SSIDs more than %d\n", __FUNCTION__, idx));
7523 			return BCME_BADARG;
7524 		}
7525 
7526 		/* Get proper CSCAN_TLV_TYPE_SSID_IE */
7527 		*bytes_left -= 1;
7528 		if (*bytes_left == 0) {
7529 			DHD_ERROR(("%s no length field.\n", __FUNCTION__));
7530 			return BCME_BADARG;
7531 		}
7532 		str += 1;
7533 		ssid[idx].rssi_thresh = 0;
7534 		ssid[idx].flags = 0;
7535 		len = str[0];
7536 		if (len == 0) {
7537 			/* Broadcast SSID */
7538 			ssid[idx].SSID_len = 0;
7539 			memset((char*)ssid[idx].SSID, 0x0, DOT11_MAX_SSID_LEN);
7540 			*bytes_left -= 1;
7541 			str += 1;
7542 
7543 			DHD_TRACE(("BROADCAST SCAN  left=%d\n", *bytes_left));
7544 		} else if (len <= DOT11_MAX_SSID_LEN) {
7545 			/* Get proper SSID size */
7546 			ssid[idx].SSID_len = len;
7547 			*bytes_left -= 1;
7548 			/* Get SSID */
7549 			if (ssid[idx].SSID_len > *bytes_left) {
7550 				DHD_ERROR(("%s out of memory range len=%d but left=%d\n",
7551 				__FUNCTION__, ssid[idx].SSID_len, *bytes_left));
7552 				return BCME_BADARG;
7553 			}
7554 			str += 1;
7555 			memcpy((char*)ssid[idx].SSID, str, ssid[idx].SSID_len);
7556 
7557 			*bytes_left -= ssid[idx].SSID_len;
7558 			str += ssid[idx].SSID_len;
7559 			ssid[idx].hidden = TRUE;
7560 
7561 			DHD_TRACE(("%s :size=%d left=%d\n",
7562 				(char*)ssid[idx].SSID, ssid[idx].SSID_len, *bytes_left));
7563 		} else {
7564 			DHD_ERROR(("### SSID size more than %d\n", str[0]));
7565 			return BCME_BADARG;
7566 		}
7567 		idx++;
7568 	}
7569 
7570 	*list_str = str;
7571 	return idx;
7572 }
7573 
7574 #if defined(WL_WIRELESS_EXT)
7575 /* Android ComboSCAN support */
7576 
7577 /*
7578  *  data parsing from ComboScan tlv list
7579 */
7580 int
wl_iw_parse_data_tlv(char ** list_str,void * dst,int dst_size,const char token,int input_size,int * bytes_left)7581 wl_iw_parse_data_tlv(char** list_str, void *dst, int dst_size, const char token,
7582                      int input_size, int *bytes_left)
7583 {
7584 	char* str;
7585 	uint16 short_temp;
7586 	uint32 int_temp;
7587 
7588 	if ((list_str == NULL) || (*list_str == NULL) ||(bytes_left == NULL) || (*bytes_left < 0)) {
7589 		DHD_ERROR(("%s error paramters\n", __FUNCTION__));
7590 		return -1;
7591 	}
7592 	str = *list_str;
7593 
7594 	/* Clean all dest bytes */
7595 	memset(dst, 0, dst_size);
7596 	if (*bytes_left > 0) {
7597 
7598 		if (str[0] != token) {
7599 			DHD_TRACE(("%s NOT Type=%d get=%d left_parse=%d \n",
7600 				__FUNCTION__, token, str[0], *bytes_left));
7601 			return -1;
7602 		}
7603 
7604 		*bytes_left -= 1;
7605 		str += 1;
7606 
7607 		if (input_size == 1) {
7608 			memcpy(dst, str, input_size);
7609 		}
7610 		else if (input_size == 2) {
7611 			memcpy(dst, (char *)htod16(memcpy(&short_temp, str, input_size)),
7612 				input_size);
7613 		}
7614 		else if (input_size == 4) {
7615 			memcpy(dst, (char *)htod32(memcpy(&int_temp, str, input_size)),
7616 				input_size);
7617 		}
7618 
7619 		*bytes_left -= input_size;
7620 		str += input_size;
7621 		*list_str = str;
7622 		return 1;
7623 	}
7624 	return 1;
7625 }
7626 
7627 /*
7628  *  channel list parsing from cscan tlv list
7629 */
7630 int
wl_iw_parse_channel_list_tlv(char ** list_str,uint16 * channel_list,int channel_num,int * bytes_left)7631 wl_iw_parse_channel_list_tlv(char** list_str, uint16* channel_list,
7632                              int channel_num, int *bytes_left)
7633 {
7634 	char* str;
7635 	int idx = 0;
7636 
7637 	if ((list_str == NULL) || (*list_str == NULL) ||(bytes_left == NULL) || (*bytes_left < 0)) {
7638 		DHD_ERROR(("%s error paramters\n", __FUNCTION__));
7639 		return -1;
7640 	}
7641 	str = *list_str;
7642 
7643 	while (*bytes_left > 0) {
7644 
7645 		if (str[0] != CSCAN_TLV_TYPE_CHANNEL_IE) {
7646 			*list_str = str;
7647 			DHD_TRACE(("End channel=%d left_parse=%d %d\n", idx, *bytes_left, str[0]));
7648 			return idx;
7649 		}
7650 		/* Get proper CSCAN_TLV_TYPE_CHANNEL_IE */
7651 		*bytes_left -= 1;
7652 		str += 1;
7653 
7654 		if (str[0] == 0) {
7655 			/* All channels */
7656 			channel_list[idx] = 0x0;
7657 		}
7658 		else {
7659 			channel_list[idx] = (uint16)str[0];
7660 			DHD_TRACE(("%s channel=%d \n", __FUNCTION__,  channel_list[idx]));
7661 		}
7662 		*bytes_left -= 1;
7663 		str += 1;
7664 
7665 		if (idx++ > 255) {
7666 			DHD_ERROR(("%s Too many channels \n", __FUNCTION__));
7667 			return -1;
7668 		}
7669 	}
7670 
7671 	*list_str = str;
7672 	return idx;
7673 }
7674 
7675 /* Parse a comma-separated list from list_str into ssid array, starting
7676  * at index idx.  Max specifies size of the ssid array.  Parses ssids
7677  * and returns updated idx; if idx >= max not all fit, the excess have
7678  * not been copied.  Returns -1 on empty string, or on ssid too long.
7679  */
7680 int
wl_iw_parse_ssid_list(char ** list_str,wlc_ssid_t * ssid,int idx,int max)7681 wl_iw_parse_ssid_list(char** list_str, wlc_ssid_t* ssid, int idx, int max)
7682 {
7683 	char* str, *ptr;
7684 
7685 	if ((list_str == NULL) || (*list_str == NULL))
7686 		return -1;
7687 
7688 	for (str = *list_str; str != NULL; str = ptr) {
7689 
7690 		/* check for next TAG */
7691 		if (!strncmp(str, GET_CHANNEL, strlen(GET_CHANNEL))) {
7692 			*list_str	 = str + strlen(GET_CHANNEL);
7693 			return idx;
7694 		}
7695 
7696 		if ((ptr = strchr(str, ',')) != NULL) {
7697 			*ptr++ = '\0';
7698 		}
7699 
7700 		if (strlen(str) > DOT11_MAX_SSID_LEN) {
7701 			DHD_ERROR(("ssid <%s> exceeds %d\n", str, DOT11_MAX_SSID_LEN));
7702 			return -1;
7703 		}
7704 
7705 		if (strlen(str) == 0)
7706 			ssid[idx].SSID_len = 0;
7707 
7708 		if (idx < max) {
7709 			bzero(ssid[idx].SSID, sizeof(ssid[idx].SSID));
7710 			strlcpy((char*)ssid[idx].SSID, str, sizeof(ssid[idx].SSID));
7711 			ssid[idx].SSID_len = sizeof(ssid[idx].SSID);
7712 		}
7713 		idx++;
7714 	}
7715 	return idx;
7716 }
7717 
7718 /*
7719  * Parse channel list from iwpriv CSCAN
7720  */
7721 int
wl_iw_parse_channel_list(char ** list_str,uint16 * channel_list,int channel_num)7722 wl_iw_parse_channel_list(char** list_str, uint16* channel_list, int channel_num)
7723 {
7724 	int num;
7725 	int val;
7726 	char* str;
7727 	char* endptr = NULL;
7728 
7729 	if ((list_str == NULL)||(*list_str == NULL))
7730 		return -1;
7731 
7732 	str = *list_str;
7733 	num = 0;
7734 	while (strncmp(str, GET_NPROBE, strlen(GET_NPROBE))) {
7735 		val = (int)strtoul(str, &endptr, 0);
7736 		if (endptr == str) {
7737 			printf("could not parse channel number starting at"
7738 				" substring \"%s\" in list:\n%s\n",
7739 				str, *list_str);
7740 			return -1;
7741 		}
7742 		str = endptr + strspn(endptr, " ,");
7743 
7744 		if (num == channel_num) {
7745 			DHD_ERROR(("too many channels (more than %d) in channel list:\n%s\n",
7746 				channel_num, *list_str));
7747 			return -1;
7748 		}
7749 
7750 		channel_list[num++] = (uint16)val;
7751 	}
7752 	*list_str = str;
7753 	return num;
7754 }
7755 #endif
7756 #endif /* defined(OEM_ANDROID) */
7757 
7758 #if defined(BCM_ROUTER_DHD)
traffic_mgmt_add_dwm_filter(dhd_pub_t * dhd,trf_mgmt_filter_list_t * trf_mgmt_filter_list,int len)7759 static int traffic_mgmt_add_dwm_filter(dhd_pub_t *dhd,
7760 	trf_mgmt_filter_list_t * trf_mgmt_filter_list, int len)
7761 {
7762 	int ret = 0;
7763 	uint32              i;
7764 	trf_mgmt_filter_t   *trf_mgmt_filter;
7765 	uint8               dwm_tbl_entry;
7766 	uint32              dscp = 0;
7767 	uint16              dwm_filter_enabled = 0;
7768 
7769 	/* Check parameter length is adequate */
7770 	if (len < (OFFSETOF(trf_mgmt_filter_list_t, filter) +
7771 		trf_mgmt_filter_list->num_filters * sizeof(trf_mgmt_filter_t))) {
7772 		ret = BCME_BUFTOOSHORT;
7773 		return ret;
7774 	}
7775 
7776 	bzero(&dhd->dhd_tm_dwm_tbl, sizeof(dhd_trf_mgmt_dwm_tbl_t));
7777 
7778 	for (i = 0; i < trf_mgmt_filter_list->num_filters; i++) {
7779 		trf_mgmt_filter = &trf_mgmt_filter_list->filter[i];
7780 
7781 		dwm_filter_enabled = (trf_mgmt_filter->flags & TRF_FILTER_DWM);
7782 
7783 		if (dwm_filter_enabled) {
7784 			dscp = trf_mgmt_filter->dscp;
7785 			if (dscp >= DHD_DWM_TBL_SIZE) {
7786 				ret = BCME_BADARG;
7787 			return ret;
7788 			}
7789 		}
7790 
7791 		dhd->dhd_tm_dwm_tbl.dhd_dwm_enabled = 1;
7792 		/* set WMM AC bits */
7793 		dwm_tbl_entry = (uint8) trf_mgmt_filter->priority;
7794 		DHD_TRF_MGMT_DWM_SET_FILTER(dwm_tbl_entry);
7795 
7796 		/* set favored bits */
7797 		if (trf_mgmt_filter->flags & TRF_FILTER_FAVORED)
7798 			DHD_TRF_MGMT_DWM_SET_FAVORED(dwm_tbl_entry);
7799 
7800 		dhd->dhd_tm_dwm_tbl.dhd_dwm_tbl[dscp] =  dwm_tbl_entry;
7801 	}
7802 	return ret;
7803 }
7804 #endif /* BCM_ROUTER_DHD */
7805 
7806 /* Given filename and download type,  returns a buffer pointer and length
7807 * for download to f/w. Type can be FW or NVRAM.
7808 *
7809 */
dhd_get_download_buffer(dhd_pub_t * dhd,char * file_path,download_type_t component,char ** buffer,int * length)7810 int dhd_get_download_buffer(dhd_pub_t	*dhd, char *file_path, download_type_t component,
7811 	char ** buffer, int *length)
7812 
7813 {
7814 	int ret = BCME_ERROR;
7815 	int len = 0;
7816 	int file_len;
7817 	void *image = NULL;
7818 	uint8 *buf = NULL;
7819 
7820 	/* Point to cache if available. */
7821 #ifdef CACHE_FW_IMAGES
7822 	if (component == FW) {
7823 		if (dhd->cached_fw_length) {
7824 			len = dhd->cached_fw_length;
7825 			buf = dhd->cached_fw;
7826 		}
7827 	} else if (component == NVRAM) {
7828 		if (dhd->cached_nvram_length) {
7829 			len = dhd->cached_nvram_length;
7830 			buf = dhd->cached_nvram;
7831 		}
7832 	} else if (component == CLM_BLOB) {
7833 		if (dhd->cached_clm_length) {
7834 			len = dhd->cached_clm_length;
7835 			buf = dhd->cached_clm;
7836 		}
7837 	} else if (component == TXCAP_BLOB) {
7838 		if (dhd->cached_txcap_length) {
7839 			len = dhd->cached_txcap_length;
7840 			buf = dhd->cached_txcap;
7841 		}
7842 	} else {
7843 		DHD_ERROR(("%s: Invalid component arg %d\n",
7844 			__FUNCTION__, component));
7845 		ret = BCME_BADARG;
7846 		return ret;
7847 	}
7848 #endif /* CACHE_FW_IMAGES */
7849 	/* No Valid cache found on this call */
7850 	if (!len) {
7851 		file_len = *length;
7852 		*length = 0;
7853 
7854 		if (file_path) {
7855 			image = dhd_os_open_image1(dhd, file_path);
7856 			if (image == NULL) {
7857 				printf("%s: Open image file failed %s\n", __FUNCTION__, file_path);
7858 				goto err;
7859 			}
7860 		}
7861 
7862 		buf = MALLOCZ(dhd->osh, file_len);
7863 		if (buf == NULL) {
7864 			DHD_ERROR(("%s: Failed to allocate memory %d bytes\n",
7865 				__FUNCTION__, file_len));
7866 			goto err;
7867 		}
7868 
7869 		/* Download image */
7870 #if defined(BCMEMBEDIMAGE) && defined(DHD_EFI)
7871 		if (!image) {
7872 			memcpy(buf, nvram_arr, sizeof(nvram_arr));
7873 			len = sizeof(nvram_arr);
7874 		} else {
7875 			len = dhd_os_get_image_block((char *)buf, file_len, image);
7876 			if ((len <= 0 || len > file_len)) {
7877 				MFREE(dhd->osh, buf, file_len);
7878 				goto err;
7879 			}
7880 		}
7881 #else
7882 		len = dhd_os_get_image_block((char *)buf, file_len, image);
7883 		if ((len <= 0 || len > file_len)) {
7884 			MFREE(dhd->osh, buf, file_len);
7885 			goto err;
7886 		}
7887 #endif /* DHD_EFI */
7888 	}
7889 
7890 	ret = BCME_OK;
7891 	*length = len;
7892 	*buffer = (char *)buf;
7893 
7894 	/* Cache if first call. */
7895 #ifdef CACHE_FW_IMAGES
7896 	if (component == FW) {
7897 		if (!dhd->cached_fw_length) {
7898 			dhd->cached_fw = buf;
7899 			dhd->cached_fw_length = len;
7900 		}
7901 	} else if (component == NVRAM) {
7902 		if (!dhd->cached_nvram_length) {
7903 			dhd->cached_nvram = buf;
7904 			dhd->cached_nvram_length = len;
7905 		}
7906 	} else if (component == CLM_BLOB) {
7907 		if (!dhd->cached_clm_length) {
7908 			dhd->cached_clm = buf;
7909 			dhd->cached_clm_length = len;
7910 		}
7911 	} else if (component == TXCAP_BLOB) {
7912 		if (!dhd->cached_txcap_length) {
7913 			dhd->cached_txcap = buf;
7914 			dhd->cached_txcap_length = len;
7915 		}
7916 	}
7917 #endif /* CACHE_FW_IMAGES */
7918 
7919 err:
7920 	if (image)
7921 		dhd_os_close_image1(dhd, image);
7922 
7923 	return ret;
7924 }
7925 
7926 int
dhd_download_2_dongle(dhd_pub_t * dhd,char * iovar,uint16 flag,uint16 dload_type,unsigned char * dload_buf,int len)7927 dhd_download_2_dongle(dhd_pub_t	*dhd, char *iovar, uint16 flag, uint16 dload_type,
7928 	unsigned char *dload_buf, int len)
7929 {
7930 	struct wl_dload_data *dload_ptr = (struct wl_dload_data *)dload_buf;
7931 	int err = 0;
7932 	int dload_data_offset;
7933 	static char iovar_buf[WLC_IOCTL_MEDLEN];
7934 	int iovar_len;
7935 
7936 	memset(iovar_buf, 0, sizeof(iovar_buf));
7937 
7938 	dload_data_offset = OFFSETOF(wl_dload_data_t, data);
7939 	dload_ptr->flag = (DLOAD_HANDLER_VER << DLOAD_FLAG_VER_SHIFT) | flag;
7940 	dload_ptr->dload_type = dload_type;
7941 	dload_ptr->len = htod32(len - dload_data_offset);
7942 	dload_ptr->crc = 0;
7943 	len = ROUNDUP(len, 8);
7944 
7945 	iovar_len = bcm_mkiovar(iovar, (char *)dload_buf,
7946 		(uint)len, iovar_buf, sizeof(iovar_buf));
7947 	if (iovar_len == 0) {
7948 		DHD_ERROR(("%s: insufficient buffer space passed to bcm_mkiovar for '%s' \n",
7949 		           __FUNCTION__, iovar));
7950 		return BCME_BUFTOOSHORT;
7951 	}
7952 
7953 	err = dhd_wl_ioctl_cmd(dhd, WLC_SET_VAR, iovar_buf,
7954 			iovar_len, IOV_SET, 0);
7955 
7956 	return err;
7957 }
7958 
7959 int
dhd_download_blob(dhd_pub_t * dhd,unsigned char * buf,uint32 len,char * iovar)7960 dhd_download_blob(dhd_pub_t *dhd, unsigned char *buf,
7961 		uint32 len, char *iovar)
7962 
7963 {
7964 	int chunk_len;
7965 #if !defined(LINUX) && !defined(linux)
7966 	int cumulative_len = 0;
7967 #endif /* !LINUX && !linux */
7968 	int size2alloc;
7969 	unsigned char *new_buf;
7970 	int err = 0, data_offset;
7971 	uint16 dl_flag = DL_BEGIN;
7972 
7973 	data_offset = OFFSETOF(wl_dload_data_t, data);
7974 	size2alloc = data_offset + MAX_CHUNK_LEN;
7975 	size2alloc = ROUNDUP(size2alloc, 8);
7976 
7977 	if ((new_buf = (unsigned char *)MALLOCZ(dhd->osh, size2alloc)) != NULL) {
7978 		do {
7979 #if !defined(LINUX) && !defined(linux)
7980 			if (len >= MAX_CHUNK_LEN)
7981 				chunk_len = MAX_CHUNK_LEN;
7982 			else
7983 				chunk_len = len;
7984 
7985 			memcpy(new_buf + data_offset, buf + cumulative_len, chunk_len);
7986 			cumulative_len += chunk_len;
7987 #else
7988 			chunk_len = dhd_os_get_image_block((char *)(new_buf + data_offset),
7989 				MAX_CHUNK_LEN, buf);
7990 			if (chunk_len < 0) {
7991 				DHD_ERROR(("%s: dhd_os_get_image_block failed (%d)\n",
7992 					__FUNCTION__, chunk_len));
7993 				err = BCME_ERROR;
7994 				goto exit;
7995 			}
7996 #endif /* !LINUX && !linux */
7997 			if (len - chunk_len == 0)
7998 				dl_flag |= DL_END;
7999 
8000 			err = dhd_download_2_dongle(dhd, iovar, dl_flag, DL_TYPE_CLM,
8001 				new_buf, data_offset + chunk_len);
8002 
8003 			dl_flag &= ~DL_BEGIN;
8004 
8005 			len = len - chunk_len;
8006 		} while ((len > 0) && (err == 0));
8007 #if !defined(LINUX) && !defined(linux)
8008 		MFREE(dhd->osh, new_buf, size2alloc);
8009 #endif /* !LINUX && !linux */
8010 	} else {
8011 		err = BCME_NOMEM;
8012 	}
8013 #if defined(LINUX) || defined(linux)
8014 exit:
8015 	if (new_buf) {
8016 		MFREE(dhd->osh, new_buf, size2alloc);
8017 	}
8018 #endif /* LINUX || linux */
8019 	return err;
8020 }
8021 
8022 #if defined(CACHE_FW_IMAGES)
8023 int
dhd_download_blob_cached(dhd_pub_t * dhd,char * file_path,uint32 len,char * iovar)8024 dhd_download_blob_cached(dhd_pub_t *dhd, char *file_path,
8025 	uint32 len, char *iovar)
8026 {
8027 	int ret = BCME_ERROR;
8028 	uint chunk_len, size2alloc, data_offset, file_offset;
8029 	unsigned char *pay_load, *dnld_buf;
8030 	char *memblock;
8031 	uint16 dl_flag = DL_BEGIN;
8032 	download_type_t dl_type;
8033 
8034 	data_offset = OFFSETOF(wl_dload_data_t, data);
8035 	size2alloc = data_offset + MAX_CHUNK_LEN;
8036 	size2alloc = ROUNDUP(size2alloc, 8);
8037 	file_offset = 0;
8038 
8039 	if ((dnld_buf = MALLOCZ(dhd->osh, size2alloc)) == NULL) {
8040 		ret = BCME_NOMEM;
8041 		goto exit;
8042 	}
8043 	pay_load = (dnld_buf + data_offset);
8044 
8045 	if (!memcmp("clmload", iovar, strlen("clmload"))) {
8046 		dl_type = CLM_BLOB;
8047 	} else if (!memcmp("txcapload", iovar, strlen("txcapload"))) {
8048 		dl_type = TXCAP_BLOB;
8049 	} else {
8050 		DHD_ERROR(("%s Invalid iovar :%s \n", __FUNCTION__, iovar));
8051 		ret = BCME_BADARG;
8052 		goto exit;
8053 	}
8054 
8055 	ret = dhd_get_download_buffer(dhd, file_path, dl_type, &memblock, (int *)&len);
8056 	if (ret != BCME_OK) {
8057 		DHD_ERROR(("%s: error getting buffer for %s, %s \n", __FUNCTION__,
8058 			file_path, bcmerrorstr(ret)));
8059 		goto exit;
8060 	}
8061 
8062 	do {
8063 		chunk_len = MIN(len, MAX_CHUNK_LEN);
8064 		memcpy(pay_load, memblock + file_offset, chunk_len);
8065 		if (len - chunk_len == 0) {
8066 			dl_flag |= DL_END;
8067 		}
8068 
8069 		ret = dhd_download_2_dongle(dhd, iovar, dl_flag, DL_TYPE_CLM,
8070 			dnld_buf, data_offset + chunk_len);
8071 
8072 		dl_flag &= ~DL_BEGIN;
8073 		len = len - chunk_len;
8074 		file_offset += chunk_len;
8075 	} while ((len > 0) && (ret == 0));
8076 
8077 exit:
8078 	if (dnld_buf) {
8079 		MFREE(dhd->osh, dnld_buf, size2alloc);
8080 	}
8081 
8082 	return ret;
8083 }
8084 
8085 int
dhd_apply_default_txcap(dhd_pub_t * dhd,char * path)8086 dhd_apply_default_txcap(dhd_pub_t  *dhd, char *path)
8087 {
8088 	int ret = BCME_ERROR;
8089 	ret = dhd_download_blob_cached(dhd, path, MAX_TXCAP_BUF_SIZE, "txcapload");
8090 	if (ret) {
8091 		DHD_ERROR(("%s: error downloading blob: %s \n", __FUNCTION__, bcmerrorstr(ret)));
8092 	}
8093 	return ret;
8094 }
8095 
8096 int
dhd_apply_default_clm(dhd_pub_t * dhd,char * clm_path)8097 dhd_apply_default_clm(dhd_pub_t *dhd, char *clm_path)
8098 {
8099 	char *clm_blob_path;
8100 	int len;
8101 	unsigned char *imgbuf = NULL;
8102 	int err = BCME_OK;
8103 	char iovbuf[WLC_IOCTL_SMLEN];
8104 	wl_country_t *cspec;
8105 
8106 	if (clm_path[0] != '\0') {
8107 		if (strlen(clm_path) > MOD_PARAM_PATHLEN) {
8108 			DHD_ERROR(("clm path exceeds max len\n"));
8109 			return BCME_ERROR;
8110 		}
8111 		clm_blob_path = clm_path;
8112 		DHD_TRACE(("clm path from module param:%s\n", clm_path));
8113 	} else {
8114 		clm_blob_path = VENDOR_PATH CONFIG_BCMDHD_CLM_PATH;
8115 	}
8116 
8117 	/* If CLM blob file is found on the filesystem, download the file.
8118 	* After CLM file download or If the blob file is not present,
8119 	* validate the country code before proceeding with the initialization.
8120 	* If country code is not valid, fail the initialization.
8121 	*/
8122 
8123 	imgbuf = dhd_os_open_image((char *)clm_blob_path);
8124 	if (imgbuf == NULL) {
8125 		goto exit;
8126 	}
8127 
8128 	len = dhd_os_get_image_size(imgbuf);
8129 
8130 	if ((len > 0) && (len < MAX_CLM_BUF_SIZE) && imgbuf) {
8131 		len = bcm_mkiovar("country", NULL, 0, iovbuf, sizeof(iovbuf));
8132 		if (len == 0) {
8133 			err = BCME_BUFTOOSHORT;
8134 			goto exit;
8135 		}
8136 		err = dhd_wl_ioctl_cmd(dhd, WLC_GET_VAR, iovbuf, sizeof(iovbuf), FALSE, 0);
8137 		if (err) {
8138 			DHD_ERROR(("%s: country code get failed\n", __FUNCTION__));
8139 			goto exit;
8140 		}
8141 
8142 		cspec = (wl_country_t *)iovbuf;
8143 		if ((strncmp(cspec->ccode, WL_CCODE_NULL_COUNTRY, WLC_CNTRY_BUF_SZ)) != 0) {
8144 			DHD_ERROR(("%s: CLM already exist in F/W, "
8145 				"new CLM data will be added to the end of existing CLM data!\n",
8146 				__FUNCTION__));
8147 		}
8148 
8149 		/* Found blob file. Download the file */
8150 		DHD_ERROR(("clm file download from %s \n", clm_blob_path));
8151 		if (imgbuf) {
8152 			dhd_os_close_image(imgbuf);
8153 			imgbuf = NULL;
8154 		}
8155 		err = dhd_download_blob_cached(dhd, clm_blob_path, MAX_CLM_BUF_SIZE, "clmload");
8156 		if (err) {
8157 			DHD_ERROR(("%s: CLM download failed err=%d\n", __FUNCTION__, err));
8158 			if (!dhd_bus_skip_clm(dhd)) {
8159 				/* Retrieve clmload_status and print */
8160 				len = bcm_mkiovar("clmload_status", NULL, 0, iovbuf,
8161 						sizeof(iovbuf));
8162 				if (len == 0) {
8163 					err = BCME_BUFTOOSHORT;
8164 					goto exit;
8165 				}
8166 				err = dhd_wl_ioctl_cmd(dhd, WLC_GET_VAR, iovbuf,
8167 					sizeof(iovbuf), FALSE, 0);
8168 				if (err) {
8169 					DHD_ERROR(("%s: clmload_status get failed err=%d \n",
8170 						__FUNCTION__, err));
8171 				} else {
8172 					DHD_ERROR(("%s: clmload_status: %d \n",
8173 						__FUNCTION__, *((int *)iovbuf)));
8174 					if (*((int *)iovbuf) == CHIPID_MISMATCH) {
8175 						DHD_ERROR(("Chip ID mismatch error \n"));
8176 					}
8177 				}
8178 				err = BCME_ERROR;
8179 				goto exit;
8180 			}
8181 		} else {
8182 			DHD_INFO(("%s: CLM download succeeded \n", __FUNCTION__));
8183 		}
8184 	} else {
8185 		DHD_INFO(("Skipping the clm download. len:%d memblk:%p \n", len, imgbuf));
8186 #ifdef DHD_USE_CLMINFO_PARSER
8187 		err = BCME_ERROR;
8188 		goto exit;
8189 #endif /* DHD_USE_CLMINFO_PARSER */
8190 	}
8191 
8192 	/* Verify country code */
8193 	len = bcm_mkiovar("country", NULL, 0, iovbuf, sizeof(iovbuf));
8194 	if (len == 0) {
8195 		err = BCME_BUFTOOSHORT;
8196 		goto exit;
8197 	}
8198 	err = dhd_wl_ioctl_cmd(dhd, WLC_GET_VAR, iovbuf, sizeof(iovbuf), FALSE, 0);
8199 	if (err) {
8200 		DHD_ERROR(("%s: country code get failed\n", __FUNCTION__));
8201 		goto exit;
8202 	}
8203 
8204 	cspec = (wl_country_t *)iovbuf;
8205 	if ((strncmp(cspec->ccode, WL_CCODE_NULL_COUNTRY, WLC_CNTRY_BUF_SZ)) == 0) {
8206 		/* Country code not initialized or CLM download not proper */
8207 		DHD_ERROR(("country code not initialized\n"));
8208 		err = BCME_ERROR;
8209 	}
8210 exit:
8211 
8212 	if (imgbuf) {
8213 		dhd_os_close_image(imgbuf);
8214 	}
8215 
8216 	return err;
8217 }
8218 #else
8219 
8220 int
dhd_apply_default_txcap(dhd_pub_t * dhd,char * path)8221 dhd_apply_default_txcap(dhd_pub_t  *dhd, char *path)
8222 {
8223 	return 0;
8224 }
8225 
8226 int
dhd_check_current_clm_data(dhd_pub_t * dhd)8227 dhd_check_current_clm_data(dhd_pub_t *dhd)
8228 {
8229 	char iovbuf[WLC_IOCTL_SMLEN];
8230 	wl_country_t *cspec;
8231 	int err = BCME_OK;
8232 
8233 	memset(iovbuf, 0, sizeof(iovbuf));
8234 	err = bcm_mkiovar("country", NULL, 0, iovbuf, sizeof(iovbuf));
8235 	if (err == 0) {
8236 		err = BCME_BUFTOOSHORT;
8237 		DHD_ERROR(("%s: bcm_mkiovar failed.", __FUNCTION__));
8238 		return err;
8239 	}
8240 	err = dhd_wl_ioctl_cmd(dhd, WLC_GET_VAR, iovbuf, sizeof(iovbuf), FALSE, 0);
8241 	if (err) {
8242 		DHD_ERROR(("%s: country code get failed\n", __FUNCTION__));
8243 		return err;
8244 	}
8245 	cspec = (wl_country_t *)iovbuf;
8246 	if ((strncmp(cspec->ccode, WL_CCODE_NULL_COUNTRY, WLC_CNTRY_BUF_SZ)) == 0) {
8247 		DHD_ERROR(("%s: ----- This FW is not included CLM data -----\n",
8248 			__FUNCTION__));
8249 		return FALSE;
8250 	}
8251 	DHD_ERROR(("%s: ----- This FW is included CLM data -----\n",
8252 		__FUNCTION__));
8253 	return TRUE;
8254 }
8255 
8256 int
dhd_apply_default_clm(dhd_pub_t * dhd,char * clm_path)8257 dhd_apply_default_clm(dhd_pub_t *dhd, char *clm_path)
8258 {
8259 	char *clm_blob_path;
8260 	int len;
8261 	char *memblock = NULL;
8262 	int err = BCME_OK;
8263 	char iovbuf[WLC_IOCTL_SMLEN];
8264 	int status = FALSE;
8265 
8266 	if (clm_path && clm_path[0] != '\0') {
8267 		if (strlen(clm_path) > MOD_PARAM_PATHLEN) {
8268 			DHD_ERROR(("clm path exceeds max len\n"));
8269 			return BCME_ERROR;
8270 		}
8271 		clm_blob_path = clm_path;
8272 		DHD_TRACE(("clm path from module param:%s\n", clm_path));
8273 	} else {
8274 		clm_blob_path = VENDOR_PATH CONFIG_BCMDHD_CLM_PATH;
8275 	}
8276 
8277 	/* If CLM blob file is found on the filesystem, download the file.
8278 	 * After CLM file download or If the blob file is not present,
8279 	 * validate the country code before proceeding with the initialization.
8280 	 * If country code is not valid, fail the initialization.
8281 	 */
8282 #if !defined(LINUX) && !defined(linux)
8283 	len = MAX_CLM_BUF_SIZE;
8284 	dhd_get_download_buffer(dhd, clm_blob_path, CLM_BLOB, &memblock, &len);
8285 #else
8286 	memblock = dhd_os_open_image1(dhd, (char *)clm_blob_path);
8287 	if (memblock == NULL) {
8288 		printf("%s: Ignore clm file %s\n", __FUNCTION__, clm_path);
8289 #if defined(DHD_BLOB_EXISTENCE_CHECK)
8290 		if (dhd->is_blob) {
8291 			err = BCME_ERROR;
8292 		} else {
8293 			status = dhd_check_current_clm_data(dhd);
8294 			if (status == TRUE) {
8295 				err = BCME_OK;
8296 			} else {
8297 				err = status;
8298 			}
8299 		}
8300 #endif /* DHD_BLOB_EXISTENCE_CHECK */
8301 		goto exit;
8302 	}
8303 
8304 	len = dhd_os_get_image_size(memblock);
8305 #endif /* !LINUX && !linux */
8306 
8307 	if ((len > 0) && (len < MAX_CLM_BUF_SIZE) && memblock) {
8308 		status = dhd_check_current_clm_data(dhd);
8309 		if (status == TRUE) {
8310 #if defined(DHD_BLOB_EXISTENCE_CHECK)
8311 			if (dhd->op_mode != DHD_FLAG_MFG_MODE) {
8312 				if (dhd->is_blob) {
8313 					err = BCME_ERROR;
8314 				}
8315 				goto exit;
8316 			}
8317 #else
8318 			DHD_ERROR(("%s: CLM already exist in F/W, "
8319 				"new CLM data will be added to the end of existing CLM data!\n",
8320 				__FUNCTION__));
8321 #endif /* DHD_BLOB_EXISTENCE_CHECK */
8322 		} else if (status != FALSE) {
8323 			err = status;
8324 			goto exit;
8325 		}
8326 
8327 		/* Found blob file. Download the file */
8328 		DHD_TRACE(("clm file download from %s \n", clm_blob_path));
8329 		err = dhd_download_blob(dhd, (unsigned char*)memblock, len, "clmload");
8330 		if (err) {
8331 			DHD_ERROR(("%s: CLM download failed err=%d\n", __FUNCTION__, err));
8332 			/* Retrieve clmload_status and print */
8333 			memset(iovbuf, 0, sizeof(iovbuf));
8334 			len = bcm_mkiovar("clmload_status", NULL, 0, iovbuf, sizeof(iovbuf));
8335 			if (len == 0) {
8336 				err = BCME_BUFTOOSHORT;
8337 				goto exit;
8338 			}
8339 			err = dhd_wl_ioctl_cmd(dhd, WLC_GET_VAR, iovbuf, sizeof(iovbuf), FALSE, 0);
8340 			if (err) {
8341 				DHD_ERROR(("%s: clmload_status get failed err=%d \n",
8342 					__FUNCTION__, err));
8343 			} else {
8344 				DHD_ERROR(("%s: clmload_status: %d \n",
8345 					__FUNCTION__, *((int *)iovbuf)));
8346 				if (*((int *)iovbuf) == CHIPID_MISMATCH) {
8347 					DHD_ERROR(("Chip ID mismatch error \n"));
8348 				}
8349 			}
8350 			err = BCME_ERROR;
8351 			goto exit;
8352 		} else {
8353 			DHD_INFO(("%s: CLM download succeeded \n", __FUNCTION__));
8354 		}
8355 	} else {
8356 		DHD_INFO(("Skipping the clm download. len:%d memblk:%p \n", len, memblock));
8357 	}
8358 
8359 	/* Verify country code */
8360 	status = dhd_check_current_clm_data(dhd);
8361 
8362 	if (status != TRUE) {
8363 		/* Country code not initialized or CLM download not proper */
8364 		DHD_ERROR(("country code not initialized\n"));
8365 		err = status;
8366 	}
8367 exit:
8368 
8369 	if (memblock) {
8370 #if defined(LINUX) || defined(linux)
8371 		dhd_os_close_image1(dhd, memblock);
8372 #else
8373 		dhd_free_download_buffer(dhd, memblock, MAX_CLM_BUF_SIZE);
8374 #endif /* LINUX || linux */
8375 	}
8376 
8377 	return err;
8378 }
8379 #endif /* defined(CACHE_FW_IMAGES) */
8380 
dhd_free_download_buffer(dhd_pub_t * dhd,void * buffer,int length)8381 void dhd_free_download_buffer(dhd_pub_t	*dhd, void *buffer, int length)
8382 {
8383 #ifdef CACHE_FW_IMAGES
8384 	return;
8385 #endif
8386 	MFREE(dhd->osh, buffer, length);
8387 }
8388 
8389 #ifdef REPORT_FATAL_TIMEOUTS
8390 void
init_dhd_timeouts(dhd_pub_t * pub)8391 init_dhd_timeouts(dhd_pub_t *pub)
8392 {
8393 	pub->timeout_info = MALLOC(pub->osh, sizeof(timeout_info_t));
8394 	if (pub->timeout_info == NULL) {
8395 		DHD_ERROR(("%s: Failed to alloc timeout_info\n", __FUNCTION__));
8396 	} else {
8397 		DHD_INFO(("Initializing dhd_timeouts\n"));
8398 		pub->timeout_info->scan_timer_lock = osl_spin_lock_init(pub->osh);
8399 		pub->timeout_info->join_timer_lock = osl_spin_lock_init(pub->osh);
8400 		pub->timeout_info->bus_timer_lock = osl_spin_lock_init(pub->osh);
8401 		pub->timeout_info->cmd_timer_lock = osl_spin_lock_init(pub->osh);
8402 		pub->timeout_info->scan_timeout_val = SCAN_TIMEOUT_DEFAULT;
8403 		pub->timeout_info->join_timeout_val = JOIN_TIMEOUT_DEFAULT;
8404 		pub->timeout_info->cmd_timeout_val = CMD_TIMEOUT_DEFAULT;
8405 		pub->timeout_info->bus_timeout_val = BUS_TIMEOUT_DEFAULT;
8406 		pub->timeout_info->scan_timer_active = FALSE;
8407 		pub->timeout_info->join_timer_active = FALSE;
8408 		pub->timeout_info->cmd_timer_active = FALSE;
8409 		pub->timeout_info->bus_timer_active = FALSE;
8410 		pub->timeout_info->cmd_join_error = FALSE;
8411 		pub->timeout_info->cmd_request_id = 0;
8412 		OSL_ATOMIC_SET(pub->osh, &pub->set_ssid_rcvd, FALSE);
8413 	}
8414 }
8415 
8416 void
deinit_dhd_timeouts(dhd_pub_t * pub)8417 deinit_dhd_timeouts(dhd_pub_t *pub)
8418 {
8419 	/* stop the join, scan bus, cmd timers
8420 	 * as failing to do so may cause a kernel panic if
8421 	 * an rmmod is done
8422 	 */
8423 	if (!pub->timeout_info) {
8424 		DHD_ERROR(("%s timeout_info pointer is NULL\n", __FUNCTION__));
8425 		ASSERT(0);
8426 		return;
8427 	}
8428 	if (dhd_stop_scan_timer(pub, FALSE, 0)) {
8429 		DHD_ERROR(("%s dhd_stop_scan_timer failed\n", __FUNCTION__));
8430 		ASSERT(0);
8431 	}
8432 	if (dhd_stop_bus_timer(pub)) {
8433 		DHD_ERROR(("%s dhd_stop_bus_timer failed\n", __FUNCTION__));
8434 		ASSERT(0);
8435 	}
8436 	if (dhd_stop_cmd_timer(pub)) {
8437 		DHD_ERROR(("%s dhd_stop_cmd_timer failed\n", __FUNCTION__));
8438 		ASSERT(0);
8439 	}
8440 	if (dhd_stop_join_timer(pub)) {
8441 		DHD_ERROR(("%s dhd_stop_join_timer failed\n", __FUNCTION__));
8442 		ASSERT(0);
8443 	}
8444 
8445 	osl_spin_lock_deinit(pub->osh, pub->timeout_info->scan_timer_lock);
8446 	osl_spin_lock_deinit(pub->osh, pub->timeout_info->join_timer_lock);
8447 	osl_spin_lock_deinit(pub->osh, pub->timeout_info->bus_timer_lock);
8448 	osl_spin_lock_deinit(pub->osh, pub->timeout_info->cmd_timer_lock);
8449 	MFREE(pub->osh, pub->timeout_info, sizeof(timeout_info_t));
8450 }
8451 
8452 static void
dhd_cmd_timeout(void * ctx)8453 dhd_cmd_timeout(void *ctx)
8454 {
8455 	dhd_pub_t *pub = (dhd_pub_t *)ctx;
8456 	unsigned long flags;
8457 
8458 	if (!pub->timeout_info) {
8459 		DHD_ERROR(("DHD: timeout_info NULL\n"));
8460 		ASSERT(0);
8461 		return;
8462 	}
8463 
8464 	DHD_TIMER_LOCK(pub->timeout_info->cmd_timer_lock, flags);
8465 	if (pub->timeout_info && pub->timeout_info->cmd_timer_active) {
8466 		DHD_ERROR(("\nERROR COMMAND TIMEOUT TO:%d\n", pub->timeout_info->cmd_timeout_val));
8467 		DHD_TIMER_UNLOCK(pub->timeout_info->cmd_timer_lock, flags);
8468 #ifdef PCIE_OOB
8469 		/* Assert device_wake so that UART_Rx is available */
8470 		if (dhd_bus_set_device_wake(pub->bus, TRUE)) {
8471 			DHD_ERROR(("%s: dhd_bus_set_device_wake() failed\n", __FUNCTION__));
8472 			ASSERT(0);
8473 		}
8474 #endif /* PCIE_OOB */
8475 		if (dhd_stop_cmd_timer(pub)) {
8476 			DHD_ERROR(("%s: dhd_stop_cmd_timer() failed\n", __FUNCTION__));
8477 			ASSERT(0);
8478 		}
8479 		dhd_wakeup_ioctl_event(pub, IOCTL_RETURN_ON_ERROR);
8480 		if (!dhd_query_bus_erros(pub))
8481 			dhd_send_trap_to_fw_for_timeout(pub, DHD_REASON_COMMAND_TO);
8482 	} else {
8483 		DHD_TIMER_UNLOCK(pub->timeout_info->cmd_timer_lock, flags);
8484 	}
8485 }
8486 
8487 int
dhd_start_cmd_timer(dhd_pub_t * pub)8488 dhd_start_cmd_timer(dhd_pub_t *pub)
8489 {
8490 	int ret = BCME_OK;
8491 	unsigned long flags = 0;
8492 	uint32 cmd_to_ms;
8493 
8494 	if (!pub->timeout_info) {
8495 		DHD_ERROR(("DHD: timeout_info NULL\n"));
8496 		ret = BCME_ERROR;
8497 		ASSERT(0);
8498 		goto exit_null;
8499 	}
8500 	DHD_TIMER_LOCK(pub->timeout_info->cmd_timer_lock, flags);
8501 	cmd_to_ms = pub->timeout_info->cmd_timeout_val;
8502 
8503 	if (pub->timeout_info->cmd_timeout_val == 0) {
8504 		/* Disable Command timer timeout */
8505 		DHD_INFO(("DHD: Command Timeout Disabled\n"));
8506 		goto exit;
8507 	}
8508 	if (pub->timeout_info->cmd_timer_active) {
8509 		DHD_ERROR(("%s:Timer already active\n", __FUNCTION__));
8510 		ret = BCME_ERROR;
8511 		ASSERT(0);
8512 	} else {
8513 		pub->timeout_info->cmd_timer = osl_timer_init(pub->osh,
8514 			"cmd_timer", dhd_cmd_timeout, pub);
8515 		osl_timer_update(pub->osh, pub->timeout_info->cmd_timer,
8516 			cmd_to_ms, 0);
8517 		pub->timeout_info->cmd_timer_active = TRUE;
8518 	}
8519 	if (ret == BCME_OK) {
8520 		DHD_INFO(("%s Cmd Timer started\n", __FUNCTION__));
8521 	}
8522 exit:
8523 	DHD_TIMER_UNLOCK(pub->timeout_info->cmd_timer_lock, flags);
8524 exit_null:
8525 	return ret;
8526 }
8527 
8528 int
dhd_stop_cmd_timer(dhd_pub_t * pub)8529 dhd_stop_cmd_timer(dhd_pub_t *pub)
8530 {
8531 	int ret = BCME_OK;
8532 	unsigned long flags = 0;
8533 
8534 	if (!pub) {
8535 		DHD_ERROR(("DHD: pub NULL\n"));
8536 		ASSERT(0);
8537 		return BCME_ERROR;
8538 	}
8539 
8540 	if (!pub->timeout_info) {
8541 		DHD_ERROR(("DHD: timeout_info NULL\n"));
8542 		ret = BCME_ERROR;
8543 		ASSERT(0);
8544 		goto exit;
8545 	}
8546 	DHD_TIMER_LOCK(pub->timeout_info->cmd_timer_lock, flags);
8547 
8548 	if (pub->timeout_info->cmd_timer_active) {
8549 		osl_timer_del(pub->osh, pub->timeout_info->cmd_timer);
8550 		pub->timeout_info->cmd_timer_active = FALSE;
8551 	}
8552 	else {
8553 		DHD_INFO(("DHD: CMD timer is not active\n"));
8554 	}
8555 	if (ret == BCME_OK) {
8556 		DHD_INFO(("%s Cmd Timer Stopped\n", __FUNCTION__));
8557 	}
8558 	DHD_TIMER_UNLOCK(pub->timeout_info->cmd_timer_lock, flags);
8559 exit:
8560 	return ret;
8561 }
8562 
8563 static int
__dhd_stop_join_timer(dhd_pub_t * pub)8564 __dhd_stop_join_timer(dhd_pub_t *pub)
8565 {
8566 	int ret = BCME_OK;
8567 	if (!pub) {
8568 		DHD_ERROR(("DHD: pub NULL\n"));
8569 		ASSERT(0);
8570 		return BCME_ERROR;
8571 	}
8572 	if (!pub->timeout_info) {
8573 		DHD_ERROR(("DHD: timeout_info NULL\n"));
8574 		ASSERT(0);
8575 		return BCME_ERROR;
8576 	}
8577 
8578 	if (pub->timeout_info->join_timer_active) {
8579 		osl_timer_del(pub->osh, pub->timeout_info->join_timer);
8580 		pub->timeout_info->join_timer_active = FALSE;
8581 		DHD_INFO(("%s join timer stopped\n", __FUNCTION__));
8582 	} else {
8583 		DHD_INFO(("%s join timer is not active\n", __FUNCTION__));
8584 	}
8585 
8586 	return ret;
8587 }
8588 
8589 static void
dhd_join_timeout(void * ctx)8590 dhd_join_timeout(void *ctx)
8591 {
8592 	dhd_pub_t *pub = (dhd_pub_t *)ctx;
8593 	unsigned long flags;
8594 
8595 	if (!pub->timeout_info) {
8596 		DHD_ERROR(("%s DHD: timeout_info NULL\n", __FUNCTION__));
8597 		ASSERT(0);
8598 		return;
8599 	}
8600 
8601 	DHD_TIMER_LOCK(pub->timeout_info->join_timer_lock, flags);
8602 	if (pub->timeout_info->join_timer_active) {
8603 		if (__dhd_stop_join_timer(pub)) {
8604 			DHD_ERROR(("%s: __dhd_stop_join_timer() failed\n", __FUNCTION__));
8605 			ASSERT(0);
8606 		}
8607 		if (pub->timeout_info->cmd_join_error) {
8608 			DHD_ERROR(("\n%s ERROR JOIN TIMEOUT TO:%d:0x%x\n", __FUNCTION__,
8609 				pub->timeout_info->join_timeout_val,
8610 				pub->timeout_info->cmd_join_error));
8611 			if (!dhd_query_bus_erros(pub)) {
8612 				dhd_send_trap_to_fw_for_timeout(pub, DHD_REASON_JOIN_TO);
8613 			}
8614 			pub->timeout_info->cmd_join_error = 0;
8615 		}
8616 	}
8617 	DHD_TIMER_UNLOCK(pub->timeout_info->join_timer_lock, flags);
8618 }
8619 
8620 int
dhd_start_join_timer(dhd_pub_t * pub)8621 dhd_start_join_timer(dhd_pub_t *pub)
8622 {
8623 	int ret = BCME_OK;
8624 	unsigned long flags = 0;
8625 	uint32 join_to_ms;
8626 
8627 	if (!pub->timeout_info) {
8628 		DHD_ERROR(("%s DHD: timeout_info NULL\n", __FUNCTION__));
8629 		ret = BCME_ERROR;
8630 		ASSERT(0);
8631 		goto exit;
8632 	}
8633 
8634 	join_to_ms = pub->timeout_info->join_timeout_val;
8635 	DHD_TIMER_LOCK(pub->timeout_info->join_timer_lock, flags);
8636 	if (pub->timeout_info->join_timer_active) {
8637 		DHD_ERROR(("%s: stopping active timer\n", __FUNCTION__));
8638 		__dhd_stop_join_timer(pub);
8639 	}
8640 	if (pub->timeout_info->join_timeout_val == 0) {
8641 		/* Disable Join timer timeout */
8642 		DHD_INFO(("%s DHD: join timeout disabled\n", __FUNCTION__));
8643 	} else {
8644 		pub->timeout_info->join_timer = osl_timer_init(pub->osh,
8645 			"join_timer", dhd_join_timeout, pub);
8646 		osl_timer_update(pub->osh, pub->timeout_info->join_timer, join_to_ms, 0);
8647 		pub->timeout_info->join_timer_active = TRUE;
8648 		pub->timeout_info->cmd_join_error = 0;
8649 		dhd_set_join_error(pub, WLC_SSID_MASK);
8650 		if (pub->secure_join) {
8651 			dhd_set_join_error(pub, WLC_WPA_MASK);
8652 		}
8653 		DHD_ERROR(("%s: join timer started 0x%x\n", __FUNCTION__,
8654 			pub->timeout_info->cmd_join_error));
8655 	}
8656 	DHD_TIMER_UNLOCK(pub->timeout_info->join_timer_lock, flags);
8657 exit:
8658 	return ret;
8659 }
8660 
8661 int
dhd_stop_join_timer(dhd_pub_t * pub)8662 dhd_stop_join_timer(dhd_pub_t *pub)
8663 {
8664 	int ret = BCME_OK;
8665 	unsigned long flags;
8666 
8667 	if (!pub) {
8668 		DHD_ERROR(("%s DHD: pub NULL\n", __FUNCTION__));
8669 		ASSERT(0);
8670 		return BCME_ERROR;
8671 	}
8672 
8673 	DHD_TIMER_LOCK(pub->timeout_info->join_timer_lock, flags);
8674 	ret = __dhd_stop_join_timer(pub);
8675 	DHD_TIMER_UNLOCK(pub->timeout_info->join_timer_lock, flags);
8676 	return ret;
8677 }
8678 
8679 static void
dhd_set_join_error(dhd_pub_t * pub,uint32 mask)8680 dhd_set_join_error(dhd_pub_t *pub, uint32 mask)
8681 {
8682 	DHD_INFO(("Setting join Error %d\n", mask));
8683 	if (pub->timeout_info) {
8684 		pub->timeout_info->cmd_join_error |= mask;
8685 	}
8686 }
8687 
8688 void
dhd_clear_join_error(dhd_pub_t * pub,uint32 mask)8689 dhd_clear_join_error(dhd_pub_t *pub, uint32 mask)
8690 {
8691 	unsigned long flags;
8692 
8693 	DHD_INFO(("%s clear join error %d\n", __FUNCTION__, mask));
8694 	if (!(pub->timeout_info)) {
8695 		return;
8696 	}
8697 
8698 	DHD_TIMER_LOCK(pub->timeout_info->join_timer_lock, flags);
8699 	pub->timeout_info->cmd_join_error &= ~mask;
8700 	/* If both WLC_SSID_MASK, WLC_WPA_MASK are received cancel the timer */
8701 	if (!(pub->timeout_info->cmd_join_error)) {
8702 		if (__dhd_stop_join_timer(pub)) {
8703 			DHD_ERROR(("%s: dhd_stop_join_timer failed\n", __FUNCTION__));
8704 			ASSERT(0);
8705 		}
8706 	}
8707 	DHD_TIMER_UNLOCK(pub->timeout_info->join_timer_lock, flags);
8708 }
8709 
8710 static void
dhd_scan_timeout(void * ctx)8711 dhd_scan_timeout(void *ctx)
8712 {
8713 	dhd_pub_t *pub = (dhd_pub_t *)ctx;
8714 	unsigned long flags;
8715 
8716 	if (!pub) {
8717 		DHD_ERROR(("DHD: pub NULL\n"));
8718 		ASSERT(0);
8719 		return;
8720 	}
8721 
8722 	if (pub->timeout_info == NULL) {
8723 		DHD_ERROR(("timeout_info pointer is NULL\n"));
8724 		ASSERT(0);
8725 		return;
8726 	}
8727 	DHD_TIMER_LOCK(pub->timeout_info->scan_timer_lock, flags);
8728 	if (pub->timeout_info->scan_timer_active) {
8729 		DHD_ERROR(("\nERROR SCAN TIMEOUT TO:%d\n", pub->timeout_info->scan_timeout_val));
8730 		DHD_TIMER_UNLOCK(pub->timeout_info->scan_timer_lock, flags);
8731 		dhd_stop_scan_timer(pub, FALSE, 0);
8732 		if (!dhd_query_bus_erros(pub))
8733 			dhd_send_trap_to_fw_for_timeout(pub, DHD_REASON_SCAN_TO);
8734 	} else {
8735 		DHD_TIMER_UNLOCK(pub->timeout_info->scan_timer_lock, flags);
8736 	}
8737 }
8738 
8739 int
dhd_start_scan_timer(dhd_pub_t * pub,bool is_escan)8740 dhd_start_scan_timer(dhd_pub_t *pub, bool is_escan)
8741 {
8742 	int ret = BCME_OK;
8743 	unsigned long flags = 0;
8744 	uint32 scan_to_ms;
8745 
8746 	if (!pub->timeout_info) {
8747 		DHD_ERROR(("DHD: timeout_info NULL\n"));
8748 		ret = BCME_ERROR;
8749 		ASSERT(0);
8750 		goto exit_null;
8751 	}
8752 	DHD_TIMER_LOCK(pub->timeout_info->scan_timer_lock, flags);
8753 	scan_to_ms = pub->timeout_info->scan_timeout_val;
8754 
8755 	if (is_escan) {
8756 		if (pub->timeout_info->escan_aborted &&
8757 				pub->esync_id == pub->timeout_info->abort_syncid) {
8758 			pub->timeout_info->escan_aborted = FALSE;
8759 			DHD_INFO(("%s: escan already aborted, do not start timer \n",
8760 				__FUNCTION__));
8761 			goto exit;
8762 		}
8763 		pub->timeout_info->escan_syncid = pub->esync_id;
8764 	} else {
8765 		pub->timeout_info->escan_syncid = 0;
8766 	}
8767 
8768 	if (pub->timeout_info->scan_timer_active) {
8769 		/* cancel any earlier running timer */
8770 		DHD_INFO(("%s:Timer already active, stopping it.\n", __FUNCTION__));
8771 		osl_timer_del(pub->osh, pub->timeout_info->scan_timer);
8772 		pub->timeout_info->scan_timer_active = FALSE;
8773 	}
8774 
8775 	if (pub->timeout_info->scan_timeout_val == 0) {
8776 		/* Disable Scan timer timeout */
8777 		DHD_INFO(("DHD: Scan Timeout Disabled\n"));
8778 	} else {
8779 		pub->timeout_info->scan_timer = osl_timer_init(pub->osh, "scan_timer",
8780 			dhd_scan_timeout, pub);
8781 		pub->timeout_info->scan_timer_active = TRUE;
8782 		osl_timer_update(pub->osh, pub->timeout_info->scan_timer, scan_to_ms, 0);
8783 		DHD_INFO(("%s Scan Timer started\n", __FUNCTION__));
8784 	}
8785 
8786 exit:
8787 	DHD_TIMER_UNLOCK(pub->timeout_info->scan_timer_lock, flags);
8788 exit_null:
8789 	return ret;
8790 }
8791 
8792 int
dhd_stop_scan_timer(dhd_pub_t * pub,bool is_escan,uint16 sync_id)8793 dhd_stop_scan_timer(dhd_pub_t *pub, bool is_escan, uint16 sync_id)
8794 {
8795 	int ret = BCME_OK;
8796 	unsigned long flags = 0;
8797 
8798 	if (!pub) {
8799 		DHD_ERROR(("DHD: pub NULL\n"));
8800 		ASSERT(0);
8801 		return BCME_ERROR;
8802 	}
8803 
8804 	if (!pub->timeout_info) {
8805 		DHD_ERROR(("DHD: timeout_info NULL\n"));
8806 		ret = BCME_ERROR;
8807 		ASSERT(0);
8808 		goto exit_null;
8809 	}
8810 
8811 	DHD_TIMER_LOCK(pub->timeout_info->scan_timer_lock, flags);
8812 
8813 	if (pub->timeout_info->scan_timer_active) {
8814 		if (is_escan) {
8815 			if (sync_id == pub->timeout_info->escan_syncid) {
8816 				osl_timer_del(pub->osh, pub->timeout_info->scan_timer);
8817 				pub->timeout_info->scan_timer_active = FALSE;
8818 				DHD_INFO(("%s Scan Timer Stopped\n", __FUNCTION__));
8819 			}
8820 		} else {
8821 			osl_timer_del(pub->osh, pub->timeout_info->scan_timer);
8822 			pub->timeout_info->scan_timer_active = FALSE;
8823 			DHD_INFO(("%s Scan Timer Stopped\n", __FUNCTION__));
8824 		}
8825 
8826 	} else {
8827 		DHD_INFO(("DHD: SCAN timer is not active\n"));
8828 	}
8829 
8830 	DHD_TIMER_UNLOCK(pub->timeout_info->scan_timer_lock, flags);
8831 
8832 exit_null:
8833 	return ret;
8834 }
8835 
8836 static void
dhd_bus_timeout(void * ctx)8837 dhd_bus_timeout(void *ctx)
8838 {
8839 	dhd_pub_t *pub = (dhd_pub_t *)ctx;
8840 	unsigned long flags;
8841 
8842 	if (pub->timeout_info == NULL) {
8843 		DHD_ERROR(("timeout_info pointer is NULL\n"));
8844 		ASSERT(0);
8845 		return;
8846 	}
8847 
8848 	DHD_TIMER_LOCK(pub->timeout_info->bus_timer_lock, flags);
8849 	if (pub->timeout_info && pub->timeout_info->bus_timer_active) {
8850 		DHD_ERROR(("\nERROR BUS TIMEOUT TO:%d\n", pub->timeout_info->bus_timeout_val));
8851 		DHD_TIMER_UNLOCK(pub->timeout_info->bus_timer_lock, flags);
8852 #ifdef PCIE_OOB
8853 		/* Assert device_wake so that UART_Rx is available */
8854 		if (dhd_bus_set_device_wake(pub->bus, TRUE)) {
8855 			DHD_ERROR(("%s: dhd_bus_set_device_wake() failed\n", __FUNCTION__));
8856 			ASSERT(0);
8857 		}
8858 #endif /* PCIE_OOB */
8859 		if (dhd_stop_bus_timer(pub)) {
8860 			DHD_ERROR(("%s: dhd_stop_bus_timer() failed\n", __FUNCTION__));
8861 			ASSERT(0);
8862 		}
8863 		if (!dhd_query_bus_erros(pub)) {
8864 			dhd_send_trap_to_fw_for_timeout(pub, DHD_REASON_OQS_TO);
8865 		}
8866 #ifdef BCMPCIE
8867 		dhd_msgbuf_iovar_timeout_dump(pub);
8868 #endif /* BCMPCIE */
8869 	} else {
8870 		DHD_TIMER_UNLOCK(pub->timeout_info->bus_timer_lock, flags);
8871 	}
8872 }
8873 
8874 int
dhd_start_bus_timer(dhd_pub_t * pub)8875 dhd_start_bus_timer(dhd_pub_t *pub)
8876 {
8877 	int ret = BCME_OK;
8878 	unsigned long flags = 0;
8879 	uint32 bus_to_ms;
8880 
8881 	if (!pub->timeout_info) {
8882 		DHD_ERROR(("DHD: timeout_info NULL\n"));
8883 		ret = BCME_ERROR;
8884 		ASSERT(0);
8885 		goto exit_null;
8886 	}
8887 	DHD_TIMER_LOCK(pub->timeout_info->bus_timer_lock, flags);
8888 	bus_to_ms = pub->timeout_info->bus_timeout_val;
8889 
8890 	if (pub->timeout_info->bus_timeout_val == 0) {
8891 		/* Disable Bus timer timeout */
8892 		DHD_INFO(("DHD: Bus Timeout Disabled\n"));
8893 		goto exit;
8894 	}
8895 	if (pub->timeout_info->bus_timer_active) {
8896 		DHD_ERROR(("%s:Timer already active\n", __FUNCTION__));
8897 		ret = BCME_ERROR;
8898 		ASSERT(0);
8899 	} else {
8900 		pub->timeout_info->bus_timer = osl_timer_init(pub->osh,
8901 			"bus_timer", dhd_bus_timeout, pub);
8902 		pub->timeout_info->bus_timer_active = TRUE;
8903 		osl_timer_update(pub->osh, pub->timeout_info->bus_timer, bus_to_ms, 0);
8904 	}
8905 	if (ret == BCME_OK) {
8906 		DHD_INFO(("%s: BUS Timer started\n", __FUNCTION__));
8907 	}
8908 exit:
8909 	DHD_TIMER_UNLOCK(pub->timeout_info->bus_timer_lock, flags);
8910 exit_null:
8911 	return ret;
8912 }
8913 
8914 int
dhd_stop_bus_timer(dhd_pub_t * pub)8915 dhd_stop_bus_timer(dhd_pub_t *pub)
8916 {
8917 	int ret = BCME_OK;
8918 	unsigned long flags;
8919 
8920 	if (!pub) {
8921 		DHD_ERROR(("DHD: pub NULL\n"));
8922 		ASSERT(0);
8923 		return BCME_ERROR;
8924 	}
8925 
8926 	if (!pub->timeout_info) {
8927 		DHD_ERROR(("DHD: timeout_info NULL\n"));
8928 		ret = BCME_ERROR;
8929 		ASSERT(0);
8930 		goto exit;
8931 	}
8932 
8933 	DHD_TIMER_LOCK(pub->timeout_info->bus_timer_lock, flags);
8934 
8935 	if (pub->timeout_info->bus_timer_active) {
8936 		osl_timer_del(pub->osh, pub->timeout_info->bus_timer);
8937 		pub->timeout_info->bus_timer_active = FALSE;
8938 	}
8939 	else {
8940 		DHD_INFO(("DHD: BUS timer is not active\n"));
8941 	}
8942 	if (ret == BCME_OK) {
8943 		DHD_INFO(("%s: Bus Timer Stopped\n", __FUNCTION__));
8944 	}
8945 	DHD_TIMER_UNLOCK(pub->timeout_info->bus_timer_lock, flags);
8946 exit:
8947 	return ret;
8948 }
8949 
8950 int
dhd_set_request_id(dhd_pub_t * pub,uint16 id,uint32 cmd)8951 dhd_set_request_id(dhd_pub_t *pub, uint16 id, uint32 cmd)
8952 {
8953 	DHD_INFO(("%s: id:%d\n", __FUNCTION__, id));
8954 	if (pub->timeout_info) {
8955 		pub->timeout_info->cmd_request_id = id;
8956 		pub->timeout_info->cmd = cmd;
8957 		return BCME_OK;
8958 	} else {
8959 		return BCME_ERROR;
8960 	}
8961 }
8962 
8963 uint16
dhd_get_request_id(dhd_pub_t * pub)8964 dhd_get_request_id(dhd_pub_t *pub)
8965 {
8966 	if (pub->timeout_info) {
8967 		return (pub->timeout_info->cmd_request_id);
8968 	} else {
8969 		return 0;
8970 	}
8971 }
8972 
8973 void
dhd_get_scan_to_val(dhd_pub_t * pub,uint32 * to_val)8974 dhd_get_scan_to_val(dhd_pub_t *pub, uint32 *to_val)
8975 {
8976 	if (pub->timeout_info) {
8977 		*to_val = pub->timeout_info->scan_timeout_val;
8978 	} else {
8979 		*to_val = 0;
8980 	}
8981 }
8982 
8983 void
dhd_set_scan_to_val(dhd_pub_t * pub,uint32 to_val)8984 dhd_set_scan_to_val(dhd_pub_t *pub, uint32 to_val)
8985 {
8986 	if (pub->timeout_info) {
8987 		DHD_INFO(("Setting scan TO val:%d\n", to_val));
8988 		pub->timeout_info->scan_timeout_val = to_val;
8989 	}
8990 }
8991 
8992 void
dhd_get_join_to_val(dhd_pub_t * pub,uint32 * to_val)8993 dhd_get_join_to_val(dhd_pub_t *pub, uint32 *to_val)
8994 {
8995 	if (pub->timeout_info) {
8996 		*to_val = pub->timeout_info->join_timeout_val;
8997 	} else {
8998 		*to_val = 0;
8999 	}
9000 }
9001 
9002 void
dhd_set_join_to_val(dhd_pub_t * pub,uint32 to_val)9003 dhd_set_join_to_val(dhd_pub_t *pub, uint32 to_val)
9004 {
9005 	if (pub->timeout_info) {
9006 		DHD_INFO(("Setting join TO val:%d\n", to_val));
9007 		pub->timeout_info->join_timeout_val = to_val;
9008 	}
9009 }
9010 
9011 void
dhd_get_cmd_to_val(dhd_pub_t * pub,uint32 * to_val)9012 dhd_get_cmd_to_val(dhd_pub_t *pub, uint32 *to_val)
9013 {
9014 	if (pub->timeout_info) {
9015 		*to_val = pub->timeout_info->cmd_timeout_val;
9016 	} else {
9017 		*to_val = 0;
9018 	}
9019 }
9020 
9021 void
dhd_set_cmd_to_val(dhd_pub_t * pub,uint32 to_val)9022 dhd_set_cmd_to_val(dhd_pub_t *pub, uint32 to_val)
9023 {
9024 	if (pub->timeout_info) {
9025 		DHD_INFO(("Setting cmd TO val:%d\n", to_val));
9026 		pub->timeout_info->cmd_timeout_val = to_val;
9027 	}
9028 }
9029 
9030 void
dhd_get_bus_to_val(dhd_pub_t * pub,uint32 * to_val)9031 dhd_get_bus_to_val(dhd_pub_t *pub, uint32 *to_val)
9032 {
9033 	if (pub->timeout_info) {
9034 		*to_val = pub->timeout_info->bus_timeout_val;
9035 	} else {
9036 		*to_val = 0;
9037 	}
9038 }
9039 
9040 void
dhd_set_bus_to_val(dhd_pub_t * pub,uint32 to_val)9041 dhd_set_bus_to_val(dhd_pub_t *pub, uint32 to_val)
9042 {
9043 	if (pub->timeout_info) {
9044 		DHD_INFO(("Setting bus TO val:%d\n", to_val));
9045 		pub->timeout_info->bus_timeout_val = to_val;
9046 	}
9047 }
9048 #endif /* REPORT_FATAL_TIMEOUTS */
9049 
9050 #ifdef SHOW_LOGTRACE
9051 int
dhd_parse_logstrs_file(osl_t * osh,char * raw_fmts,int logstrs_size,dhd_event_log_t * event_log)9052 dhd_parse_logstrs_file(osl_t *osh, char *raw_fmts, int logstrs_size,
9053 		dhd_event_log_t *event_log)
9054 {
9055 	uint32 *lognums = NULL;
9056 	char *logstrs = NULL;
9057 	logstr_trailer_t *trailer = NULL;
9058 	int ram_index = 0;
9059 	char **fmts = NULL;
9060 	int num_fmts = 0;
9061 	bool match_fail = TRUE;
9062 	int32 i = 0;
9063 	uint8 *pfw_id = NULL;
9064 	uint32 fwid = 0;
9065 	void *file = NULL;
9066 	int file_len = 0;
9067 	char fwid_str[FWID_STR_LEN];
9068 	uint32 hdr_logstrs_size = 0;
9069 
9070 	/* Read last three words in the logstrs.bin file */
9071 	trailer = (logstr_trailer_t *) (raw_fmts + logstrs_size -
9072 		sizeof(logstr_trailer_t));
9073 
9074 	if (trailer->log_magic == LOGSTRS_MAGIC) {
9075 		/*
9076 		* logstrs.bin has a header.
9077 		*/
9078 		if (trailer->version == 1) {
9079 			logstr_header_v1_t *hdr_v1 = (logstr_header_v1_t *) (raw_fmts +
9080 					logstrs_size - sizeof(logstr_header_v1_t));
9081 			DHD_INFO(("%s: logstr header version = %u\n",
9082 					__FUNCTION__, hdr_v1->version));
9083 			num_fmts =	hdr_v1->rom_logstrs_offset / sizeof(uint32);
9084 			ram_index = (hdr_v1->ram_lognums_offset -
9085 				hdr_v1->rom_lognums_offset) / sizeof(uint32);
9086 			lognums = (uint32 *) &raw_fmts[hdr_v1->rom_lognums_offset];
9087 			logstrs = (char *)	 &raw_fmts[hdr_v1->rom_logstrs_offset];
9088 			hdr_logstrs_size = hdr_v1->logstrs_size;
9089 		} else if (trailer->version == 2) {
9090 			logstr_header_t *hdr = (logstr_header_t *) (raw_fmts + logstrs_size -
9091 					sizeof(logstr_header_t));
9092 			DHD_INFO(("%s: logstr header version = %u; flags = %x\n",
9093 					__FUNCTION__, hdr->version, hdr->flags));
9094 
9095 			/* For ver. 2 of the header, need to match fwid of
9096 			 *  both logstrs.bin and fw bin
9097 			 */
9098 
9099 			/* read the FWID from fw bin */
9100 			file = dhd_os_open_image1(NULL, st_str_file_path);
9101 			if (!file) {
9102 				DHD_ERROR(("%s: cannot open fw file !\n", __FUNCTION__));
9103 				goto error;
9104 			}
9105 			file_len = dhd_os_get_image_size(file);
9106 			if (file_len <= 0) {
9107 				DHD_ERROR(("%s: bad fw file length !\n", __FUNCTION__));
9108 				goto error;
9109 			}
9110 			/* fwid is at the end of fw bin in string format */
9111 			if (dhd_os_seek_file(file, file_len - (sizeof(fwid_str) - 1)) < 0) {
9112 				DHD_ERROR(("%s: can't seek file \n", __FUNCTION__));
9113 				goto error;
9114 			}
9115 
9116 			memset(fwid_str, 0, sizeof(fwid_str));
9117 			if (dhd_os_get_image_block(fwid_str, sizeof(fwid_str) - 1, file) <= 0) {
9118 				DHD_ERROR(("%s: read fw file failed !\n", __FUNCTION__));
9119 				goto error;
9120 			}
9121 			pfw_id = (uint8 *)bcmstrnstr(fwid_str, sizeof(fwid_str) - 1,
9122 					FWID_STR_1, strlen(FWID_STR_1));
9123 			if (!pfw_id) {
9124 				pfw_id = (uint8 *)bcmstrnstr(fwid_str, sizeof(fwid_str) - 1,
9125 					FWID_STR_2, strlen(FWID_STR_2));
9126 				if (!pfw_id) {
9127 					DHD_ERROR(("%s: could not find id in FW bin!\n",
9128 							__FUNCTION__));
9129 					goto error;
9130 				}
9131 			}
9132 			/* search for the '-' in the fw id str, after which the
9133 			 * actual 4 byte fw id is present
9134 			 */
9135 			while (pfw_id && *pfw_id != '-') {
9136 				++pfw_id;
9137 			}
9138 			++pfw_id;
9139 			fwid = bcm_strtoul((char *)pfw_id, NULL, 16);
9140 
9141 			/* check if fw id in logstrs.bin matches the fw one */
9142 			if (hdr->fw_id != fwid) {
9143 				DHD_ERROR(("%s: logstr id does not match FW!"
9144 					"logstrs_fwid:0x%x, rtecdc_fwid:0x%x\n",
9145 					__FUNCTION__, hdr->fw_id, fwid));
9146 				goto error;
9147 			}
9148 
9149 			match_fail = FALSE;
9150 			num_fmts = hdr->rom_logstrs_offset / sizeof(uint32);
9151 			ram_index = (hdr->ram_lognums_offset -
9152 				hdr->rom_lognums_offset) / sizeof(uint32);
9153 			lognums = (uint32 *) &raw_fmts[hdr->rom_lognums_offset];
9154 			logstrs = (char *)	 &raw_fmts[hdr->rom_logstrs_offset];
9155 			hdr_logstrs_size = hdr->logstrs_size;
9156 
9157 error:
9158 			if (file) {
9159 				dhd_os_close_image1(NULL, file);
9160 			}
9161 			if (match_fail) {
9162 				return BCME_DECERR;
9163 			}
9164 		} else {
9165 			DHD_ERROR(("%s: Invalid logstr version %u\n", __FUNCTION__,
9166 					trailer->version));
9167 			return BCME_ERROR;
9168 		}
9169 		if (logstrs_size != hdr_logstrs_size) {
9170 			DHD_ERROR(("%s: bad logstrs_size %d\n", __FUNCTION__, hdr_logstrs_size));
9171 			return BCME_ERROR;
9172 		}
9173 	} else {
9174 		/*
9175 		 * Legacy logstrs.bin format without header.
9176 		 */
9177 		num_fmts = *((uint32 *) (raw_fmts)) / sizeof(uint32);
9178 
9179 		/* Legacy RAM-only logstrs.bin format:
9180 		 *	  - RAM 'lognums' section
9181 		 *	  - RAM 'logstrs' section.
9182 		 *
9183 		 * 'lognums' is an array of indexes for the strings in the
9184 		 * 'logstrs' section. The first uint32 is an index to the
9185 		 * start of 'logstrs'. Therefore, if this index is divided
9186 		 * by 'sizeof(uint32)' it provides the number of logstr
9187 		 *	entries.
9188 		 */
9189 		ram_index = 0;
9190 		lognums = (uint32 *) raw_fmts;
9191 		logstrs = (char *) &raw_fmts[num_fmts << 2];
9192 	}
9193 	if (num_fmts) {
9194 		if (event_log->fmts != NULL) {
9195 			fmts = event_log->fmts;	/* reuse existing malloced fmts */
9196 		} else {
9197 			fmts = MALLOC(osh, num_fmts  * sizeof(char *));
9198 		}
9199 	}
9200 	if (fmts == NULL) {
9201 		DHD_ERROR(("%s: Failed to allocate fmts memory\n", __FUNCTION__));
9202 		return BCME_ERROR;
9203 	}
9204 	event_log->fmts_size = num_fmts  * sizeof(char *);
9205 
9206 	for (i = 0; i < num_fmts; i++) {
9207 		/* ROM lognums index into logstrs using 'rom_logstrs_offset' as a base
9208 		* (they are 0-indexed relative to 'rom_logstrs_offset').
9209 		*
9210 		* RAM lognums are already indexed to point to the correct RAM logstrs (they
9211 		* are 0-indexed relative to the start of the logstrs.bin file).
9212 		*/
9213 		if (i == ram_index) {
9214 			logstrs = raw_fmts;
9215 		}
9216 		fmts[i] = &logstrs[lognums[i]];
9217 	}
9218 	event_log->fmts = fmts;
9219 	event_log->raw_fmts_size = logstrs_size;
9220 	event_log->raw_fmts = raw_fmts;
9221 	event_log->num_fmts = num_fmts;
9222 	return BCME_OK;
9223 } /* dhd_parse_logstrs_file */
9224 
dhd_parse_map_file(osl_t * osh,void * file,uint32 * ramstart,uint32 * rodata_start,uint32 * rodata_end)9225 int dhd_parse_map_file(osl_t *osh, void *file, uint32 *ramstart, uint32 *rodata_start,
9226 		uint32 *rodata_end)
9227 {
9228 	char *raw_fmts =  NULL, *raw_fmts_loc = NULL;
9229 	uint32 read_size = READ_NUM_BYTES;
9230 	int error = 0;
9231 	char * cptr = NULL;
9232 	char c;
9233 	uint8 count = 0;
9234 
9235 	*ramstart = 0;
9236 	*rodata_start = 0;
9237 	*rodata_end = 0;
9238 
9239 	/* Allocate 1 byte more than read_size to terminate it with NULL */
9240 	raw_fmts = MALLOCZ(osh, read_size + 1);
9241 	if (raw_fmts == NULL) {
9242 		DHD_ERROR(("%s: Failed to allocate raw_fmts memory \n", __FUNCTION__));
9243 		goto fail;
9244 	}
9245 
9246 	/* read ram start, rodata_start and rodata_end values from map  file */
9247 	while (count != ALL_MAP_VAL)
9248 	{
9249 		error = dhd_os_read_file(file, raw_fmts, read_size);
9250 		if (error < 0) {
9251 			DHD_ERROR(("%s: map file read failed err:%d \n", __FUNCTION__,
9252 					error));
9253 			goto fail;
9254 		}
9255 
9256 		/* End raw_fmts with NULL as strstr expects NULL terminated strings */
9257 		raw_fmts[read_size] = '\0';
9258 
9259 		/* Get ramstart address */
9260 		raw_fmts_loc = raw_fmts;
9261 		if (!(count & RAMSTART_BIT) &&
9262 			(cptr = bcmstrnstr(raw_fmts_loc, read_size, ramstart_str,
9263 			strlen(ramstart_str)))) {
9264 			cptr = cptr - BYTES_AHEAD_NUM;
9265 			sscanf(cptr, "%x %c text_start", ramstart, &c);
9266 			count |= RAMSTART_BIT;
9267 		}
9268 
9269 		/* Get ram rodata start address */
9270 		raw_fmts_loc = raw_fmts;
9271 		if (!(count & RDSTART_BIT) &&
9272 			(cptr = bcmstrnstr(raw_fmts_loc, read_size, rodata_start_str,
9273 			strlen(rodata_start_str)))) {
9274 			cptr = cptr - BYTES_AHEAD_NUM;
9275 			sscanf(cptr, "%x %c rodata_start", rodata_start, &c);
9276 			count |= RDSTART_BIT;
9277 		}
9278 
9279 		/* Get ram rodata end address */
9280 		raw_fmts_loc = raw_fmts;
9281 		if (!(count & RDEND_BIT) &&
9282 			(cptr = bcmstrnstr(raw_fmts_loc, read_size, rodata_end_str,
9283 			strlen(rodata_end_str)))) {
9284 			cptr = cptr - BYTES_AHEAD_NUM;
9285 			sscanf(cptr, "%x %c rodata_end", rodata_end, &c);
9286 			count |= RDEND_BIT;
9287 		}
9288 
9289 		if (error < (int)read_size) {
9290 			/*
9291 			* since we reset file pos back to earlier pos by
9292 			* GO_BACK_FILE_POS_NUM_BYTES bytes we won't reach EOF.
9293 			* The reason for this is if string is spreaded across
9294 			* bytes, the read function should not miss it.
9295 			* So if ret value is less than read_size, reached EOF don't read further
9296 			*/
9297 			break;
9298 		}
9299 		memset(raw_fmts, 0, read_size);
9300 		/*
9301 		* go back to predefined NUM of bytes so that we won't miss
9302 		* the string and  addr even if it comes as splited in next read.
9303 		*/
9304 		dhd_os_seek_file(file, -GO_BACK_FILE_POS_NUM_BYTES);
9305 	}
9306 
9307 fail:
9308 	if (raw_fmts) {
9309 		MFREE(osh, raw_fmts, read_size + 1);
9310 		raw_fmts = NULL;
9311 	}
9312 	if (count == ALL_MAP_VAL) {
9313 		return BCME_OK;
9314 	}
9315 	else {
9316 		DHD_ERROR(("%s: readmap error 0X%x \n", __FUNCTION__,
9317 				count));
9318 		return BCME_ERROR;
9319 	}
9320 
9321 } /* dhd_parse_map_file */
9322 
9323 #ifdef PCIE_FULL_DONGLE
9324 int
dhd_event_logtrace_infobuf_pkt_process(dhd_pub_t * dhdp,void * pktbuf,dhd_event_log_t * event_data)9325 dhd_event_logtrace_infobuf_pkt_process(dhd_pub_t *dhdp, void *pktbuf,
9326 		dhd_event_log_t *event_data)
9327 {
9328 	uint32 infobuf_version;
9329 	info_buf_payload_hdr_t *payload_hdr_ptr;
9330 	uint16 payload_hdr_type;
9331 	uint16 payload_hdr_length;
9332 
9333 	DHD_TRACE(("%s:Enter\n", __FUNCTION__));
9334 
9335 	if (PKTLEN(dhdp->osh, pktbuf) < sizeof(uint32)) {
9336 		DHD_ERROR(("%s: infobuf too small for version field\n",
9337 			__FUNCTION__));
9338 		goto exit;
9339 	}
9340 	infobuf_version = *((uint32 *)PKTDATA(dhdp->osh, pktbuf));
9341 	PKTPULL(dhdp->osh, pktbuf, sizeof(uint32));
9342 	if (infobuf_version != PCIE_INFOBUF_V1) {
9343 		DHD_ERROR(("%s: infobuf version %d is not PCIE_INFOBUF_V1\n",
9344 			__FUNCTION__, infobuf_version));
9345 		goto exit;
9346 	}
9347 
9348 	/* Version 1 infobuf has a single type/length (and then value) field */
9349 	if (PKTLEN(dhdp->osh, pktbuf) < sizeof(info_buf_payload_hdr_t)) {
9350 		DHD_ERROR(("%s: infobuf too small for v1 type/length  fields\n",
9351 			__FUNCTION__));
9352 		goto exit;
9353 	}
9354 	/* Process/parse the common info payload header (type/length) */
9355 	payload_hdr_ptr = (info_buf_payload_hdr_t *)PKTDATA(dhdp->osh, pktbuf);
9356 	payload_hdr_type = ltoh16(payload_hdr_ptr->type);
9357 	payload_hdr_length = ltoh16(payload_hdr_ptr->length);
9358 	if (payload_hdr_type != PCIE_INFOBUF_V1_TYPE_LOGTRACE) {
9359 		DHD_ERROR(("%s: payload_hdr_type %d is not V1_TYPE_LOGTRACE\n",
9360 			__FUNCTION__, payload_hdr_type));
9361 		goto exit;
9362 	}
9363 	PKTPULL(dhdp->osh, pktbuf, sizeof(info_buf_payload_hdr_t));
9364 
9365 	/* Validate that the specified length isn't bigger than the
9366 	 * provided data.
9367 	 */
9368 	if (payload_hdr_length > PKTLEN(dhdp->osh, pktbuf)) {
9369 		DHD_ERROR(("%s: infobuf logtrace length is bigger"
9370 			" than actual buffer data\n", __FUNCTION__));
9371 		goto exit;
9372 	}
9373 	dhd_dbg_trace_evnt_handler(dhdp, PKTDATA(dhdp->osh, pktbuf),
9374 		event_data, payload_hdr_length);
9375 
9376 	return BCME_OK;
9377 
9378 exit:
9379 	return BCME_ERROR;
9380 } /* dhd_event_logtrace_infobuf_pkt_process */
9381 #endif /* PCIE_FULL_DONGLE */
9382 #endif /* SHOW_LOGTRACE */
9383 
9384 #ifdef BTLOG
9385 int
dhd_bt_log_pkt_process(dhd_pub_t * dhdp,void * pktbuf)9386 dhd_bt_log_pkt_process(dhd_pub_t *dhdp, void *pktbuf)
9387 {
9388 	DHD_TRACE(("%s:Enter\n", __FUNCTION__));
9389 
9390 	dhd_dbg_bt_log_handler(dhdp,
9391 		PKTDATA(dhdp->osh, pktbuf), PKTLEN(dhdp->osh, pktbuf));
9392 
9393 	return BCME_OK;
9394 }
9395 #endif /* BTLOG */
9396 
9397 #if defined(WLTDLS) && defined(PCIE_FULL_DONGLE)
9398 
9399 /* To handle the TDLS event in the dhd_common.c
9400  */
dhd_tdls_event_handler(dhd_pub_t * dhd_pub,wl_event_msg_t * event)9401 int dhd_tdls_event_handler(dhd_pub_t *dhd_pub, wl_event_msg_t *event)
9402 {
9403 	int ret = BCME_OK;
9404 
9405 	GCC_DIAGNOSTIC_PUSH_SUPPRESS_CAST()
9406 	ret = dhd_tdls_update_peer_info(dhd_pub, event);
9407 	GCC_DIAGNOSTIC_POP()
9408 
9409 	return ret;
9410 }
9411 
dhd_free_tdls_peer_list(dhd_pub_t * dhd_pub)9412 int dhd_free_tdls_peer_list(dhd_pub_t *dhd_pub)
9413 {
9414 	tdls_peer_node_t *cur = NULL, *prev = NULL;
9415 	if (!dhd_pub)
9416 		return BCME_ERROR;
9417 	cur = dhd_pub->peer_tbl.node;
9418 
9419 	if ((dhd_pub->peer_tbl.node == NULL) && !dhd_pub->peer_tbl.tdls_peer_count)
9420 		return BCME_ERROR;
9421 
9422 	while (cur != NULL) {
9423 		prev = cur;
9424 		cur = cur->next;
9425 		MFREE(dhd_pub->osh, prev, sizeof(tdls_peer_node_t));
9426 	}
9427 	dhd_pub->peer_tbl.tdls_peer_count = 0;
9428 	dhd_pub->peer_tbl.node = NULL;
9429 	return BCME_OK;
9430 }
9431 #endif	/* #if defined(WLTDLS) && defined(PCIE_FULL_DONGLE) */
9432 
9433 /* pretty hex print a contiguous buffer
9434 * based on the debug level specified
9435 */
9436 void
dhd_prhex(const char * msg,volatile uchar * buf,uint nbytes,uint8 dbg_level)9437 dhd_prhex(const char *msg, volatile uchar *buf, uint nbytes, uint8 dbg_level)
9438 {
9439 	char line[128], *p;
9440 	int len = sizeof(line);
9441 	int nchar;
9442 	uint i;
9443 
9444 	if (msg && (msg[0] != '\0')) {
9445 		if (dbg_level == DHD_ERROR_VAL)
9446 			DHD_ERROR(("%s:\n", msg));
9447 		else if (dbg_level == DHD_INFO_VAL)
9448 			DHD_INFO(("%s:\n", msg));
9449 		else if (dbg_level == DHD_TRACE_VAL)
9450 			DHD_TRACE(("%s:\n", msg));
9451 	}
9452 
9453 	p = line;
9454 	for (i = 0; i < nbytes; i++) {
9455 		if (i % 16 == 0) {
9456 			nchar = snprintf(p, len, "  %04x: ", i);	/* line prefix */
9457 			p += nchar;
9458 			len -= nchar;
9459 		}
9460 		if (len > 0) {
9461 			nchar = snprintf(p, len, "%02x ", buf[i]);
9462 			p += nchar;
9463 			len -= nchar;
9464 		}
9465 
9466 		if (i % 16 == 15) {
9467 			/* flush line */
9468 			if (dbg_level == DHD_ERROR_VAL)
9469 				DHD_ERROR(("%s:\n", line));
9470 			else if (dbg_level == DHD_INFO_VAL)
9471 				DHD_INFO(("%s:\n", line));
9472 			else if (dbg_level == DHD_TRACE_VAL)
9473 				DHD_TRACE(("%s:\n", line));
9474 			p = line;
9475 			len = sizeof(line);
9476 		}
9477 	}
9478 
9479 	/* flush last partial line */
9480 	if (p != line) {
9481 		if (dbg_level == DHD_ERROR_VAL)
9482 			DHD_ERROR(("%s:\n", line));
9483 		else if (dbg_level == DHD_INFO_VAL)
9484 			DHD_INFO(("%s:\n", line));
9485 		else if (dbg_level == DHD_TRACE_VAL)
9486 			DHD_TRACE(("%s:\n", line));
9487 	}
9488 }
9489 
9490 int
dhd_tput_test(dhd_pub_t * dhd,tput_test_t * tput_data)9491 dhd_tput_test(dhd_pub_t *dhd, tput_test_t *tput_data)
9492 {
9493 	struct ether_header ether_hdr;
9494 	tput_pkt_t tput_pkt;
9495 	void *pkt = NULL;
9496 	uint8 *pktdata = NULL;
9497 	uint32 pktsize = 0;
9498 	uint64 total_size = 0;
9499 	uint32 *crc = 0;
9500 	uint32 pktid = 0;
9501 	uint32 total_num_tx_pkts = 0;
9502 	int err = 0, err_exit = 0;
9503 	uint32 i = 0;
9504 	uint64 time_taken = 0;
9505 	int max_txbufs = 0;
9506 	uint32 n_batches = 0;
9507 	uint32 n_remain = 0;
9508 	uint8 tput_pkt_hdr_size = 0;
9509 	bool batch_cnt = FALSE;
9510 	bool tx_stop_pkt = FALSE;
9511 
9512 #if defined(DHD_EFI) && defined(DHD_INTR_POLL_PERIOD_DYNAMIC)
9513 	uint32 cur_intr_poll_period = 0;
9514 	cur_intr_poll_period = dhd_os_get_intr_poll_period();
9515 	/* before running tput_test, set interrupt poll period to a lesser value */
9516 	dhd_os_set_intr_poll_period(dhd->bus, INTR_POLL_PERIOD_CRITICAL);
9517 #endif	/* DHD_EFI && DHD_INTR_POLL_PERIOD_DYNAMIC */
9518 
9519 	if (tput_data->version != TPUT_TEST_T_VER ||
9520 		tput_data->length != TPUT_TEST_T_LEN) {
9521 		DHD_ERROR(("%s: wrong structure ver/len! \n", __FUNCTION__));
9522 		err_exit = BCME_BADARG;
9523 		goto exit_error;
9524 	}
9525 
9526 	if (dhd->tput_data.tput_test_running) {
9527 		DHD_ERROR(("%s: tput test already running ! \n", __FUNCTION__));
9528 		err_exit = BCME_BUSY;
9529 		goto exit_error;
9530 	}
9531 #ifdef PCIE_FULL_DONGLE
9532 	/*
9533 	 * 100 bytes to accommodate ether header and tput header. As of today
9534 	 * both occupy 30 bytes. Rest is reserved.
9535 	 */
9536 	if ((tput_data->payload_size > TPUT_TEST_MAX_PAYLOAD) ||
9537 		(tput_data->payload_size > (DHD_FLOWRING_RX_BUFPOST_PKTSZ - 100))) {
9538 		DHD_ERROR(("%s: payload size is too large! max_payload=%u rx_bufpost_size=%u\n",
9539 			__FUNCTION__, TPUT_TEST_MAX_PAYLOAD,
9540 			(DHD_FLOWRING_RX_BUFPOST_PKTSZ - 100)));
9541 		err_exit = BCME_BUFTOOLONG;
9542 		goto exit_error;
9543 	}
9544 #endif
9545 	max_txbufs = dhd_get_max_txbufs(dhd);
9546 	max_txbufs = MIN(max_txbufs, DHD_TPUT_MAX_TX_PKTS_BATCH);
9547 
9548 	if (!(tput_data->num_pkts > 0)) {
9549 		DHD_ERROR(("%s: invalid num_pkts: %d to tx\n",
9550 			__FUNCTION__, tput_data->num_pkts));
9551 		err_exit = BCME_ERROR;
9552 		goto exit_error;
9553 	}
9554 
9555 	memset(&dhd->tput_data, 0, sizeof(dhd->tput_data));
9556 	memcpy(&dhd->tput_data, tput_data, sizeof(*tput_data));
9557 	dhd->tput_data.pkts_bad = dhd->tput_data.pkts_good = 0;
9558 	dhd->tput_data.pkts_cmpl = 0;
9559 	dhd->tput_start_ts = dhd->tput_stop_ts = 0;
9560 
9561 	if (tput_data->flags & TPUT_TEST_USE_ETHERNET_HDR) {
9562 		pktsize = sizeof(ether_hdr) + sizeof(tput_pkt_t) +
9563 				(tput_data->payload_size - 12);
9564 	} else {
9565 		pktsize = sizeof(tput_pkt_t) +
9566 				(tput_data->payload_size - 12);
9567 	}
9568 
9569 	tput_pkt_hdr_size = (uint8)((uint8 *)&tput_pkt.crc32 -
9570 			(uint8 *)&tput_pkt.mac_sta);
9571 
9572 	/* mark the tput test as started */
9573 	dhd->tput_data.tput_test_running = TRUE;
9574 
9575 	if (tput_data->direction == TPUT_DIR_TX) {
9576 		/* for ethernet header */
9577 		memcpy(ether_hdr.ether_shost, tput_data->mac_sta, ETHER_ADDR_LEN);
9578 		memcpy(ether_hdr.ether_dhost, tput_data->mac_ap, ETHER_ADDR_LEN);
9579 		ether_hdr.ether_type = hton16(ETHER_TYPE_IP);
9580 
9581 		/* fill in the tput pkt */
9582 		memset(&tput_pkt, 0, sizeof(tput_pkt));
9583 		memcpy(tput_pkt.mac_ap, tput_data->mac_ap, ETHER_ADDR_LEN);
9584 		memcpy(tput_pkt.mac_sta, tput_data->mac_sta, ETHER_ADDR_LEN);
9585 		tput_pkt.pkt_type = hton16(TPUT_PKT_TYPE_NORMAL);
9586 		tput_pkt.num_pkts = hton32(tput_data->num_pkts);
9587 
9588 		if (tput_data->num_pkts > (uint32)max_txbufs) {
9589 			n_batches = tput_data->num_pkts / max_txbufs;
9590 			n_remain = tput_data->num_pkts % max_txbufs;
9591 		} else {
9592 			n_batches = 0;
9593 			n_remain = tput_data->num_pkts;
9594 		}
9595 		DHD_ERROR(("%s: num_pkts: %u n_batches: %u n_remain: %u\n",
9596 			__FUNCTION__, tput_data->num_pkts, n_batches, n_remain));
9597 
9598 		do {
9599 			/* reset before every batch */
9600 			dhd->batch_tx_pkts_cmpl = 0;
9601 			if (n_batches) {
9602 				dhd->batch_tx_num_pkts = max_txbufs;
9603 				--n_batches;
9604 			} else if (n_remain) {
9605 				dhd->batch_tx_num_pkts = n_remain;
9606 				n_remain = 0;
9607 			} else {
9608 				DHD_ERROR(("Invalid. This should not hit\n"));
9609 			}
9610 
9611 			dhd->tput_start_ts = OSL_SYSUPTIME_US();
9612 			for (i = 0; (i < dhd->batch_tx_num_pkts) || (tx_stop_pkt); ++i) {
9613 				pkt = PKTGET(dhd->osh, pktsize, TRUE);
9614 				if (!pkt) {
9615 					dhd->tput_data.tput_test_running = FALSE;
9616 					DHD_ERROR(("%s: PKTGET fails ! Not enough Tx buffers\n",
9617 						__FUNCTION__));
9618 					DHD_ERROR(("%s: pkts_good:%u; pkts_bad:%u; pkts_cmpl:%u\n",
9619 						__FUNCTION__, dhd->tput_data.pkts_good,
9620 						dhd->tput_data.pkts_bad, dhd->tput_data.pkts_cmpl));
9621 					err_exit = BCME_NOMEM;
9622 					goto exit_error;
9623 				}
9624 				pktdata = PKTDATA(dhd->osh, pkt);
9625 				PKTSETLEN(dhd->osh, pkt, pktsize);
9626 				memset(pktdata, 0, pktsize);
9627 				if (tput_data->flags & TPUT_TEST_USE_ETHERNET_HDR) {
9628 					memcpy(pktdata, &ether_hdr, sizeof(ether_hdr));
9629 					pktdata += sizeof(ether_hdr);
9630 				}
9631 				/* send stop pkt as last pkt */
9632 				if (tx_stop_pkt) {
9633 					tput_pkt.pkt_type = hton16(TPUT_PKT_TYPE_STOP);
9634 					tx_stop_pkt = FALSE;
9635 				} else
9636 					tput_pkt.pkt_type = hton16(TPUT_PKT_TYPE_NORMAL);
9637 				tput_pkt.pkt_id = hton32(pktid++);
9638 				tput_pkt.crc32 = 0;
9639 				memcpy(pktdata, &tput_pkt, sizeof(tput_pkt));
9640 				/* compute crc32 over the pkt-id, num-pkts and data fields */
9641 				crc = (uint32 *)(pktdata + tput_pkt_hdr_size);
9642 				*crc = hton32(hndcrc32(pktdata + tput_pkt_hdr_size + 4,
9643 						8 + (tput_data->payload_size - 12),
9644 						CRC32_INIT_VALUE));
9645 
9646 				err = dhd_sendpkt(dhd, 0, pkt);
9647 				if (err != BCME_OK) {
9648 					DHD_INFO(("%s: send pkt (id = %u) fails (err = %d) ! \n",
9649 						__FUNCTION__, pktid, err));
9650 					dhd->tput_data.pkts_bad++;
9651 				}
9652 				total_num_tx_pkts++;
9653 				if ((total_num_tx_pkts == tput_data->num_pkts) && (!tx_stop_pkt)) {
9654 					tx_stop_pkt = TRUE;
9655 				}
9656 			}
9657 			DHD_INFO(("%s: TX done, wait for completion...\n", __FUNCTION__));
9658 			if (!dhd_os_tput_test_wait(dhd, NULL,
9659 					TPUT_TEST_WAIT_TIMEOUT_DEFAULT)) {
9660 				dhd->tput_stop_ts = OSL_SYSUPTIME_US();
9661 				dhd->tput_data.tput_test_running = FALSE;
9662 				DHD_ERROR(("%s: TX completion timeout !"
9663 					" Total Tx pkts (including STOP) = %u; pkts cmpl = %u; \n",
9664 					__FUNCTION__, total_num_tx_pkts, dhd->batch_tx_pkts_cmpl));
9665 				err_exit = BCME_ERROR;
9666 				goto exit_error;
9667 			}
9668 			if ((dhd->tput_start_ts && dhd->tput_stop_ts &&
9669 				(dhd->tput_stop_ts > dhd->tput_start_ts)) || (time_taken)) {
9670 				if (!time_taken) {
9671 					time_taken = dhd->tput_stop_ts - dhd->tput_start_ts;
9672 				}
9673 			} else {
9674 				dhd->tput_data.tput_test_running = FALSE;
9675 				DHD_ERROR(("%s: bad timestamp while cal tx batch time\n",
9676 					__FUNCTION__));
9677 				err_exit = BCME_ERROR;
9678 				goto exit_error;
9679 			}
9680 			if (n_batches || n_remain) {
9681 				batch_cnt = TRUE;
9682 			} else {
9683 				batch_cnt = FALSE;
9684 			}
9685 		} while (batch_cnt);
9686 	} else {
9687 		/* TPUT_DIR_RX */
9688 		DHD_INFO(("%s: waiting for RX completion... \n", __FUNCTION__));
9689 		if (!dhd_os_tput_test_wait(dhd, NULL, tput_data->timeout_ms)) {
9690 			DHD_ERROR(("%s: RX completion timeout ! \n", __FUNCTION__));
9691 			dhd->tput_stop_ts = OSL_SYSUPTIME_US();
9692 		}
9693 	}
9694 
9695 	/* calculate the throughput in bits per sec */
9696 	if (dhd->tput_start_ts && dhd->tput_stop_ts &&
9697 		(dhd->tput_stop_ts > dhd->tput_start_ts)) {
9698 		time_taken = dhd->tput_stop_ts - dhd->tput_start_ts;
9699 		time_taken = DIV_U64_BY_U32(time_taken, MSEC_PER_SEC); /* convert to ms */
9700 		dhd->tput_data.time_ms = time_taken;
9701 		if (time_taken) {
9702 			total_size = pktsize * dhd->tput_data.pkts_cmpl * 8;
9703 			dhd->tput_data.tput_bps = DIV_U64_BY_U64(total_size, time_taken);
9704 			/* convert from ms to seconds */
9705 			dhd->tput_data.tput_bps = dhd->tput_data.tput_bps * 1000;
9706 		}
9707 	} else {
9708 		DHD_ERROR(("%s: bad timestamp !\n", __FUNCTION__));
9709 	}
9710 	DHD_INFO(("%s: DONE. tput = %llu bps, time = %llu ms\n", __FUNCTION__,
9711 		dhd->tput_data.tput_bps, dhd->tput_data.time_ms));
9712 
9713 	memcpy(tput_data, &dhd->tput_data, sizeof(dhd->tput_data));
9714 
9715 	dhd->tput_data.tput_test_running = FALSE;
9716 
9717 	err_exit = BCME_OK;
9718 
9719 exit_error:
9720 	DHD_ERROR(("%s: pkts_good = %u; pkts_bad = %u; pkts_cmpl = %u\n",
9721 		__FUNCTION__, dhd->tput_data.pkts_good,
9722 		dhd->tput_data.pkts_bad, dhd->tput_data.pkts_cmpl));
9723 #if defined(DHD_EFI) && defined(DHD_INTR_POLL_PERIOD_DYNAMIC)
9724 	/* restore interrupt poll period to the previous existing value */
9725 	dhd_os_set_intr_poll_period(dhd->bus, cur_intr_poll_period);
9726 #endif	/* DHD_EFI && DHD_INTR_POLL_PERIOD_DYNAMIC */
9727 
9728 	return err_exit;
9729 }
9730 
9731 void
dhd_tput_test_rx(dhd_pub_t * dhd,void * pkt)9732 dhd_tput_test_rx(dhd_pub_t *dhd, void *pkt)
9733 {
9734 	uint8 *pktdata = NULL;
9735 	tput_pkt_t *tput_pkt = NULL;
9736 	uint32 crc = 0;
9737 	uint8 tput_pkt_hdr_size = 0;
9738 
9739 	pktdata = PKTDATA(dhd->osh, pkt);
9740 	if (dhd->tput_data.flags & TPUT_TEST_USE_ETHERNET_HDR)
9741 		pktdata += sizeof(struct ether_header);
9742 	tput_pkt = (tput_pkt_t *)pktdata;
9743 
9744 	/* record the timestamp of the first packet received */
9745 	if (dhd->tput_data.pkts_cmpl == 0) {
9746 		dhd->tput_start_ts = OSL_SYSUPTIME_US();
9747 	}
9748 
9749 	if (ntoh16(tput_pkt->pkt_type) != TPUT_PKT_TYPE_STOP &&
9750 			dhd->tput_data.pkts_cmpl <= dhd->tput_data.num_pkts) {
9751 		dhd->tput_data.pkts_cmpl++;
9752 	}
9753 	/* drop rx packets received beyond the specified # */
9754 	if (dhd->tput_data.pkts_cmpl > dhd->tput_data.num_pkts)
9755 		return;
9756 
9757 	DHD_TRACE(("%s: Rx tput test pkt, id = %u ; type = %u\n", __FUNCTION__,
9758 		ntoh32(tput_pkt->pkt_id), ntoh16(tput_pkt->pkt_type)));
9759 
9760 	/* discard if mac addr of AP/STA does not match the specified ones */
9761 	if ((memcmp(tput_pkt->mac_ap, dhd->tput_data.mac_ap,
9762 			ETHER_ADDR_LEN) != 0) ||
9763 		(memcmp(tput_pkt->mac_sta, dhd->tput_data.mac_sta,
9764 			ETHER_ADDR_LEN) != 0)) {
9765 		dhd->tput_data.pkts_bad++;
9766 		DHD_INFO(("%s: dropping tput pkt with id %u due to bad AP/STA mac !\n",
9767 			__FUNCTION__, ntoh32(tput_pkt->pkt_id)));
9768 		return;
9769 	}
9770 
9771 	tput_pkt_hdr_size = (uint8)((uint8 *)&tput_pkt->crc32 -
9772 			(uint8 *)&tput_pkt->mac_sta);
9773 	pktdata += tput_pkt_hdr_size + 4;
9774 	crc = hndcrc32(pktdata, 8 + (dhd->tput_data.payload_size - 12),
9775 			CRC32_INIT_VALUE);
9776 	if (crc != ntoh32(tput_pkt->crc32)) {
9777 		DHD_INFO(("%s: dropping tput pkt with id %u due to bad CRC !\n",
9778 			__FUNCTION__, ntoh32(tput_pkt->pkt_id)));
9779 		dhd->tput_data.pkts_bad++;
9780 		return;
9781 	}
9782 
9783 	if (ntoh16(tput_pkt->pkt_type) != TPUT_PKT_TYPE_STOP)
9784 		dhd->tput_data.pkts_good++;
9785 
9786 	/* if we have received the stop packet or all the # of pkts, we're done */
9787 	if (ntoh16(tput_pkt->pkt_type) == TPUT_PKT_TYPE_STOP ||
9788 			dhd->tput_data.pkts_cmpl == dhd->tput_data.num_pkts) {
9789 		dhd->tput_stop_ts = OSL_SYSUPTIME_US();
9790 		dhd_os_tput_test_wake(dhd);
9791 	}
9792 }
9793 
9794 #ifdef DUMP_IOCTL_IOV_LIST
9795 void
dhd_iov_li_append(dhd_pub_t * dhd,dll_t * list_head,dll_t * node)9796 dhd_iov_li_append(dhd_pub_t *dhd, dll_t *list_head, dll_t *node)
9797 {
9798 	dll_t *item;
9799 	dhd_iov_li_t *iov_li;
9800 	dhd->dump_iovlist_len++;
9801 
9802 	if (dhd->dump_iovlist_len == IOV_LIST_MAX_LEN+1) {
9803 		item = dll_head_p(list_head);
9804 		iov_li = (dhd_iov_li_t *)CONTAINEROF(item, dhd_iov_li_t, list);
9805 		dll_delete(item);
9806 		MFREE(dhd->osh, iov_li, sizeof(*iov_li));
9807 		dhd->dump_iovlist_len--;
9808 	}
9809 	dll_append(list_head, node);
9810 }
9811 
9812 void
dhd_iov_li_print(dll_t * list_head)9813 dhd_iov_li_print(dll_t *list_head)
9814 {
9815 	dhd_iov_li_t *iov_li;
9816 	dll_t *item, *next;
9817 	uint8 index = 0;
9818 	for (item = dll_head_p(list_head); !dll_end(list_head, item); item = next) {
9819 		next = dll_next_p(item);
9820 		iov_li = (dhd_iov_li_t *)CONTAINEROF(item, dhd_iov_li_t, list);
9821 		DHD_ERROR(("%d:cmd_name = %s, cmd = %d.\n", ++index, iov_li->buff, iov_li->cmd));
9822 	}
9823 }
9824 
9825 void
dhd_iov_li_delete(dhd_pub_t * dhd,dll_t * list_head)9826 dhd_iov_li_delete(dhd_pub_t *dhd, dll_t *list_head)
9827 {
9828 	dll_t *item;
9829 	dhd_iov_li_t *iov_li;
9830 	while (!(dll_empty(list_head))) {
9831 		item = dll_head_p(list_head);
9832 		iov_li = (dhd_iov_li_t *)CONTAINEROF(item, dhd_iov_li_t, list);
9833 		dll_delete(item);
9834 		MFREE(dhd->osh, iov_li, sizeof(*iov_li));
9835 	}
9836 }
9837 #endif /* DUMP_IOCTL_IOV_LIST */
9838 
9839 #ifdef EWP_EDL
9840 /* For now we are allocating memory for EDL ring using DMA_ALLOC_CONSISTENT
9841 * The reason being that, in hikey, if we try to DMA_MAP prealloced memory
9842 * it is failing with an 'out of space in SWIOTLB' error
9843 */
9844 int
dhd_edl_mem_init(dhd_pub_t * dhd)9845 dhd_edl_mem_init(dhd_pub_t *dhd)
9846 {
9847 	int ret = 0;
9848 
9849 	memset(&dhd->edl_ring_mem, 0, sizeof(dhd->edl_ring_mem));
9850 	ret = dhd_dma_buf_alloc(dhd, &dhd->edl_ring_mem, DHD_EDL_RING_SIZE);
9851 	if (ret != BCME_OK) {
9852 		DHD_ERROR(("%s: alloc of edl_ring_mem failed\n",
9853 			__FUNCTION__));
9854 		return BCME_ERROR;
9855 	}
9856 	return BCME_OK;
9857 }
9858 
9859 /*
9860  * NOTE:- that dhd_edl_mem_deinit need NOT be called explicitly, because the dma_buf
9861  * for EDL is freed during 'dhd_prot_detach_edl_rings' which is called during de-init.
9862  */
9863 void
dhd_edl_mem_deinit(dhd_pub_t * dhd)9864 dhd_edl_mem_deinit(dhd_pub_t *dhd)
9865 {
9866 	if (dhd->edl_ring_mem.va != NULL)
9867 		dhd_dma_buf_free(dhd, &dhd->edl_ring_mem);
9868 }
9869 
9870 int
dhd_event_logtrace_process_edl(dhd_pub_t * dhdp,uint8 * data,void * evt_decode_data)9871 dhd_event_logtrace_process_edl(dhd_pub_t *dhdp, uint8 *data,
9872 		void *evt_decode_data)
9873 {
9874 	msg_hdr_edl_t *msg = NULL;
9875 	cmn_msg_hdr_t *cmn_msg_hdr = NULL;
9876 	uint8 *buf = NULL;
9877 
9878 	if (!data || !dhdp || !evt_decode_data) {
9879 		DHD_ERROR(("%s: invalid args ! \n", __FUNCTION__));
9880 		return BCME_ERROR;
9881 	}
9882 
9883 	/* format of data in each work item in the EDL ring:
9884 	* |cmn_msg_hdr_t |payload (var len)|cmn_msg_hdr_t|
9885 	* payload = |infobuf_ver(u32)|info_buf_payload_hdr_t|msgtrace_hdr_t|<var len data>|
9886 	*/
9887 	cmn_msg_hdr = (cmn_msg_hdr_t *)data;
9888 	msg = (msg_hdr_edl_t *)(data + sizeof(cmn_msg_hdr_t));
9889 	buf = (uint8 *)msg;
9890 	/* validate the fields */
9891 	if (ltoh32(msg->infobuf_ver) != PCIE_INFOBUF_V1) {
9892 		DHD_ERROR(("%s: Skipping msg with invalid infobuf ver (0x%x)"
9893 			" expected (0x%x)\n", __FUNCTION__,
9894 			msg->infobuf_ver, PCIE_INFOBUF_V1));
9895 		return BCME_VERSION;
9896 	}
9897 
9898 	/* in EDL, the request_id field of cmn_msg_hdr is overloaded to carry payload length */
9899 	if (sizeof(info_buf_payload_hdr_t) > cmn_msg_hdr->request_id) {
9900 		DHD_ERROR(("%s: infobuf too small for v1 type/length fields\n",
9901 			__FUNCTION__));
9902 		return BCME_BUFTOOLONG;
9903 	}
9904 
9905 	if (ltoh16(msg->pyld_hdr.type) != PCIE_INFOBUF_V1_TYPE_LOGTRACE) {
9906 		DHD_ERROR(("%s: payload_hdr_type %d is not V1_TYPE_LOGTRACE\n",
9907 			__FUNCTION__, ltoh16(msg->pyld_hdr.type)));
9908 		return BCME_BADOPTION;
9909 	}
9910 
9911 	if (ltoh16(msg->pyld_hdr.length) > cmn_msg_hdr->request_id) {
9912 		DHD_ERROR(("%s: infobuf logtrace length %u is bigger"
9913 			" than available buffer size %u\n", __FUNCTION__,
9914 			ltoh16(msg->pyld_hdr.length), cmn_msg_hdr->request_id));
9915 		return BCME_BADLEN;
9916 	}
9917 
9918 	/* dhd_dbg_trace_evnt_handler expects the data to start from msgtrace_hdr_t */
9919 	buf += sizeof(msg->infobuf_ver) + sizeof(msg->pyld_hdr);
9920 	dhd_dbg_trace_evnt_handler(dhdp, buf, evt_decode_data,
9921 		ltoh16(msg->pyld_hdr.length));
9922 
9923 	/*
9924 	 * check 'dhdp->logtrace_pkt_sendup' and if true alloc an skb
9925 	 * copy the event data to the skb and send it up the stack
9926 	 */
9927 	if (dhdp->logtrace_pkt_sendup) {
9928 		DHD_INFO(("%s: send up event log, len %u bytes\n", __FUNCTION__,
9929 				(uint32)(ltoh16(msg->pyld_hdr.length) +
9930 				sizeof(info_buf_payload_hdr_t) + 4)));
9931 		dhd_sendup_info_buf(dhdp, (uint8 *)msg);
9932 	}
9933 
9934 	return BCME_OK;
9935 }
9936 #endif /* EWP_EDL */
9937 
9938 #ifdef DHD_LOG_DUMP
9939 #define DEBUG_DUMP_TRIGGER_INTERVAL_SEC	4
9940 void
dhd_log_dump_trigger(dhd_pub_t * dhdp,int subcmd)9941 dhd_log_dump_trigger(dhd_pub_t *dhdp, int subcmd)
9942 {
9943 #if defined(DHD_DUMP_FILE_WRITE_FROM_KERNEL)
9944 	log_dump_type_t *flush_type;
9945 #endif /* DHD_DUMP_FILE_WRITE_FROM_KERNEL */
9946 	uint64 current_time_sec;
9947 
9948 	if (!dhdp) {
9949 		DHD_ERROR(("dhdp is NULL !\n"));
9950 		return;
9951 	}
9952 
9953 	if (subcmd >= CMD_MAX || subcmd < CMD_DEFAULT) {
9954 		DHD_ERROR(("%s : Invalid subcmd \n", __FUNCTION__));
9955 		return;
9956 	}
9957 
9958 	current_time_sec = DIV_U64_BY_U32(OSL_LOCALTIME_NS(), NSEC_PER_SEC);
9959 
9960 	DHD_ERROR(("%s: current_time_sec=%lld debug_dump_time_sec=%lld interval=%d\n",
9961 		__FUNCTION__, current_time_sec, dhdp->debug_dump_time_sec,
9962 		DEBUG_DUMP_TRIGGER_INTERVAL_SEC));
9963 
9964 	if ((current_time_sec - dhdp->debug_dump_time_sec) < DEBUG_DUMP_TRIGGER_INTERVAL_SEC) {
9965 		DHD_ERROR(("%s : Last debug dump triggered(%lld) within %d seconds, so SKIP\n",
9966 			__FUNCTION__, dhdp->debug_dump_time_sec, DEBUG_DUMP_TRIGGER_INTERVAL_SEC));
9967 		return;
9968 	}
9969 
9970 	clear_debug_dump_time(dhdp->debug_dump_time_str);
9971 #ifdef DHD_PCIE_RUNTIMEPM
9972 	/* wake up RPM if SYSDUMP is triggered */
9973 	dhdpcie_runtime_bus_wake(dhdp, TRUE, __builtin_return_address(0));
9974 #endif /* DHD_PCIE_RUNTIMEPM */
9975 	/*  */
9976 
9977 	dhdp->debug_dump_subcmd = subcmd;
9978 
9979 	dhdp->debug_dump_time_sec = DIV_U64_BY_U32(OSL_LOCALTIME_NS(), NSEC_PER_SEC);
9980 
9981 #if defined(DHD_DUMP_FILE_WRITE_FROM_KERNEL)
9982 	/* flush_type is freed at do_dhd_log_dump function */
9983 	flush_type = MALLOCZ(dhdp->osh, sizeof(log_dump_type_t));
9984 	if (flush_type) {
9985 		*flush_type = DLD_BUF_TYPE_ALL;
9986 		dhd_schedule_log_dump(dhdp, flush_type);
9987 	} else {
9988 		DHD_ERROR(("%s Fail to malloc flush_type\n", __FUNCTION__));
9989 		return;
9990 	}
9991 #endif /* DHD_DUMP_FILE_WRITE_FROM_KERNEL */
9992 
9993 	/* Inside dhd_mem_dump, event notification will be sent to HAL and
9994 	 * from other context DHD pushes memdump, debug_dump and pktlog dump
9995 	 * to HAL and HAL will write into file
9996 	 */
9997 #if (defined(BCMPCIE) || defined(BCMSDIO)) && defined(DHD_FW_COREDUMP)
9998 	dhdp->memdump_type = DUMP_TYPE_BY_SYSDUMP;
9999 	dhd_bus_mem_dump(dhdp);
10000 #endif /* BCMPCIE && DHD_FW_COREDUMP */
10001 
10002 #if defined(DHD_PKT_LOGGING) && defined(DHD_DUMP_FILE_WRITE_FROM_KERNEL)
10003 	dhd_schedule_pktlog_dump(dhdp);
10004 #endif /* DHD_PKT_LOGGING && DHD_DUMP_FILE_WRITE_FROM_KERNEL */
10005 }
10006 #endif /* DHD_LOG_DUMP */
10007 
10008 #if (defined(LINUX) || defined(DHD_EFI)) && defined(SHOW_LOGTRACE)
10009 int
dhd_print_fw_ver_from_file(dhd_pub_t * dhdp,char * fwpath)10010 dhd_print_fw_ver_from_file(dhd_pub_t *dhdp, char *fwpath)
10011 {
10012 	void *file = NULL;
10013 	int size = 0;
10014 	char buf[FW_VER_STR_LEN];
10015 	char *str = NULL;
10016 	int ret = BCME_OK;
10017 
10018 	if (!fwpath)
10019 		return BCME_BADARG;
10020 
10021 	file = dhd_os_open_image1(dhdp, fwpath);
10022 	if (!file) {
10023 		ret = BCME_ERROR;
10024 		goto exit;
10025 	}
10026 	size = dhd_os_get_image_size(file);
10027 	if (!size) {
10028 		ret = BCME_ERROR;
10029 		goto exit;
10030 	}
10031 
10032 	/* seek to the last 'X' bytes in the file */
10033 	if (dhd_os_seek_file(file, size - FW_VER_STR_LEN) != BCME_OK) {
10034 		ret = BCME_ERROR;
10035 		goto exit;
10036 	}
10037 
10038 	/* read the last 'X' bytes of the file to a buffer */
10039 	memset(buf, 0, FW_VER_STR_LEN);
10040 	if (dhd_os_get_image_block(buf, FW_VER_STR_LEN - 1, file) < 0) {
10041 		ret = BCME_ERROR;
10042 		goto exit;
10043 	}
10044 	/* search for 'Version' in the buffer */
10045 	str = bcmstrnstr(buf, FW_VER_STR_LEN, FW_VER_STR, strlen(FW_VER_STR));
10046 	if (!str) {
10047 		ret = BCME_ERROR;
10048 		goto exit;
10049 	}
10050 	/* go back in the buffer to the last ascii character */
10051 	while (str != buf &&
10052 		(*str >= ' ' && *str <= '~')) {
10053 		--str;
10054 	}
10055 	/* reverse the final decrement, so that str is pointing
10056 	* to the first ascii character in the buffer
10057 	*/
10058 	++str;
10059 
10060 	if (strlen(str) > (FW_VER_STR_LEN - 1)) {
10061 		ret = BCME_BADLEN;
10062 		goto exit;
10063 	}
10064 
10065 	DHD_ERROR(("FW version in file '%s': %s\n", fwpath, str));
10066 	/* copy to global variable, so that in case FW load fails, the
10067 	* core capture logs will contain FW version read from the file
10068 	*/
10069 	memset(fw_version, 0, FW_VER_STR_LEN);
10070 	strlcpy(fw_version, str, FW_VER_STR_LEN);
10071 
10072 exit:
10073 	if (file)
10074 		dhd_os_close_image1(dhdp, file);
10075 
10076 	return ret;
10077 }
10078 #endif /* LINUX || DHD_EFI */
10079 
10080 #if defined(DHD_AWDL) && defined(AWDL_SLOT_STATS)
10081 void
dhd_clear_awdl_stats(dhd_pub_t * dhd)10082 dhd_clear_awdl_stats(dhd_pub_t *dhd)
10083 {
10084 	unsigned long flags;
10085 	/*
10086 	 * Since event path(ex: WLC_E_AWDL_AW) and bus path(tx status process) update
10087 	 * the AWDL data acquire lock before clearing the AWDL stats.
10088 	 */
10089 	DHD_AWDL_STATS_LOCK(dhd->awdl_stats_lock, flags);
10090 	memset(dhd->awdl_stats, 0, sizeof(dhd->awdl_stats));
10091 	DHD_AWDL_STATS_UNLOCK(dhd->awdl_stats_lock, flags);
10092 }
10093 #endif /* DHD_AWDL && AWDL_SLOT_STATS */
10094 
10095 #ifdef WL_CFGVENDOR_SEND_HANG_EVENT
10096 
10097 static void
copy_hang_info_ioctl_timeout(dhd_pub_t * dhd,int ifidx,wl_ioctl_t * ioc)10098 copy_hang_info_ioctl_timeout(dhd_pub_t *dhd, int ifidx, wl_ioctl_t *ioc)
10099 {
10100 	int remain_len;
10101 	int i;
10102 	int *cnt;
10103 	char *dest;
10104 	int bytes_written;
10105 	uint32 ioc_dwlen = 0;
10106 
10107 	if (!dhd || !dhd->hang_info) {
10108 		DHD_ERROR(("%s dhd=%p hang_info=%p\n",
10109 			__FUNCTION__, dhd, (dhd ? dhd->hang_info : NULL)));
10110 		return;
10111 	}
10112 
10113 	cnt = &dhd->hang_info_cnt;
10114 	dest = dhd->hang_info;
10115 
10116 	memset(dest, 0, VENDOR_SEND_HANG_EXT_INFO_LEN);
10117 	(*cnt) = 0;
10118 
10119 	bytes_written = 0;
10120 	remain_len = VENDOR_SEND_HANG_EXT_INFO_LEN - bytes_written;
10121 
10122 	get_debug_dump_time(dhd->debug_dump_time_hang_str);
10123 	copy_debug_dump_time(dhd->debug_dump_time_str, dhd->debug_dump_time_hang_str);
10124 
10125 	bytes_written += scnprintf(&dest[bytes_written], remain_len, "%d %d %s %d %d %d %d %d %d ",
10126 			HANG_REASON_IOCTL_RESP_TIMEOUT, VENDOR_SEND_HANG_EXT_INFO_VER,
10127 			dhd->debug_dump_time_hang_str,
10128 			ifidx, ioc->cmd, ioc->len, ioc->set, ioc->used, ioc->needed);
10129 	(*cnt) = HANG_FIELD_IOCTL_RESP_TIMEOUT_CNT;
10130 
10131 	clear_debug_dump_time(dhd->debug_dump_time_hang_str);
10132 
10133 	/* Access ioc->buf only if the ioc->len is more than 4 bytes */
10134 	ioc_dwlen = (uint32)(ioc->len / sizeof(uint32));
10135 	if (ioc_dwlen > 0) {
10136 		const uint32 *ioc_buf = (const uint32 *)ioc->buf;
10137 
10138 		remain_len = VENDOR_SEND_HANG_EXT_INFO_LEN - bytes_written;
10139 		GCC_DIAGNOSTIC_PUSH_SUPPRESS_CAST();
10140 		bytes_written += scnprintf(&dest[bytes_written], remain_len,
10141 			"%08x", *(uint32 *)(ioc_buf++));
10142 		GCC_DIAGNOSTIC_POP();
10143 		(*cnt)++;
10144 		if ((*cnt) >= HANG_FIELD_CNT_MAX) {
10145 			return;
10146 		}
10147 
10148 		for (i = 1; i < ioc_dwlen && *cnt <= HANG_FIELD_CNT_MAX;
10149 			i++, (*cnt)++) {
10150 			remain_len = VENDOR_SEND_HANG_EXT_INFO_LEN - bytes_written;
10151 			GCC_DIAGNOSTIC_PUSH_SUPPRESS_CAST();
10152 			bytes_written += scnprintf(&dest[bytes_written], remain_len, "%c%08x",
10153 				HANG_RAW_DEL, *(uint32 *)(ioc_buf++));
10154 			GCC_DIAGNOSTIC_POP();
10155 		}
10156 	}
10157 
10158 	DHD_INFO(("%s hang info len: %d data: %s\n",
10159 		__FUNCTION__, (int)strlen(dhd->hang_info), dhd->hang_info));
10160 }
10161 
10162 #endif /* WL_CFGVENDOR_SEND_HANG_EVENT */
10163 
10164 #if defined(DHD_H2D_LOG_TIME_SYNC)
10165 /*
10166  * Helper function:
10167  * Used for Dongle console message time syncing with Host printk
10168  */
dhd_h2d_log_time_sync(dhd_pub_t * dhd)10169 void dhd_h2d_log_time_sync(dhd_pub_t *dhd)
10170 {
10171 	uint64 ts;
10172 
10173 	/*
10174 	 * local_clock() returns time in nano seconds.
10175 	 * Dongle understand only milli seconds time.
10176 	 */
10177 	ts = local_clock();
10178 	/* Nano seconds to milli seconds */
10179 	do_div(ts, 1000000);
10180 	if (dhd_wl_ioctl_set_intiovar(dhd,  "rte_timesync", ts, WLC_SET_VAR, TRUE, 0)) {
10181 		DHD_ERROR(("%s rte_timesync **** FAILED ****\n", __FUNCTION__));
10182 		/* Stopping HOST Dongle console time syncing */
10183 		dhd->dhd_rte_time_sync_ms = 0;
10184 	}
10185 }
10186 #endif /* DHD_H2D_LOG_TIME_SYNC */
10187 
10188 #if defined(LINUX) || defined(linux)
10189 /* configuations of ecounters to be enabled by default in FW */
10190 static ecounters_cfg_t ecounters_cfg_tbl[] = {
10191 	/* Global ecounters */
10192 	{ECOUNTERS_STATS_TYPES_FLAG_GLOBAL, 0x0, WL_IFSTATS_XTLV_BUS_PCIE},
10193 	// {ECOUNTERS_STATS_TYPES_FLAG_GLOBAL, 0x0, WL_IFSTATS_XTLV_TX_AMPDU_STATS},
10194 	// {ECOUNTERS_STATS_TYPES_FLAG_GLOBAL, 0x0, WL_IFSTATS_XTLV_RX_AMPDU_STATS},
10195 
10196 	/* Slice specific ecounters */
10197 	{ECOUNTERS_STATS_TYPES_FLAG_SLICE, 0x0, WL_SLICESTATS_XTLV_PERIODIC_STATE},
10198 	{ECOUNTERS_STATS_TYPES_FLAG_SLICE, 0x1, WL_SLICESTATS_XTLV_PERIODIC_STATE},
10199 	{ECOUNTERS_STATS_TYPES_FLAG_SLICE, 0x1, WL_IFSTATS_XTLV_WL_SLICE_BTCOEX},
10200 
10201 	/* Interface specific ecounters */
10202 	{ECOUNTERS_STATS_TYPES_FLAG_IFACE, 0x0, WL_IFSTATS_XTLV_IF_PERIODIC_STATE},
10203 	{ECOUNTERS_STATS_TYPES_FLAG_IFACE, 0x0, WL_IFSTATS_XTLV_GENERIC},
10204 	{ECOUNTERS_STATS_TYPES_FLAG_IFACE, 0x0, WL_IFSTATS_XTLV_INFRA_SPECIFIC},
10205 	{ECOUNTERS_STATS_TYPES_FLAG_IFACE, 0x0, WL_IFSTATS_XTLV_MGT_CNT},
10206 
10207 	/* secondary interface */
10208 	/* XXX REMOVE for temporal, will be enabled after decision
10209 	{ECOUNTERS_STATS_TYPES_FLAG_IFACE, 0x1, WL_IFSTATS_XTLV_IF_PERIODIC_STATE},
10210 	{ECOUNTERS_STATS_TYPES_FLAG_IFACE, 0x1, WL_IFSTATS_XTLV_GENERIC},
10211 	{ECOUNTERS_STATS_TYPES_FLAG_IFACE, 0x1, WL_IFSTATS_XTLV_INFRA_SPECIFIC},
10212 	{ECOUNTERS_STATS_TYPES_FLAG_IFACE, 0x1, WL_IFSTATS_XTLV_MGT_CNT},
10213 	*/
10214 };
10215 
10216 /* XXX: Same event id shall be defined in consecutive order in the below table */
10217 static event_ecounters_cfg_t event_ecounters_cfg_tbl[] = {
10218 	/* Interface specific event ecounters */
10219 	{WLC_E_DEAUTH_IND, ECOUNTERS_STATS_TYPES_FLAG_IFACE, 0x0, WL_IFSTATS_XTLV_IF_EVENT_STATS},
10220 };
10221 
10222 /* Accepts an argument to -s, -g or -f and creates an XTLV */
10223 int
dhd_create_ecounters_params(dhd_pub_t * dhd,uint16 type,uint16 if_slice_idx,uint16 stats_rep,uint8 ** xtlv)10224 dhd_create_ecounters_params(dhd_pub_t *dhd, uint16 type, uint16 if_slice_idx,
10225 	uint16 stats_rep, uint8 **xtlv)
10226 {
10227 	uint8 *req_xtlv = NULL;
10228 	ecounters_stats_types_report_req_t *req;
10229 	bcm_xtlvbuf_t xtlvbuf, container_xtlvbuf;
10230 	ecountersv2_xtlv_list_elt_t temp;
10231 	uint16 xtlv_len = 0, total_len = 0;
10232 	int rc = BCME_OK;
10233 
10234 	/* fill in the stat type XTLV. For now there is no explicit TLV for the stat type. */
10235 	temp.id = stats_rep;
10236 	temp.len = 0;
10237 
10238 	/* Hence len/data = 0/NULL */
10239 	xtlv_len += temp.len + BCM_XTLV_HDR_SIZE;
10240 
10241 	/* Total length of the container */
10242 	total_len = BCM_XTLV_HDR_SIZE +
10243 		OFFSETOF(ecounters_stats_types_report_req_t, stats_types_req) + xtlv_len;
10244 
10245 	/* Now allocate a structure for the entire request */
10246 	if ((req_xtlv = (uint8 *)MALLOCZ(dhd->osh, total_len)) == NULL) {
10247 		rc = BCME_NOMEM;
10248 		goto fail;
10249 	}
10250 
10251 	/* container XTLV context */
10252 	bcm_xtlv_buf_init(&container_xtlvbuf, (uint8 *)req_xtlv, total_len,
10253 		BCM_XTLV_OPTION_ALIGN32);
10254 
10255 	/* Fill other XTLVs in the container. Leave space for XTLV headers */
10256 	req = (ecounters_stats_types_report_req_t *)(req_xtlv + BCM_XTLV_HDR_SIZE);
10257 	req->flags = type;
10258 	if (type == ECOUNTERS_STATS_TYPES_FLAG_SLICE) {
10259 		req->slice_mask = 0x1 << if_slice_idx;
10260 	} else if (type == ECOUNTERS_STATS_TYPES_FLAG_IFACE) {
10261 		req->if_index = if_slice_idx;
10262 	}
10263 
10264 	/* Fill remaining XTLVs */
10265 	bcm_xtlv_buf_init(&xtlvbuf, (uint8*) req->stats_types_req, xtlv_len,
10266 		BCM_XTLV_OPTION_ALIGN32);
10267 	if (bcm_xtlv_put_data(&xtlvbuf, temp.id, NULL, temp.len)) {
10268 		DHD_ERROR(("Error creating XTLV for requested stats type = %d\n", temp.id));
10269 		rc = BCME_ERROR;
10270 		goto fail;
10271 	}
10272 
10273 	/* fill the top level container and get done with the XTLV container */
10274 	rc = bcm_xtlv_put_data(&container_xtlvbuf, WL_ECOUNTERS_XTLV_REPORT_REQ, NULL,
10275 		bcm_xtlv_buf_len(&xtlvbuf) + OFFSETOF(ecounters_stats_types_report_req_t,
10276 		stats_types_req));
10277 
10278 	if (rc) {
10279 		DHD_ERROR(("Error creating parent XTLV for type = %d\n", req->flags));
10280 		goto fail;
10281 	}
10282 
10283 fail:
10284 	if (rc && req_xtlv) {
10285 		MFREE(dhd->osh, req_xtlv, total_len);
10286 		req_xtlv = NULL;
10287 	}
10288 
10289 	/* update the xtlv pointer */
10290 	*xtlv = req_xtlv;
10291 	return rc;
10292 }
10293 
10294 static int
dhd_ecounter_autoconfig(dhd_pub_t * dhd)10295 dhd_ecounter_autoconfig(dhd_pub_t *dhd)
10296 {
10297 	int rc = BCME_OK;
10298 	uint32 buf;
10299 	rc = dhd_iovar(dhd, 0, "ecounters_autoconfig", NULL, 0, (char *)&buf, sizeof(buf), FALSE);
10300 
10301 	if (rc != BCME_OK) {
10302 
10303 		if (rc != BCME_UNSUPPORTED) {
10304 			rc = BCME_OK;
10305 			DHD_ERROR(("%s Ecounter autoconfig in fw failed : %d\n", __FUNCTION__, rc));
10306 		} else {
10307 			DHD_ERROR(("%s Ecounter autoconfig in FW not supported\n", __FUNCTION__));
10308 		}
10309 	}
10310 
10311 	return rc;
10312 }
10313 
10314 int
dhd_ecounter_configure(dhd_pub_t * dhd,bool enable)10315 dhd_ecounter_configure(dhd_pub_t *dhd, bool enable)
10316 {
10317 	int rc = BCME_OK;
10318 	if (enable) {
10319 		if (dhd_ecounter_autoconfig(dhd) != BCME_OK) {
10320 			if ((rc = dhd_start_ecounters(dhd)) != BCME_OK) {
10321 				DHD_ERROR(("%s Ecounters start failed\n", __FUNCTION__));
10322 			} else if ((rc = dhd_start_event_ecounters(dhd)) != BCME_OK) {
10323 				DHD_ERROR(("%s Event_Ecounters start failed\n", __FUNCTION__));
10324 			}
10325 		}
10326 	} else {
10327 		if ((rc = dhd_stop_ecounters(dhd)) != BCME_OK) {
10328 			DHD_ERROR(("%s Ecounters stop failed\n", __FUNCTION__));
10329 		} else if ((rc = dhd_stop_event_ecounters(dhd)) != BCME_OK) {
10330 			DHD_ERROR(("%s Event_Ecounters stop failed\n", __FUNCTION__));
10331 		}
10332 	}
10333 	return rc;
10334 }
10335 
10336 int
dhd_start_ecounters(dhd_pub_t * dhd)10337 dhd_start_ecounters(dhd_pub_t *dhd)
10338 {
10339 	uint8 i = 0;
10340 	uint8 *start_ptr;
10341 	int rc = BCME_OK;
10342 	bcm_xtlv_t *elt;
10343 	ecounters_config_request_v2_t *req = NULL;
10344 	ecountersv2_processed_xtlv_list_elt *list_elt, *tail = NULL;
10345 	ecountersv2_processed_xtlv_list_elt *processed_containers_list = NULL;
10346 	uint16 total_processed_containers_len = 0;
10347 
10348 	for (i = 0; i < ARRAYSIZE(ecounters_cfg_tbl); i++) {
10349 		ecounters_cfg_t *ecounter_stat = &ecounters_cfg_tbl[i];
10350 
10351 		if ((list_elt = (ecountersv2_processed_xtlv_list_elt *)
10352 			MALLOCZ(dhd->osh, sizeof(*list_elt))) == NULL) {
10353 			DHD_ERROR(("Ecounters v2: No memory to process\n"));
10354 			goto fail;
10355 		}
10356 
10357 		rc = dhd_create_ecounters_params(dhd, ecounter_stat->type,
10358 			ecounter_stat->if_slice_idx, ecounter_stat->stats_rep, &list_elt->data);
10359 
10360 		if (rc) {
10361 			DHD_ERROR(("Ecounters v2: Could not process: stat: %d return code: %d\n",
10362 				ecounter_stat->stats_rep, rc));
10363 
10364 			/* Free allocated memory and go to fail to release any memories allocated
10365 			 * in previous iterations. Note that list_elt->data gets populated in
10366 			 * dhd_create_ecounters_params() and gets freed there itself.
10367 			 */
10368 			MFREE(dhd->osh, list_elt, sizeof(*list_elt));
10369 			list_elt = NULL;
10370 			goto fail;
10371 		}
10372 		elt = (bcm_xtlv_t *) list_elt->data;
10373 
10374 		/* Put the elements in the order they are processed */
10375 		if (processed_containers_list == NULL) {
10376 			processed_containers_list = list_elt;
10377 		} else {
10378 			tail->next = list_elt;
10379 		}
10380 		tail = list_elt;
10381 		/* Size of the XTLV returned */
10382 		total_processed_containers_len += BCM_XTLV_LEN(elt) + BCM_XTLV_HDR_SIZE;
10383 	}
10384 
10385 	/* Now create ecounters config request with totallength */
10386 	req = (ecounters_config_request_v2_t *)MALLOCZ(dhd->osh, sizeof(*req) +
10387 		total_processed_containers_len);
10388 
10389 	if (req == NULL) {
10390 		rc = BCME_NOMEM;
10391 		goto fail;
10392 	}
10393 
10394 	req->version = ECOUNTERS_VERSION_2;
10395 	req->logset = EVENT_LOG_SET_ECOUNTERS;
10396 	req->reporting_period = ECOUNTERS_DEFAULT_PERIOD;
10397 	req->num_reports = ECOUNTERS_NUM_REPORTS;
10398 	req->len = total_processed_containers_len +
10399 		OFFSETOF(ecounters_config_request_v2_t, ecounters_xtlvs);
10400 
10401 	/* Copy config */
10402 	start_ptr = req->ecounters_xtlvs;
10403 
10404 	/* Now go element by element in the list */
10405 	while (processed_containers_list) {
10406 		list_elt = processed_containers_list;
10407 
10408 		elt = (bcm_xtlv_t *)list_elt->data;
10409 
10410 		memcpy(start_ptr, list_elt->data, BCM_XTLV_LEN(elt) + BCM_XTLV_HDR_SIZE);
10411 		start_ptr += (size_t)(BCM_XTLV_LEN(elt) + BCM_XTLV_HDR_SIZE);
10412 		processed_containers_list = processed_containers_list->next;
10413 
10414 		/* Free allocated memories */
10415 		MFREE(dhd->osh, elt, elt->len + BCM_XTLV_HDR_SIZE);
10416 		MFREE(dhd->osh, list_elt, sizeof(*list_elt));
10417 	}
10418 
10419 	if ((rc = dhd_iovar(dhd, 0, "ecounters", (char *)req, req->len, NULL, 0, TRUE)) < 0) {
10420 		DHD_ERROR(("failed to start ecounters\n"));
10421 	}
10422 
10423 fail:
10424 	if (req) {
10425 		MFREE(dhd->osh, req, sizeof(*req) + total_processed_containers_len);
10426 	}
10427 
10428 	/* Now go element by element in the list */
10429 	while (processed_containers_list) {
10430 		list_elt = processed_containers_list;
10431 		elt = (bcm_xtlv_t *)list_elt->data;
10432 		processed_containers_list = processed_containers_list->next;
10433 
10434 		/* Free allocated memories */
10435 		MFREE(dhd->osh, elt, elt->len + BCM_XTLV_HDR_SIZE);
10436 		MFREE(dhd->osh, list_elt, sizeof(*list_elt));
10437 	}
10438 	return rc;
10439 }
10440 
10441 int
dhd_stop_ecounters(dhd_pub_t * dhd)10442 dhd_stop_ecounters(dhd_pub_t *dhd)
10443 {
10444 	int rc = BCME_OK;
10445 	ecounters_config_request_v2_t *req;
10446 
10447 	/* Now create ecounters config request with totallength */
10448 	req = (ecounters_config_request_v2_t *)MALLOCZ(dhd->osh, sizeof(*req));
10449 
10450 	if (req == NULL) {
10451 		rc = BCME_NOMEM;
10452 		goto fail;
10453 	}
10454 
10455 	req->version = ECOUNTERS_VERSION_2;
10456 	req->len = OFFSETOF(ecounters_config_request_v2_t, ecounters_xtlvs);
10457 
10458 	if ((rc = dhd_iovar(dhd, 0, "ecounters", (char *)req, req->len, NULL, 0, TRUE)) < 0) {
10459 		DHD_ERROR(("failed to stop ecounters\n"));
10460 	}
10461 
10462 fail:
10463 	if (req) {
10464 		MFREE(dhd->osh, req, sizeof(*req));
10465 	}
10466 	return rc;
10467 }
10468 
10469 /* configured event_id_array for event ecounters */
10470 typedef struct event_id_array {
10471 	uint8	event_id;
10472 	uint8	str_idx;
10473 } event_id_array_t;
10474 
10475 /* get event id array only from event_ecounters_cfg_tbl[] */
__dhd_event_ecounters_get_event_id_array(event_id_array_t * event_array)10476 static inline int __dhd_event_ecounters_get_event_id_array(event_id_array_t *event_array)
10477 {
10478 	uint8 i;
10479 	uint8 idx = 0;
10480 	int32 prev_evt_id = -1;
10481 
10482 	for (i = 0; i < (uint8)ARRAYSIZE(event_ecounters_cfg_tbl); i++) {
10483 		if (prev_evt_id != event_ecounters_cfg_tbl[i].event_id) {
10484 			if (prev_evt_id >= 0)
10485 				idx++;
10486 			event_array[idx].event_id = event_ecounters_cfg_tbl[i].event_id;
10487 			event_array[idx].str_idx = i;
10488 		}
10489 		prev_evt_id = event_ecounters_cfg_tbl[i].event_id;
10490 	}
10491 	return idx;
10492 }
10493 
10494 /* One event id has limit xtlv num to request based on wl_ifstats_xtlv_id * 2 interface */
10495 #define ECNTRS_MAX_XTLV_NUM (31 * 2)
10496 
10497 int
dhd_start_event_ecounters(dhd_pub_t * dhd)10498 dhd_start_event_ecounters(dhd_pub_t *dhd)
10499 {
10500 	uint8 i, j = 0;
10501 	uint8 event_id_cnt = 0;
10502 	uint16 processed_containers_len = 0;
10503 	uint16 max_xtlv_len = 0;
10504 	int rc = BCME_OK;
10505 	uint8 *ptr;
10506 	uint8 *data;
10507 	event_id_array_t *id_array;
10508 	bcm_xtlv_t *elt = NULL;
10509 	event_ecounters_config_request_v2_t *req = NULL;
10510 
10511 	/* XXX: the size of id_array is limited by the size of event_ecounters_cfg_tbl */
10512 	id_array = (event_id_array_t *)MALLOCZ(dhd->osh, sizeof(event_id_array_t) *
10513 		ARRAYSIZE(event_ecounters_cfg_tbl));
10514 
10515 	if (id_array == NULL) {
10516 		rc = BCME_NOMEM;
10517 		goto fail;
10518 	}
10519 	event_id_cnt = __dhd_event_ecounters_get_event_id_array(id_array);
10520 
10521 	max_xtlv_len = ((BCM_XTLV_HDR_SIZE +
10522 		OFFSETOF(event_ecounters_config_request_v2_t, ecounters_xtlvs)) *
10523 		ECNTRS_MAX_XTLV_NUM);
10524 
10525 	/* Now create ecounters config request with max allowed length */
10526 	req = (event_ecounters_config_request_v2_t *)MALLOCZ(dhd->osh,
10527 		sizeof(event_ecounters_config_request_v2_t *) + max_xtlv_len);
10528 
10529 	if (req == NULL) {
10530 		rc = BCME_NOMEM;
10531 		goto fail;
10532 	}
10533 
10534 	for (i = 0; i <= event_id_cnt; i++) {
10535 		/* req initialization by event id */
10536 		req->version = ECOUNTERS_VERSION_2;
10537 		req->logset = EVENT_LOG_SET_ECOUNTERS;
10538 		req->event_id = id_array[i].event_id;
10539 		req->flags = EVENT_ECOUNTERS_FLAGS_ADD;
10540 		req->len = 0;
10541 		processed_containers_len = 0;
10542 
10543 		/* Copy config */
10544 		ptr = req->ecounters_xtlvs;
10545 
10546 		for (j = id_array[i].str_idx; j < (uint8)ARRAYSIZE(event_ecounters_cfg_tbl); j++) {
10547 			event_ecounters_cfg_t *event_ecounter_stat = &event_ecounters_cfg_tbl[j];
10548 			if (id_array[i].event_id != event_ecounter_stat->event_id)
10549 				break;
10550 
10551 			rc = dhd_create_ecounters_params(dhd, event_ecounter_stat->type,
10552 				event_ecounter_stat->if_slice_idx, event_ecounter_stat->stats_rep,
10553 				&data);
10554 
10555 			if (rc) {
10556 				DHD_ERROR(("%s: Could not process: stat: %d return code: %d\n",
10557 					__FUNCTION__, event_ecounter_stat->stats_rep, rc));
10558 				goto fail;
10559 			}
10560 
10561 			elt = (bcm_xtlv_t *)data;
10562 
10563 			memcpy(ptr, elt, BCM_XTLV_LEN(elt) + BCM_XTLV_HDR_SIZE);
10564 			ptr += (size_t)(BCM_XTLV_LEN(elt) + BCM_XTLV_HDR_SIZE);
10565 			processed_containers_len += BCM_XTLV_LEN(elt) + BCM_XTLV_HDR_SIZE;
10566 
10567 			/* Free allocated memories alloced by dhd_create_ecounters_params */
10568 			MFREE(dhd->osh, elt, elt->len + BCM_XTLV_HDR_SIZE);
10569 
10570 			if (processed_containers_len > max_xtlv_len) {
10571 				DHD_ERROR(("%s XTLV NUM IS OVERFLOWED THAN ALLOWED!!\n",
10572 					__FUNCTION__));
10573 				rc = BCME_BADLEN;
10574 				goto fail;
10575 			}
10576 		}
10577 
10578 		req->len = processed_containers_len +
10579 			OFFSETOF(event_ecounters_config_request_v2_t, ecounters_xtlvs);
10580 
10581 		DHD_INFO(("%s req version %d logset %d event_id %d flags %d len %d\n",
10582 			__FUNCTION__, req->version, req->logset, req->event_id,
10583 			req->flags, req->len));
10584 
10585 		rc = dhd_iovar(dhd, 0, "event_ecounters", (char *)req, req->len, NULL, 0, TRUE);
10586 
10587 		if (rc < 0) {
10588 			DHD_ERROR(("failed to start event_ecounters(event id %d) with rc %d\n",
10589 				req->event_id, rc));
10590 			goto fail;
10591 		}
10592 	}
10593 
10594 fail:
10595 	/* Free allocated memories */
10596 	if (req) {
10597 		MFREE(dhd->osh, req, sizeof(event_ecounters_config_request_v2_t *) + max_xtlv_len);
10598 	}
10599 	if (id_array) {
10600 		MFREE(dhd->osh, id_array, sizeof(event_id_array_t) *
10601 			ARRAYSIZE(event_ecounters_cfg_tbl));
10602 	}
10603 
10604 	return rc;
10605 }
10606 
10607 int
dhd_stop_event_ecounters(dhd_pub_t * dhd)10608 dhd_stop_event_ecounters(dhd_pub_t *dhd)
10609 {
10610 	int rc = BCME_OK;
10611 	event_ecounters_config_request_v2_t *req;
10612 
10613 	/* Now create ecounters config request with totallength */
10614 	req = (event_ecounters_config_request_v2_t *)MALLOCZ(dhd->osh, sizeof(*req));
10615 
10616 	if (req == NULL) {
10617 		rc = BCME_NOMEM;
10618 		goto fail;
10619 	}
10620 
10621 	req->version = ECOUNTERS_VERSION_2;
10622 	req->flags = EVENT_ECOUNTERS_FLAGS_DEL_ALL;
10623 	req->len = OFFSETOF(event_ecounters_config_request_v2_t, ecounters_xtlvs);
10624 
10625 	if ((rc = dhd_iovar(dhd, 0, "event_ecounters", (char *)req, req->len, NULL, 0, TRUE)) < 0) {
10626 		DHD_ERROR(("failed to stop event_ecounters\n"));
10627 	}
10628 
10629 fail:
10630 	if (req) {
10631 		MFREE(dhd->osh, req, sizeof(*req));
10632 	}
10633 	return rc;
10634 }
10635 #ifdef DHD_LOG_DUMP
10636 int
dhd_dump_debug_ring(dhd_pub_t * dhdp,void * ring_ptr,const void * user_buf,log_dump_section_hdr_t * sec_hdr,char * text_hdr,int buflen,uint32 sec_type)10637 dhd_dump_debug_ring(dhd_pub_t *dhdp, void *ring_ptr, const void *user_buf,
10638 		log_dump_section_hdr_t *sec_hdr,
10639 		char *text_hdr, int buflen, uint32 sec_type)
10640 {
10641 	uint32 rlen = 0;
10642 	uint32 data_len = 0;
10643 	void *data = NULL;
10644 	unsigned long flags = 0;
10645 	int ret = 0;
10646 	dhd_dbg_ring_t *ring = (dhd_dbg_ring_t *)ring_ptr;
10647 	int pos = 0;
10648 	int fpos_sechdr = 0;
10649 
10650 	if (!dhdp || !ring || !user_buf || !sec_hdr || !text_hdr) {
10651 		return BCME_BADARG;
10652 	}
10653 	/* do not allow further writes to the ring
10654 	 * till we flush it
10655 	 */
10656 	DHD_DBG_RING_LOCK(ring->lock, flags);
10657 	ring->state = RING_SUSPEND;
10658 	DHD_DBG_RING_UNLOCK(ring->lock, flags);
10659 
10660 	if (dhdp->concise_dbg_buf) {
10661 		/* re-use concise debug buffer temporarily
10662 		 * to pull ring data, to write
10663 		 * record by record to file
10664 		 */
10665 		data_len = CONCISE_DUMP_BUFLEN;
10666 		data = dhdp->concise_dbg_buf;
10667 		ret = dhd_export_debug_data(text_hdr, NULL, user_buf, strlen(text_hdr), &pos);
10668 		/* write the section header now with zero length,
10669 		 * once the correct length is found out, update
10670 		 * it later
10671 		 */
10672 		fpos_sechdr = pos;
10673 		sec_hdr->type = sec_type;
10674 		sec_hdr->length = 0;
10675 		ret = dhd_export_debug_data((char *)sec_hdr, NULL, user_buf,
10676 			sizeof(*sec_hdr), &pos);
10677 		do {
10678 			rlen = dhd_dbg_ring_pull_single(ring, data, data_len, TRUE);
10679 			if (rlen > 0) {
10680 				/* write the log */
10681 				ret = dhd_export_debug_data(data, NULL, user_buf, rlen, &pos);
10682 			}
10683 			DHD_DBGIF(("%s: rlen : %d\n", __FUNCTION__, rlen));
10684 		} while ((rlen > 0));
10685 		/* now update the section header length in the file */
10686 		/* Complete ring size is dumped by HAL, hence updating length to ring size */
10687 		sec_hdr->length = ring->ring_size;
10688 		ret = dhd_export_debug_data((char *)sec_hdr, NULL, user_buf,
10689 			sizeof(*sec_hdr), &fpos_sechdr);
10690 	} else {
10691 		DHD_ERROR(("%s: No concise buffer available !\n", __FUNCTION__));
10692 	}
10693 	DHD_DBG_RING_LOCK(ring->lock, flags);
10694 	ring->state = RING_ACTIVE;
10695 	/* Resetting both read and write pointer,
10696 	 * since all items are read.
10697 	 */
10698 	ring->rp = ring->wp = 0;
10699 	DHD_DBG_RING_UNLOCK(ring->lock, flags);
10700 
10701 	return ret;
10702 }
10703 
10704 int
dhd_log_dump_ring_to_file(dhd_pub_t * dhdp,void * ring_ptr,void * file,unsigned long * file_posn,log_dump_section_hdr_t * sec_hdr,char * text_hdr,uint32 sec_type)10705 dhd_log_dump_ring_to_file(dhd_pub_t *dhdp, void *ring_ptr, void *file,
10706 		unsigned long *file_posn, log_dump_section_hdr_t *sec_hdr,
10707 		char *text_hdr, uint32 sec_type)
10708 {
10709 	uint32 rlen = 0;
10710 	uint32 data_len = 0, total_len = 0;
10711 	void *data = NULL;
10712 	unsigned long fpos_sechdr = 0;
10713 	unsigned long flags = 0;
10714 	int ret = 0;
10715 	dhd_dbg_ring_t *ring = (dhd_dbg_ring_t *)ring_ptr;
10716 
10717 	if (!dhdp || !ring || !file || !sec_hdr ||
10718 		!file_posn || !text_hdr)
10719 		return BCME_BADARG;
10720 
10721 	/* do not allow further writes to the ring
10722 	 * till we flush it
10723 	 */
10724 	DHD_DBG_RING_LOCK(ring->lock, flags);
10725 	ring->state = RING_SUSPEND;
10726 	DHD_DBG_RING_UNLOCK(ring->lock, flags);
10727 
10728 	if (dhdp->concise_dbg_buf) {
10729 		/* re-use concise debug buffer temporarily
10730 		 * to pull ring data, to write
10731 		 * record by record to file
10732 		 */
10733 		data_len = CONCISE_DUMP_BUFLEN;
10734 		data = dhdp->concise_dbg_buf;
10735 		dhd_os_write_file_posn(file, file_posn, text_hdr,
10736 				strlen(text_hdr));
10737 		/* write the section header now with zero length,
10738 		 * once the correct length is found out, update
10739 		 * it later
10740 		 */
10741 		dhd_init_sec_hdr(sec_hdr);
10742 		fpos_sechdr = *file_posn;
10743 		sec_hdr->type = sec_type;
10744 		sec_hdr->length = 0;
10745 		dhd_os_write_file_posn(file, file_posn, (char *)sec_hdr,
10746 				sizeof(*sec_hdr));
10747 		do {
10748 			rlen = dhd_dbg_ring_pull_single(ring, data, data_len, TRUE);
10749 			if (rlen > 0) {
10750 				/* write the log */
10751 				ret = dhd_os_write_file_posn(file, file_posn, data, rlen);
10752 				if (ret < 0) {
10753 					DHD_ERROR(("%s: write file error !\n", __FUNCTION__));
10754 					DHD_DBG_RING_LOCK(ring->lock, flags);
10755 					ring->state = RING_ACTIVE;
10756 					DHD_DBG_RING_UNLOCK(ring->lock, flags);
10757 					return BCME_ERROR;
10758 				}
10759 			}
10760 			total_len += rlen;
10761 		} while (rlen > 0);
10762 		/* now update the section header length in the file */
10763 		sec_hdr->length = total_len;
10764 		dhd_os_write_file_posn(file, &fpos_sechdr, (char *)sec_hdr, sizeof(*sec_hdr));
10765 	} else {
10766 		DHD_ERROR(("%s: No concise buffer available !\n", __FUNCTION__));
10767 	}
10768 
10769 	DHD_DBG_RING_LOCK(ring->lock, flags);
10770 	ring->state = RING_ACTIVE;
10771 	/* Resetting both read and write pointer,
10772 	 * since all items are read.
10773 	 */
10774 	ring->rp = ring->wp = 0;
10775 	DHD_DBG_RING_UNLOCK(ring->lock, flags);
10776 	return BCME_OK;
10777 }
10778 
10779 /* logdump cookie */
10780 #define MAX_LOGUDMP_COOKIE_CNT	10u
10781 #define LOGDUMP_COOKIE_STR_LEN	50u
10782 int
dhd_logdump_cookie_init(dhd_pub_t * dhdp,uint8 * buf,uint32 buf_size)10783 dhd_logdump_cookie_init(dhd_pub_t *dhdp, uint8 *buf, uint32 buf_size)
10784 {
10785 	uint32 ring_size;
10786 
10787 	if (!dhdp || !buf) {
10788 		DHD_ERROR(("INVALID PTR: dhdp:%p buf:%p\n", dhdp, buf));
10789 		return BCME_ERROR;
10790 	}
10791 
10792 	ring_size = dhd_ring_get_hdr_size() + LOGDUMP_COOKIE_STR_LEN * MAX_LOGUDMP_COOKIE_CNT;
10793 	if (buf_size < ring_size) {
10794 		DHD_ERROR(("BUF SIZE IS TO SHORT: req:%d buf_size:%d\n",
10795 			ring_size, buf_size));
10796 		return BCME_ERROR;
10797 	}
10798 
10799 	dhdp->logdump_cookie = dhd_ring_init(dhdp, buf, buf_size,
10800 		LOGDUMP_COOKIE_STR_LEN, MAX_LOGUDMP_COOKIE_CNT,
10801 		DHD_RING_TYPE_FIXED);
10802 	if (!dhdp->logdump_cookie) {
10803 		DHD_ERROR(("FAIL TO INIT COOKIE RING\n"));
10804 		return BCME_ERROR;
10805 	}
10806 
10807 	return BCME_OK;
10808 }
10809 
10810 void
dhd_logdump_cookie_deinit(dhd_pub_t * dhdp)10811 dhd_logdump_cookie_deinit(dhd_pub_t *dhdp)
10812 {
10813 	if (!dhdp) {
10814 		return;
10815 	}
10816 	if (dhdp->logdump_cookie) {
10817 		dhd_ring_deinit(dhdp, dhdp->logdump_cookie);
10818 	}
10819 
10820 	return;
10821 }
10822 
10823 #ifdef DHD_TX_PROFILE
10824 int
dhd_tx_profile_detach(dhd_pub_t * dhdp)10825 dhd_tx_profile_detach(dhd_pub_t *dhdp)
10826 {
10827 	int result = BCME_ERROR;
10828 
10829 	if (dhdp != NULL && dhdp->protocol_filters != NULL) {
10830 		MFREE(dhdp->osh, dhdp->protocol_filters, DHD_MAX_PROFILES *
10831 				sizeof(*(dhdp->protocol_filters)));
10832 		dhdp->protocol_filters = NULL;
10833 
10834 		result = BCME_OK;
10835 	}
10836 
10837 	return result;
10838 }
10839 
10840 int
dhd_tx_profile_attach(dhd_pub_t * dhdp)10841 dhd_tx_profile_attach(dhd_pub_t *dhdp)
10842 {
10843 	int result = BCME_ERROR;
10844 
10845 	if (dhdp != NULL) {
10846 		dhdp->protocol_filters = (dhd_tx_profile_protocol_t*)MALLOCZ(dhdp->osh,
10847 				DHD_MAX_PROFILES * sizeof(*(dhdp->protocol_filters)));
10848 
10849 		if (dhdp->protocol_filters != NULL) {
10850 			result = BCME_OK;
10851 		}
10852 	}
10853 
10854 	if (result != BCME_OK) {
10855 		DHD_ERROR(("%s:\tMALLOC of tx profile protocol filters failed\n",
10856 			__FUNCTION__));
10857 	}
10858 
10859 	return result;
10860 }
10861 #endif /* defined(DHD_TX_PROFILE) */
10862 
10863 void
dhd_logdump_cookie_save(dhd_pub_t * dhdp,char * cookie,char * type)10864 dhd_logdump_cookie_save(dhd_pub_t *dhdp, char *cookie, char *type)
10865 {
10866 	char *ptr;
10867 
10868 	if (!dhdp || !cookie || !type || !dhdp->logdump_cookie) {
10869 		DHD_ERROR(("%s: At least one buffer ptr is NULL dhdp=%p cookie=%p"
10870 			" type = %p, cookie_cfg:%p\n", __FUNCTION__,
10871 			dhdp, cookie, type, dhdp?dhdp->logdump_cookie: NULL));
10872 		return;
10873 	}
10874 	ptr = (char *)dhd_ring_get_empty(dhdp->logdump_cookie);
10875 	if (ptr == NULL) {
10876 		DHD_ERROR(("%s : Skip to save due to locking\n", __FUNCTION__));
10877 		return;
10878 	}
10879 	scnprintf(ptr, LOGDUMP_COOKIE_STR_LEN, "%s: %s\n", type, cookie);
10880 	return;
10881 }
10882 
10883 int
dhd_logdump_cookie_get(dhd_pub_t * dhdp,char * ret_cookie,uint32 buf_size)10884 dhd_logdump_cookie_get(dhd_pub_t *dhdp, char *ret_cookie, uint32 buf_size)
10885 {
10886 	char *ptr;
10887 
10888 	if (!dhdp || !ret_cookie || !dhdp->logdump_cookie) {
10889 		DHD_ERROR(("%s: At least one buffer ptr is NULL dhdp=%p"
10890 			"cookie=%p cookie_cfg:%p\n", __FUNCTION__,
10891 			dhdp, ret_cookie, dhdp?dhdp->logdump_cookie: NULL));
10892 		return BCME_ERROR;
10893 	}
10894 	ptr = (char *)dhd_ring_get_first(dhdp->logdump_cookie);
10895 	if (ptr == NULL) {
10896 		DHD_ERROR(("%s : Skip to save due to locking\n", __FUNCTION__));
10897 		return BCME_ERROR;
10898 	}
10899 	memcpy(ret_cookie, ptr, MIN(buf_size, strlen(ptr)));
10900 	dhd_ring_free_first(dhdp->logdump_cookie);
10901 	return BCME_OK;
10902 }
10903 
10904 int
dhd_logdump_cookie_count(dhd_pub_t * dhdp)10905 dhd_logdump_cookie_count(dhd_pub_t *dhdp)
10906 {
10907 	if (!dhdp || !dhdp->logdump_cookie) {
10908 		DHD_ERROR(("%s: At least one buffer ptr is NULL dhdp=%p cookie=%p\n",
10909 			__FUNCTION__, dhdp, dhdp?dhdp->logdump_cookie: NULL));
10910 		return 0;
10911 	}
10912 	return dhd_ring_get_cur_size(dhdp->logdump_cookie);
10913 }
10914 
10915 static inline int
__dhd_log_dump_cookie_to_file(dhd_pub_t * dhdp,void * fp,const void * user_buf,unsigned long * f_pos,char * buf,uint32 buf_size)10916 __dhd_log_dump_cookie_to_file(
10917 	dhd_pub_t *dhdp, void *fp, const void *user_buf, unsigned long *f_pos,
10918 	char *buf, uint32 buf_size)
10919 {
10920 
10921 	uint32 remain = buf_size;
10922 	int ret = BCME_ERROR;
10923 	char tmp_buf[LOGDUMP_COOKIE_STR_LEN];
10924 	log_dump_section_hdr_t sec_hdr;
10925 	uint32 read_idx;
10926 	uint32 write_idx;
10927 
10928 	read_idx = dhd_ring_get_read_idx(dhdp->logdump_cookie);
10929 	write_idx = dhd_ring_get_write_idx(dhdp->logdump_cookie);
10930 	while (dhd_logdump_cookie_count(dhdp) > 0) {
10931 		memset(tmp_buf, 0, sizeof(tmp_buf));
10932 		ret = dhd_logdump_cookie_get(dhdp, tmp_buf, LOGDUMP_COOKIE_STR_LEN);
10933 		if (ret != BCME_OK) {
10934 			return ret;
10935 		}
10936 		remain -= scnprintf(&buf[buf_size - remain], remain, "%s", tmp_buf);
10937 	}
10938 	dhd_ring_set_read_idx(dhdp->logdump_cookie, read_idx);
10939 	dhd_ring_set_write_idx(dhdp->logdump_cookie, write_idx);
10940 
10941 	ret = dhd_export_debug_data(COOKIE_LOG_HDR, fp, user_buf, strlen(COOKIE_LOG_HDR), f_pos);
10942 	if (ret < 0) {
10943 		DHD_ERROR(("%s : Write file Error for cookie hdr\n", __FUNCTION__));
10944 		return ret;
10945 	}
10946 	sec_hdr.magic = LOG_DUMP_MAGIC;
10947 	sec_hdr.timestamp = local_clock();
10948 	sec_hdr.type = LOG_DUMP_SECTION_COOKIE;
10949 	sec_hdr.length = buf_size - remain;
10950 
10951 	ret = dhd_export_debug_data((char *)&sec_hdr, fp, user_buf, sizeof(sec_hdr), f_pos);
10952 	if (ret < 0) {
10953 		DHD_ERROR(("%s : Write file Error for section hdr\n", __FUNCTION__));
10954 		return ret;
10955 	}
10956 
10957 	ret = dhd_export_debug_data(buf, fp, user_buf, sec_hdr.length, f_pos);
10958 	if (ret < 0) {
10959 		DHD_ERROR(("%s : Write file Error for cookie data\n", __FUNCTION__));
10960 	}
10961 
10962 	return ret;
10963 }
10964 
10965 uint32
dhd_log_dump_cookie_len(dhd_pub_t * dhdp)10966 dhd_log_dump_cookie_len(dhd_pub_t *dhdp)
10967 {
10968 	int len = 0;
10969 	char tmp_buf[LOGDUMP_COOKIE_STR_LEN];
10970 	log_dump_section_hdr_t sec_hdr;
10971 	char *buf = NULL;
10972 	int ret = BCME_ERROR;
10973 	uint32 buf_size = MAX_LOGUDMP_COOKIE_CNT * LOGDUMP_COOKIE_STR_LEN;
10974 	uint32 read_idx;
10975 	uint32 write_idx;
10976 	uint32 remain;
10977 
10978 	remain = buf_size;
10979 
10980 	if (!dhdp || !dhdp->logdump_cookie) {
10981 		DHD_ERROR(("%s At least one ptr is NULL "
10982 			"dhdp = %p cookie %p\n",
10983 			__FUNCTION__, dhdp, dhdp?dhdp->logdump_cookie:NULL));
10984 		goto exit;
10985 	}
10986 
10987 	buf = (char *)MALLOCZ(dhdp->osh, buf_size);
10988 	if (!buf) {
10989 		DHD_ERROR(("%s Fail to malloc buffer\n", __FUNCTION__));
10990 		goto exit;
10991 	}
10992 
10993 	read_idx = dhd_ring_get_read_idx(dhdp->logdump_cookie);
10994 	write_idx = dhd_ring_get_write_idx(dhdp->logdump_cookie);
10995 	while (dhd_logdump_cookie_count(dhdp) > 0) {
10996 		memset(tmp_buf, 0, sizeof(tmp_buf));
10997 		ret = dhd_logdump_cookie_get(dhdp, tmp_buf, LOGDUMP_COOKIE_STR_LEN);
10998 		if (ret != BCME_OK) {
10999 			goto exit;
11000 		}
11001 		remain -= (uint32)strlen(tmp_buf);
11002 	}
11003 	dhd_ring_set_read_idx(dhdp->logdump_cookie, read_idx);
11004 	dhd_ring_set_write_idx(dhdp->logdump_cookie, write_idx);
11005 	len += strlen(COOKIE_LOG_HDR);
11006 	len += sizeof(sec_hdr);
11007 	len += (buf_size - remain);
11008 exit:
11009 	if (buf)
11010 		MFREE(dhdp->osh, buf, buf_size);
11011 	return len;
11012 }
11013 
11014 int
dhd_log_dump_cookie(dhd_pub_t * dhdp,const void * user_buf)11015 dhd_log_dump_cookie(dhd_pub_t *dhdp, const void *user_buf)
11016 {
11017 	int ret = BCME_ERROR;
11018 	char tmp_buf[LOGDUMP_COOKIE_STR_LEN];
11019 	log_dump_section_hdr_t sec_hdr;
11020 	char *buf = NULL;
11021 	uint32 buf_size = MAX_LOGUDMP_COOKIE_CNT * LOGDUMP_COOKIE_STR_LEN;
11022 	int pos = 0;
11023 	uint32 read_idx;
11024 	uint32 write_idx;
11025 	uint32 remain;
11026 
11027 	remain = buf_size;
11028 
11029 	if (!dhdp || !dhdp->logdump_cookie) {
11030 		DHD_ERROR(("%s At least one ptr is NULL "
11031 			"dhdp = %p cookie %p\n",
11032 			__FUNCTION__, dhdp, dhdp?dhdp->logdump_cookie:NULL));
11033 		goto exit;
11034 	}
11035 
11036 	buf = (char *)MALLOCZ(dhdp->osh, buf_size);
11037 	if (!buf) {
11038 		DHD_ERROR(("%s Fail to malloc buffer\n", __FUNCTION__));
11039 		goto exit;
11040 	}
11041 
11042 	read_idx = dhd_ring_get_read_idx(dhdp->logdump_cookie);
11043 	write_idx = dhd_ring_get_write_idx(dhdp->logdump_cookie);
11044 	while (dhd_logdump_cookie_count(dhdp) > 0) {
11045 		memset(tmp_buf, 0, sizeof(tmp_buf));
11046 		ret = dhd_logdump_cookie_get(dhdp, tmp_buf, LOGDUMP_COOKIE_STR_LEN);
11047 		if (ret != BCME_OK) {
11048 			goto exit;
11049 		}
11050 		remain -= scnprintf(&buf[buf_size - remain], remain, "%s", tmp_buf);
11051 	}
11052 	dhd_ring_set_read_idx(dhdp->logdump_cookie, read_idx);
11053 	dhd_ring_set_write_idx(dhdp->logdump_cookie, write_idx);
11054 	ret = dhd_export_debug_data(COOKIE_LOG_HDR, NULL, user_buf, strlen(COOKIE_LOG_HDR), &pos);
11055 	sec_hdr.magic = LOG_DUMP_MAGIC;
11056 	sec_hdr.timestamp = local_clock();
11057 	sec_hdr.type = LOG_DUMP_SECTION_COOKIE;
11058 	sec_hdr.length = buf_size - remain;
11059 	ret = dhd_export_debug_data((char *)&sec_hdr, NULL, user_buf, sizeof(sec_hdr), &pos);
11060 	ret = dhd_export_debug_data(buf, NULL, user_buf, sec_hdr.length, &pos);
11061 exit:
11062 	if (buf)
11063 		MFREE(dhdp->osh, buf, buf_size);
11064 	return ret;
11065 }
11066 
11067 int
dhd_log_dump_cookie_to_file(dhd_pub_t * dhdp,void * fp,const void * user_buf,unsigned long * f_pos)11068 dhd_log_dump_cookie_to_file(dhd_pub_t *dhdp, void *fp, const void *user_buf, unsigned long *f_pos)
11069 {
11070 	char *buf;
11071 	int ret = BCME_ERROR;
11072 	uint32 buf_size = MAX_LOGUDMP_COOKIE_CNT * LOGDUMP_COOKIE_STR_LEN;
11073 
11074 	if (!dhdp || !dhdp->logdump_cookie || (!fp && !user_buf) || !f_pos) {
11075 		DHD_ERROR(("%s At least one ptr is NULL "
11076 			"dhdp = %p cookie %p fp = %p f_pos = %p\n",
11077 			__FUNCTION__, dhdp, dhdp?dhdp->logdump_cookie:NULL, fp, f_pos));
11078 		return ret;
11079 	}
11080 
11081 	buf = (char *)MALLOCZ(dhdp->osh, buf_size);
11082 	if (!buf) {
11083 		DHD_ERROR(("%s Fail to malloc buffer\n", __FUNCTION__));
11084 		return ret;
11085 	}
11086 	ret = __dhd_log_dump_cookie_to_file(dhdp, fp, user_buf, f_pos, buf, buf_size);
11087 	MFREE(dhdp->osh, buf, buf_size);
11088 
11089 	return ret;
11090 }
11091 #endif /* DHD_LOG_DUMP */
11092 #endif /* LINUX || linux */
11093 
11094 #if defined(DISABLE_HE_ENAB) || defined(CUSTOM_CONTROL_HE_ENAB)
11095 int
dhd_control_he_enab(dhd_pub_t * dhd,uint8 he_enab)11096 dhd_control_he_enab(dhd_pub_t * dhd, uint8 he_enab)
11097 {
11098 	int ret = BCME_OK;
11099 	bcm_xtlv_t *pxtlv = NULL;
11100 	uint8 mybuf[DHD_IOVAR_BUF_SIZE];
11101 	uint16 mybuf_len = sizeof(mybuf);
11102 	pxtlv = (bcm_xtlv_t *)mybuf;
11103 
11104 	ret = bcm_pack_xtlv_entry((uint8**)&pxtlv, &mybuf_len, WL_HE_CMD_ENAB, sizeof(he_enab),
11105 			&he_enab, BCM_XTLV_OPTION_ALIGN32);
11106 
11107 	if (ret != BCME_OK) {
11108 		ret = -EINVAL;
11109 		DHD_ERROR(("%s failed to pack he enab, err: %s\n", __FUNCTION__, bcmerrorstr(ret)));
11110 		return ret;
11111 	}
11112 
11113 	ret = dhd_iovar(dhd, 0, "he", (char *)&mybuf, sizeof(mybuf), NULL, 0, TRUE);
11114 	if (ret < 0) {
11115 		DHD_ERROR(("%s he_enab (%d) set failed, err: %s\n",
11116 				__FUNCTION__, he_enab, bcmerrorstr(ret)));
11117 	} else {
11118 		DHD_ERROR(("%s he_enab (%d) set successed\n", __FUNCTION__, he_enab));
11119 	}
11120 
11121 	return ret;
11122 }
11123 #endif /* DISABLE_HE_ENAB || CUSTOM_CONTROL_HE_ENAB */
11124 
11125 #ifdef CONFIG_ROAM_RSSI_LIMIT
11126 int
dhd_roam_rssi_limit_get(dhd_pub_t * dhd,int * lmt2g,int * lmt5g)11127 dhd_roam_rssi_limit_get(dhd_pub_t *dhd, int *lmt2g, int *lmt5g)
11128 {
11129 	wlc_roam_rssi_limit_t *plmt;
11130 	wlc_roam_rssi_lmt_info_v1_t *pinfo;
11131 	int ret = BCME_OK;
11132 	int plmt_len = sizeof(*pinfo) + ROAMRSSI_HDRLEN;
11133 
11134 	plmt = (wlc_roam_rssi_limit_t *)MALLOCZ(dhd->osh, plmt_len);
11135 	if (!plmt) {
11136 		DHD_ERROR(("%s Fail to malloc buffer\n", __FUNCTION__));
11137 		return BCME_NOMEM;
11138 	}
11139 
11140 	/* Get roam rssi limit */
11141 	ret = dhd_iovar(dhd, 0, "roam_rssi_limit", NULL, 0, (char *)plmt, plmt_len, FALSE);
11142 	if (ret < 0) {
11143 		DHD_ERROR(("%s Failed to Get roam_rssi_limit %d\n", __FUNCTION__, ret));
11144 		goto done;
11145 	}
11146 
11147 	if (plmt->ver != WLC_ROAM_RSSI_LMT_VER_1) {
11148 	    ret = BCME_VERSION;
11149 	    goto done;
11150 	}
11151 
11152 	pinfo = (wlc_roam_rssi_lmt_info_v1_t *)plmt->data;
11153 	*lmt2g = (int)pinfo->rssi_limit_2g;
11154 	*lmt5g = (int)pinfo->rssi_limit_5g;
11155 
11156 done:
11157 	if (plmt) {
11158 		MFREE(dhd->osh, plmt, plmt_len);
11159 	}
11160 	return ret;
11161 }
11162 
11163 int
dhd_roam_rssi_limit_set(dhd_pub_t * dhd,int lmt2g,int lmt5g)11164 dhd_roam_rssi_limit_set(dhd_pub_t *dhd, int lmt2g, int lmt5g)
11165 {
11166 	wlc_roam_rssi_limit_t *plmt;
11167 	wlc_roam_rssi_lmt_info_v1_t *pinfo;
11168 	int ret = BCME_OK;
11169 	int plmt_len = sizeof(*pinfo) + ROAMRSSI_HDRLEN;
11170 
11171 	/* Sanity check RSSI limit Value */
11172 	if ((lmt2g < ROAMRSSI_2G_MIN) || (lmt2g > ROAMRSSI_2G_MAX)) {
11173 		DHD_ERROR(("%s Not In Range 2G ROAM RSSI Limit\n", __FUNCTION__));
11174 		return BCME_RANGE;
11175 	}
11176 	if ((lmt2g < ROAMRSSI_5G_MIN) || (lmt2g > ROAMRSSI_5G_MAX)) {
11177 		DHD_ERROR(("%s Not In Range 5G ROAM RSSI Limit\n", __FUNCTION__));
11178 		return BCME_RANGE;
11179 	}
11180 
11181 	plmt = (wlc_roam_rssi_limit_t *)MALLOCZ(dhd->osh, plmt_len);
11182 	if (!plmt) {
11183 		DHD_ERROR(("%s Fail to malloc buffer\n", __FUNCTION__));
11184 		return BCME_NOMEM;
11185 	}
11186 	plmt->ver = WLC_ROAM_RSSI_LMT_VER_1;
11187 	plmt->len = sizeof(*pinfo);
11188 	pinfo = (wlc_roam_rssi_lmt_info_v1_t *)plmt->data;
11189 	pinfo->rssi_limit_2g = (int16)lmt2g;
11190 	pinfo->rssi_limit_5g = (int16)lmt5g;
11191 
11192 	/* Set roam rssi limit */
11193 	ret = dhd_iovar(dhd, 0, "roam_rssi_limit", (char *)plmt, plmt_len, NULL, 0, TRUE);
11194 	if (ret < 0) {
11195 		DHD_ERROR(("%s Failed to Get roam_rssi_limit %d\n", __FUNCTION__, ret));
11196 		goto done;
11197 	}
11198 done:
11199 	if (plmt) {
11200 		MFREE(dhd->osh, plmt, plmt_len);
11201 	}
11202 	return ret;
11203 }
11204 #endif /* CONFIG_ROAM_RSSI_LIMIT */
11205 
11206 #ifdef CONFIG_ROAM_MIN_DELTA
11207 int
dhd_roam_min_delta_get(dhd_pub_t * dhd,uint32 * dt2g,uint32 * dt5g)11208 dhd_roam_min_delta_get(dhd_pub_t *dhd, uint32 *dt2g, uint32 *dt5g)
11209 {
11210 	wlc_roam_min_delta_t *pmin_delta;
11211 	wlc_roam_min_delta_info_v1_t *pmin_delta_info;
11212 	int ret = BCME_OK;
11213 	int plen = sizeof(*pmin_delta_info) + ROAM_MIN_DELTA_HDRLEN;
11214 
11215 	pmin_delta = (wlc_roam_min_delta_t *)MALLOCZ(dhd->osh, plen);
11216 	if (!pmin_delta) {
11217 		DHD_ERROR(("%s Fail to malloc buffer\n", __FUNCTION__));
11218 		return BCME_NOMEM;
11219 	}
11220 
11221 	/* Get Minimum ROAM score delta */
11222 	ret = dhd_iovar(dhd, 0, "roam_min_delta", NULL, 0, (char *)pmin_delta, plen, FALSE);
11223 	if (ret < 0) {
11224 		DHD_ERROR(("%s Failed to Get roam_min_delta %d\n", __FUNCTION__, ret));
11225 		goto done;
11226 	}
11227 
11228 	if (pmin_delta->ver != WLC_ROAM_MIN_DELTA_VER_1) {
11229 		ret = BCME_VERSION;
11230 		goto done;
11231 	}
11232 
11233 	pmin_delta_info = (wlc_roam_min_delta_info_v1_t *)pmin_delta->data;
11234 	*dt2g = (uint32)pmin_delta_info->roam_min_delta_2g;
11235 	*dt5g = (uint32)pmin_delta_info->roam_min_delta_5g;
11236 
11237 done:
11238 	if (pmin_delta) {
11239 		MFREE(dhd->osh, pmin_delta, plen);
11240 	}
11241 	return ret;
11242 }
11243 
11244 int
dhd_roam_min_delta_set(dhd_pub_t * dhd,uint32 dt2g,uint32 dt5g)11245 dhd_roam_min_delta_set(dhd_pub_t *dhd, uint32 dt2g, uint32 dt5g)
11246 {
11247 	wlc_roam_min_delta_t *pmin_delta;
11248 	wlc_roam_min_delta_info_v1_t *pmin_delta_info;
11249 	int ret = BCME_OK;
11250 	int plen = sizeof(*pmin_delta_info) + ROAM_MIN_DELTA_HDRLEN;
11251 
11252 	/* Sanity check Minimum ROAM score delta */
11253 	if ((dt2g > ROAM_MIN_DELTA_MAX) || (dt5g > ROAM_MIN_DELTA_MAX)) {
11254 		DHD_ERROR(("%s Not In Range Minimum ROAM score delta, 2G: %d, 5G: %d\n",
11255 			__FUNCTION__, dt2g, dt5g));
11256 		return BCME_RANGE;
11257 	}
11258 
11259 	pmin_delta = (wlc_roam_min_delta_t *)MALLOCZ(dhd->osh, plen);
11260 	if (!pmin_delta) {
11261 		DHD_ERROR(("%s Fail to malloc buffer\n", __FUNCTION__));
11262 		return BCME_NOMEM;
11263 	}
11264 	pmin_delta->ver = WLC_ROAM_MIN_DELTA_VER_1;
11265 	pmin_delta->len = sizeof(*pmin_delta_info);
11266 	pmin_delta_info = (wlc_roam_min_delta_info_v1_t *)pmin_delta->data;
11267 	pmin_delta_info->roam_min_delta_2g = (uint32)dt2g;
11268 	pmin_delta_info->roam_min_delta_5g = (uint32)dt5g;
11269 
11270 	/* Set Minimum ROAM score delta */
11271 	ret = dhd_iovar(dhd, 0, "roam_min_delta", (char *)pmin_delta, plen, NULL, 0, TRUE);
11272 	if (ret < 0) {
11273 		DHD_ERROR(("%s Failed to Set roam_min_delta %d\n", __FUNCTION__, ret));
11274 		goto done;
11275 	}
11276 done:
11277 	if (pmin_delta) {
11278 		MFREE(dhd->osh, pmin_delta, plen);
11279 	}
11280 	return ret;
11281 }
11282 #endif /* CONFIG_ROAM_MIN_DELTA */
11283 
11284 #ifdef HOST_SFH_LLC
11285 #define SSTLOOKUP(proto) (((proto) == 0x80f3) || ((proto) == 0x8137))
11286 /** Convert Ethernet to 802.3 per 802.1H (use bridge-tunnel if type in SST)
11287  * Note:- This function will overwrite the ethernet header in the pkt
11288  * with a 802.3 ethernet + LLC/SNAP header by utilising the headroom
11289  * in the packet. The pkt data pointer should be pointing to the
11290  * start of the packet (at the ethernet header) when the function is called.
11291  * The pkt data pointer will be pointing to the
11292  * start of the new 802.3 header if the function returns successfully
11293  *
11294  *
11295  * Original Ethernet (header length = 14):
11296  * ----------------------------------------------------------------------------------------
11297  * |                                                     |   DA   |   SA   | T |  Data... |
11298  * ----------------------------------------------------------------------------------------
11299  *                                                            6        6     2
11300  *
11301  * Conversion to 802.3 (header length = 22):
11302  *                     (LLC includes ether_type in last 2 bytes):
11303  * ----------------------------------------------------------------------------------------
11304  * |                                      |   DA   |   SA   | L | LLC/SNAP | T |  Data... |
11305  * ----------------------------------------------------------------------------------------
11306  *                                             6        6     2       6      2
11307  */
11308 int
BCMFASTPATH(dhd_ether_to_8023_hdr)11309 BCMFASTPATH(dhd_ether_to_8023_hdr)(osl_t *osh, struct ether_header *eh, void *p)
11310 {
11311 	struct ether_header *neh;
11312 	struct dot11_llc_snap_header *lsh;
11313 	uint16 plen, ether_type;
11314 
11315 	if (PKTHEADROOM(osh, p) < DOT11_LLC_SNAP_HDR_LEN) {
11316 		DHD_ERROR(("%s: FATAL! not enough pkt headroom !\n", __FUNCTION__));
11317 		ASSERT(0);
11318 		return BCME_BUFTOOSHORT;
11319 	}
11320 
11321 	ether_type = ntoh16(eh->ether_type);
11322 	neh = (struct ether_header *)PKTPUSH(osh, p, DOT11_LLC_SNAP_HDR_LEN);
11323 
11324 	/* 802.3 MAC header */
11325 	eacopy((char*)eh->ether_dhost, (char*)neh->ether_dhost);
11326 	eacopy((char*)eh->ether_shost, (char*)neh->ether_shost);
11327 	plen = (uint16)PKTLEN(osh, p) - ETHER_HDR_LEN;
11328 	neh->ether_type = hton16(plen);
11329 
11330 	/* 802.2 LLC header */
11331 	lsh = (struct dot11_llc_snap_header *)&neh[1];
11332 	lsh->dsap = 0xaa;
11333 	lsh->ssap = 0xaa;
11334 	lsh->ctl = 0x03;
11335 
11336 	/* 802.2 SNAP header Use RFC1042 or bridge-tunnel if type in SST per 802.1H */
11337 	lsh->oui[0] = 0x00;
11338 	lsh->oui[1] = 0x00;
11339 	if (SSTLOOKUP(ether_type))
11340 		lsh->oui[2] = 0xf8;
11341 	else
11342 		lsh->oui[2] = 0x00;
11343 	lsh->type = hton16(ether_type);
11344 
11345 	return BCME_OK;
11346 }
11347 
11348 /** Convert 802.3+LLC to ethernet
11349  * Note:- This function will overwrite the 802.3+LLC hdr in the pkt
11350  * with an ethernet header. The pkt data pointer should be pointing to the
11351  * start of the packet (at the 802.3 header) when the function is called.
11352  * The pkt data pointer will be pointing to the
11353  * start of the ethernet header if the function returns successfully
11354  */
11355 int
BCMFASTPATH(dhd_8023_llc_to_ether_hdr)11356 BCMFASTPATH(dhd_8023_llc_to_ether_hdr)(osl_t *osh, struct ether_header *eh8023, void *p)
11357 {
11358 	struct dot11_llc_snap_header *lsh = NULL;
11359 	uint16 ether_type = 0;
11360 	uint8 *pdata = NULL;
11361 
11362 	if (!p || !eh8023)
11363 		return BCME_BADARG;
11364 
11365 	pdata = PKTDATA(osh, p);
11366 	ether_type = ntoh16(eh8023->ether_type);
11367 	/* ether type in 802.3 hdr for sfh llc host insertion case
11368 	 * contains length, replace it with actual ether type at the
11369 	 * end of the LLC hdr
11370 	 */
11371 	if (ether_type < ETHER_TYPE_MIN) {
11372 		/* 802.2 LLC header */
11373 		lsh = (struct dot11_llc_snap_header *)(pdata + sizeof(*eh8023));
11374 		eh8023->ether_type = lsh->type;
11375 		pdata = PKTPULL(osh, p, DOT11_LLC_SNAP_HDR_LEN);
11376 		memcpy_s(pdata, sizeof(*eh8023), eh8023, sizeof(*eh8023));
11377 	 } else {
11378 		DHD_ERROR_RLMT(("ethertype 0x%x is not a length !\n", ether_type));
11379 		return BCME_BADARG;
11380 	 }
11381 
11382 	return BCME_OK;
11383 }
11384 #endif /* HOST_SFH_LLC */
11385 
11386 #ifdef DHD_AWDL
11387 
11388 #define AWDL_MIN_EXTENSION_DEFAULT 0x3u
11389 #define AWDL_PRESENCE_MODE_DEFAULT 0x4u
11390 #define AWDL_FLAGS_DEFAULT 0x0000u
11391 #define AWDL_PID 0x0800u
11392 #define AWDL_USERDATA_SIZE 6u
11393 /** Convert Ethernet to 802.3 + AWDL LLC SNAP header
11394  * Note:- This function will overwrite the ethernet header in the pkt 'p'
11395  * with a 802.3 ethernet + AWDL LLC/SNAP header by utilising the headroom
11396  * in the packet. The pkt data pointer should be pointing to the
11397  * start of the packet (at the ethernet header) when the function is called.
11398  * The pkt data pointer will be pointing to the
11399  * start of the new 802.3 header if the function returns successfully
11400  */
11401 int
BCMFASTPATH(dhd_ether_to_awdl_llc_hdr)11402 BCMFASTPATH(dhd_ether_to_awdl_llc_hdr)(struct dhd_pub *dhd, struct ether_header *eh, void *p)
11403 {
11404 	osl_t *osh = dhd->osh;
11405 	struct ether_header *neh;
11406 	struct dot11_llc_snap_header *lsh;
11407 	uint16 plen, ether_type;
11408 	uint8 *awdl_data = NULL;
11409 	uint16 *seq = NULL;
11410 	uint16 *flags = NULL;
11411 	uint16 *type = NULL;
11412 
11413 	if (PKTHEADROOM(osh, p) < (2 * DOT11_LLC_SNAP_HDR_LEN)) {
11414 		DHD_ERROR(("%s: FATAL! not enough pkt headroom !\n", __FUNCTION__));
11415 		ASSERT(0);
11416 		return BCME_BUFTOOSHORT;
11417 	}
11418 
11419 	ether_type = ntoh16(eh->ether_type);
11420 	neh = (struct ether_header *)PKTPUSH(osh, p, 2 * DOT11_LLC_SNAP_HDR_LEN);
11421 
11422 	/* 802.3 MAC header */
11423 	eacopy((char*)eh->ether_dhost, (char*)neh->ether_dhost);
11424 	eacopy((char*)eh->ether_shost, (char*)neh->ether_shost);
11425 	plen = (uint16)PKTLEN(osh, p) - ETHER_HDR_LEN;
11426 	neh->ether_type = hton16(plen);
11427 
11428 	/* 802.2 LLC header */
11429 	lsh = (struct dot11_llc_snap_header *)&neh[1];
11430 	lsh->dsap = 0xaa;
11431 	lsh->ssap = 0xaa;
11432 	lsh->ctl = 0x03;
11433 
11434 	/* 802.2 SNAP header */
11435 	lsh->oui[0] = 0x00;
11436 	lsh->oui[1] = 0x17;
11437 	lsh->oui[2] = 0xf2;
11438 	lsh->type = hton16(AWDL_PID);
11439 
11440 	/* AWDL upper layer data */
11441 	awdl_data = (uint8 *)&lsh[1];
11442 
11443 	awdl_data[0] = dhd->awdl_minext;
11444 	awdl_data[1] = dhd->awdl_presmode;
11445 
11446 	seq = (uint16 *)&awdl_data[2];
11447 	*seq = dhd->awdl_seq++;
11448 
11449 	flags = (uint16 *)&awdl_data[4];
11450 	*flags = hton16(AWDL_FLAGS_DEFAULT);
11451 
11452 	type = (uint16 *)&awdl_data[6];
11453 	*type = hton16(ether_type);
11454 
11455 	return BCME_OK;
11456 }
11457 
11458 /** Convert 802.3 + AWDL LLC SNAP header to ethernet header
11459  * Note:- This function will overwrite the existing
11460  * 802.3 ethernet + AWDL LLC/SNAP header in the packet 'p'
11461  * with a 14 byte ethernet header
11462  * The pkt data pointer should be pointing to the
11463  * start of the packet (at the 802.3 header) when the function is called.
11464  * The pkt data pointer will be pointing to the
11465  * start of the new ethernet header if the function returns successfully
11466  */
11467 int
dhd_awdl_llc_to_eth_hdr(struct dhd_pub * dhd,struct ether_header * eh,void * p)11468 dhd_awdl_llc_to_eth_hdr(struct dhd_pub *dhd, struct ether_header *eh, void *p)
11469 {
11470 	uint16 *ethertype = NULL;
11471 	uint8 *ptr = NULL;
11472 
11473 	if (!eh || !p || !dhd)
11474 		return BCME_BADARG;
11475 
11476 	ptr = PKTDATA(dhd->osh, p);
11477 
11478 	/* copy ether type instead of length from the
11479 	 * end of the awdl llc header to the ethernet header
11480 	 */
11481 	ptr += sizeof(*eh) + DOT11_LLC_SNAP_HDR_LEN + AWDL_USERDATA_SIZE;
11482 	ethertype = (uint16 *)ptr;
11483 	eh->ether_type = *ethertype;
11484 
11485 	/* overwrite awdl llc header with ethernet header */
11486 	PKTPULL(dhd->osh, p, 2 * DOT11_LLC_SNAP_HDR_LEN);
11487 	ptr = PKTDATA(dhd->osh, p);
11488 	memcpy_s(ptr, sizeof(*eh), eh, sizeof(*eh));
11489 	return BCME_OK;
11490 }
11491 #endif /* DHD_AWDL */
11492 
11493 int
dhd_iovar(dhd_pub_t * pub,int ifidx,char * name,char * param_buf,uint param_len,char * res_buf,uint res_len,bool set)11494 dhd_iovar(dhd_pub_t *pub, int ifidx, char *name, char *param_buf, uint param_len, char *res_buf,
11495 		uint res_len, bool set)
11496 {
11497 	char *buf = NULL;
11498 	uint input_len;
11499 	wl_ioctl_t ioc;
11500 	int ret;
11501 
11502 	if (res_len > WLC_IOCTL_MAXLEN || param_len > WLC_IOCTL_MAXLEN)
11503 		return BCME_BADARG;
11504 
11505 	input_len = strlen(name) + 1 + param_len;
11506 
11507 	/* WAR to fix GET iovar returning buf too short error
11508 	 * If param len is 0 for get iovar, increment input_len by sizeof(int)
11509 	 * to avoid the length check error in fw
11510 	 */
11511 	if (!set && !param_len) {
11512 		input_len += sizeof(int);
11513 	}
11514 	if (input_len > WLC_IOCTL_MAXLEN)
11515 		return BCME_BADARG;
11516 
11517 	buf = NULL;
11518 	if (set) {
11519 		if (res_buf || res_len != 0) {
11520 			DHD_ERROR(("%s: SET wrong arguemnet\n", __FUNCTION__));
11521 			ret = BCME_BADARG;
11522 			goto exit;
11523 		}
11524 		buf = MALLOCZ(pub->osh, input_len);
11525 		if (!buf) {
11526 			DHD_ERROR(("%s: mem alloc failed\n", __FUNCTION__));
11527 			ret = BCME_NOMEM;
11528 			goto exit;
11529 		}
11530 		ret = bcm_mkiovar(name, param_buf, param_len, buf, input_len);
11531 		if (!ret) {
11532 			ret = BCME_NOMEM;
11533 			goto exit;
11534 		}
11535 
11536 		ioc.cmd = WLC_SET_VAR;
11537 		ioc.buf = buf;
11538 		ioc.len = input_len;
11539 		ioc.set = set;
11540 
11541 		ret = dhd_wl_ioctl(pub, ifidx, &ioc, ioc.buf, ioc.len);
11542 	} else {
11543 		if (!res_buf || !res_len) {
11544 			DHD_ERROR(("%s: GET failed. resp_buf NULL or length 0.\n", __FUNCTION__));
11545 			ret = BCME_BADARG;
11546 			goto exit;
11547 		}
11548 
11549 		if (res_len < input_len) {
11550 			DHD_INFO(("%s: res_len(%d) < input_len(%d)\n", __FUNCTION__,
11551 					res_len, input_len));
11552 			buf = MALLOCZ(pub->osh, input_len);
11553 			if (!buf) {
11554 				DHD_ERROR(("%s: mem alloc failed\n", __FUNCTION__));
11555 				ret = BCME_NOMEM;
11556 				goto exit;
11557 			}
11558 			ret = bcm_mkiovar(name, param_buf, param_len, buf, input_len);
11559 			if (!ret) {
11560 				ret = BCME_NOMEM;
11561 				goto exit;
11562 			}
11563 
11564 			ioc.cmd = WLC_GET_VAR;
11565 			ioc.buf = buf;
11566 			ioc.len = input_len;
11567 			ioc.set = set;
11568 
11569 			ret = dhd_wl_ioctl(pub, ifidx, &ioc, ioc.buf, ioc.len);
11570 
11571 			if (ret == BCME_OK) {
11572 				memcpy(res_buf, buf, res_len);
11573 			}
11574 		} else {
11575 			memset(res_buf, 0, res_len);
11576 			ret = bcm_mkiovar(name, param_buf, param_len, res_buf, res_len);
11577 			if (!ret) {
11578 				ret = BCME_NOMEM;
11579 				goto exit;
11580 			}
11581 
11582 			ioc.cmd = WLC_GET_VAR;
11583 			ioc.buf = res_buf;
11584 			ioc.len = res_len;
11585 			ioc.set = set;
11586 
11587 			ret = dhd_wl_ioctl(pub, ifidx, &ioc, ioc.buf, ioc.len);
11588 		}
11589 	}
11590 exit:
11591 	if (buf) {
11592 		MFREE(pub->osh, buf, input_len);
11593 	}
11594 	return ret;
11595 }
11596