1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
|
/* SPDX-License-Identifier: BSD-3-Clause-Clear */
/*
* Copyright (c) 2018-2019 The Linux Foundation. All rights reserved.
*/
#ifndef DEBUG_HTT_STATS_H
#define DEBUG_HTT_STATS_H
#define HTT_STATS_COOKIE_LSB GENMASK_ULL(31, 0)
#define HTT_STATS_COOKIE_MSB GENMASK_ULL(63, 32)
#define HTT_STATS_MAGIC_VALUE 0xF0F0F0F0
enum htt_tlv_tag_t {
HTT_STATS_TX_PDEV_CMN_TAG = 0,
HTT_STATS_TX_PDEV_UNDERRUN_TAG = 1,
HTT_STATS_TX_PDEV_SIFS_TAG = 2,
HTT_STATS_TX_PDEV_FLUSH_TAG = 3,
HTT_STATS_TX_PDEV_PHY_ERR_TAG = 4,
HTT_STATS_STRING_TAG = 5,
HTT_STATS_TX_HWQ_CMN_TAG = 6,
HTT_STATS_TX_HWQ_DIFS_LATENCY_TAG = 7,
HTT_STATS_TX_HWQ_CMD_RESULT_TAG = 8,
HTT_STATS_TX_HWQ_CMD_STALL_TAG = 9,
HTT_STATS_TX_HWQ_FES_STATUS_TAG = 10,
HTT_STATS_TX_TQM_GEN_MPDU_TAG = 11,
HTT_STATS_TX_TQM_LIST_MPDU_TAG = 12,
HTT_STATS_TX_TQM_LIST_MPDU_CNT_TAG = 13,
HTT_STATS_TX_TQM_CMN_TAG = 14,
HTT_STATS_TX_TQM_PDEV_TAG = 15,
HTT_STATS_TX_TQM_CMDQ_STATUS_TAG = 16,
HTT_STATS_TX_DE_EAPOL_PACKETS_TAG = 17,
HTT_STATS_TX_DE_CLASSIFY_FAILED_TAG = 18,
HTT_STATS_TX_DE_CLASSIFY_STATS_TAG = 19,
HTT_STATS_TX_DE_CLASSIFY_STATUS_TAG = 20,
HTT_STATS_TX_DE_ENQUEUE_PACKETS_TAG = 21,
HTT_STATS_TX_DE_ENQUEUE_DISCARD_TAG = 22,
HTT_STATS_TX_DE_CMN_TAG = 23,
HTT_STATS_RING_IF_TAG = 24,
HTT_STATS_TX_PDEV_MU_MIMO_STATS_TAG = 25,
HTT_STATS_SFM_CMN_TAG = 26,
HTT_STATS_SRING_STATS_TAG = 27,
HTT_STATS_RX_PDEV_FW_STATS_TAG = 28,
HTT_STATS_RX_PDEV_FW_RING_MPDU_ERR_TAG = 29,
HTT_STATS_RX_PDEV_FW_MPDU_DROP_TAG = 30,
HTT_STATS_RX_SOC_FW_STATS_TAG = 31,
HTT_STATS_RX_SOC_FW_REFILL_RING_EMPTY_TAG = 32,
HTT_STATS_RX_SOC_FW_REFILL_RING_NUM_REFILL_TAG = 33,
HTT_STATS_TX_PDEV_RATE_STATS_TAG = 34,
HTT_STATS_RX_PDEV_RATE_STATS_TAG = 35,
HTT_STATS_TX_PDEV_SCHEDULER_TXQ_STATS_TAG = 36,
HTT_STATS_TX_SCHED_CMN_TAG = 37,
HTT_STATS_TX_PDEV_MUMIMO_MPDU_STATS_TAG = 38,
HTT_STATS_SCHED_TXQ_CMD_POSTED_TAG = 39,
HTT_STATS_RING_IF_CMN_TAG = 40,
HTT_STATS_SFM_CLIENT_USER_TAG = 41,
HTT_STATS_SFM_CLIENT_TAG = 42,
HTT_STATS_TX_TQM_ERROR_STATS_TAG = 43,
HTT_STATS_SCHED_TXQ_CMD_REAPED_TAG = 44,
HTT_STATS_SRING_CMN_TAG = 45,
HTT_STATS_TX_SELFGEN_AC_ERR_STATS_TAG = 46,
HTT_STATS_TX_SELFGEN_CMN_STATS_TAG = 47,
HTT_STATS_TX_SELFGEN_AC_STATS_TAG = 48,
HTT_STATS_TX_SELFGEN_AX_STATS_TAG = 49,
HTT_STATS_TX_SELFGEN_AX_ERR_STATS_TAG = 50,
HTT_STATS_TX_HWQ_MUMIMO_SCH_STATS_TAG = 51,
HTT_STATS_TX_HWQ_MUMIMO_MPDU_STATS_TAG = 52,
HTT_STATS_TX_HWQ_MUMIMO_CMN_STATS_TAG = 53,
HTT_STATS_HW_INTR_MISC_TAG = 54,
HTT_STATS_HW_WD_TIMEOUT_TAG = 55,
HTT_STATS_HW_PDEV_ERRS_TAG = 56,
HTT_STATS_COUNTER_NAME_TAG = 57,
HTT_STATS_TX_TID_DETAILS_TAG = 58,
HTT_STATS_RX_TID_DETAILS_TAG = 59,
HTT_STATS_PEER_STATS_CMN_TAG = 60,
HTT_STATS_PEER_DETAILS_TAG = 61,
HTT_STATS_PEER_TX_RATE_STATS_TAG = 62,
HTT_STATS_PEER_RX_RATE_STATS_TAG = 63,
HTT_STATS_PEER_MSDU_FLOWQ_TAG = 64,
HTT_STATS_TX_DE_COMPL_STATS_TAG = 65,
HTT_STATS_WHAL_TX_TAG = 66,
HTT_STATS_TX_PDEV_SIFS_HIST_TAG = 67,
HTT_STATS_RX_PDEV_FW_STATS_PHY_ERR_TAG = 68,
HTT_STATS_TX_TID_DETAILS_V1_TAG = 69,
HTT_STATS_PDEV_CCA_1SEC_HIST_TAG = 70,
HTT_STATS_PDEV_CCA_100MSEC_HIST_TAG = 71,
HTT_STATS_PDEV_CCA_STAT_CUMULATIVE_TAG = 72,
HTT_STATS_PDEV_CCA_COUNTERS_TAG = 73,
HTT_STATS_TX_PDEV_MPDU_STATS_TAG = 74,
HTT_STATS_PDEV_TWT_SESSIONS_TAG = 75,
HTT_STATS_PDEV_TWT_SESSION_TAG = 76,
HTT_STATS_RX_REFILL_RXDMA_ERR_TAG = 77,
HTT_STATS_RX_REFILL_REO_ERR_TAG = 78,
HTT_STATS_RX_REO_RESOURCE_STATS_TAG = 79,
HTT_STATS_TX_SOUNDING_STATS_TAG = 80,
HTT_STATS_TX_PDEV_TX_PPDU_STATS_TAG = 81,
HTT_STATS_TX_PDEV_TRIED_MPDU_CNT_HIST_TAG = 82,
HTT_STATS_TX_HWQ_TRIED_MPDU_CNT_HIST_TAG = 83,
HTT_STATS_TX_HWQ_TXOP_USED_CNT_HIST_TAG = 84,
HTT_STATS_TX_DE_FW2WBM_RING_FULL_HIST_TAG = 85,
HTT_STATS_SCHED_TXQ_SCHED_ORDER_SU_TAG = 86,
HTT_STATS_SCHED_TXQ_SCHED_INELIGIBILITY_TAG = 87,
HTT_STATS_PDEV_OBSS_PD_TAG = 88,
HTT_STATS_HW_WAR_TAG = 89,
HTT_STATS_RING_BACKPRESSURE_STATS_TAG = 90,
HTT_STATS_MAX_TAG,
};
#define HTT_STATS_MAX_STRING_SZ32 4
#define HTT_STATS_MACID_INVALID 0xff
#define HTT_TX_HWQ_MAX_DIFS_LATENCY_BINS 10
#define HTT_TX_HWQ_MAX_CMD_RESULT_STATS 13
#define HTT_TX_HWQ_MAX_CMD_STALL_STATS 5
#define HTT_TX_HWQ_MAX_FES_RESULT_STATS 10
enum htt_tx_pdev_underrun_enum {
HTT_STATS_TX_PDEV_NO_DATA_UNDERRUN = 0,
HTT_STATS_TX_PDEV_DATA_UNDERRUN_BETWEEN_MPDU = 1,
HTT_STATS_TX_PDEV_DATA_UNDERRUN_WITHIN_MPDU = 2,
HTT_TX_PDEV_MAX_URRN_STATS = 3,
};
#define HTT_TX_PDEV_MAX_FLUSH_REASON_STATS 71
#define HTT_TX_PDEV_MAX_SIFS_BURST_STATS 9
#define HTT_TX_PDEV_MAX_SIFS_BURST_HIST_STATS 10
#define HTT_TX_PDEV_MAX_PHY_ERR_STATS 18
#define HTT_TX_PDEV_SCHED_TX_MODE_MAX 4
#define HTT_TX_PDEV_NUM_SCHED_ORDER_LOG 20
#define HTT_RX_STATS_REFILL_MAX_RING 4
#define HTT_RX_STATS_RXDMA_MAX_ERR 16
#define HTT_RX_STATS_FW_DROP_REASON_MAX 16
/* Bytes stored in little endian order */
/* Length should be multiple of DWORD */
struct htt_stats_string_tlv {
u32 data[0]; /* Can be variable length */
} __packed;
/* == TX PDEV STATS == */
struct htt_tx_pdev_stats_cmn_tlv {
u32 mac_id__word;
u32 hw_queued;
u32 hw_reaped;
u32 underrun;
u32 hw_paused;
u32 hw_flush;
u32 hw_filt;
u32 tx_abort;
u32 mpdu_requeued;
u32 tx_xretry;
u32 data_rc;
u32 mpdu_dropped_xretry;
u32 illgl_rate_phy_err;
u32 cont_xretry;
u32 tx_timeout;
u32 pdev_resets;
u32 phy_underrun;
u32 txop_ovf;
u32 seq_posted;
u32 seq_failed_queueing;
u32 seq_completed;
u32 seq_restarted;
u32 mu_seq_posted;
u32 seq_switch_hw_paused;
u32 next_seq_posted_dsr;
u32 seq_posted_isr;
u32 seq_ctrl_cached;
u32 mpdu_count_tqm;
u32 msdu_count_tqm;
u32 mpdu_removed_tqm;
u32 msdu_removed_tqm;
u32 mpdus_sw_flush;
u32 mpdus_hw_filter;
u32 mpdus_truncated;
u32 mpdus_ack_failed;
u32 mpdus_expired;
u32 mpdus_seq_hw_retry;
u32 ack_tlv_proc;
u32 coex_abort_mpdu_cnt_valid;
u32 coex_abort_mpdu_cnt;
u32 num_total_ppdus_tried_ota;
u32 num_data_ppdus_tried_ota;
u32 local_ctrl_mgmt_enqued;
u32 local_ctrl_mgmt_freed;
u32 local_data_enqued;
u32 local_data_freed;
u32 mpdu_tried;
u32 isr_wait_seq_posted;
u32 tx_active_dur_us_low;
u32 tx_active_dur_us_high;
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_pdev_stats_urrn_tlv_v {
u32 urrn_stats[0]; /* HTT_TX_PDEV_MAX_URRN_STATS */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_pdev_stats_flush_tlv_v {
u32 flush_errs[0]; /* HTT_TX_PDEV_MAX_FLUSH_REASON_STATS */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_pdev_stats_sifs_tlv_v {
u32 sifs_status[0]; /* HTT_TX_PDEV_MAX_SIFS_BURST_STATS */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_pdev_stats_phy_err_tlv_v {
u32 phy_errs[0]; /* HTT_TX_PDEV_MAX_PHY_ERR_STATS */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_pdev_stats_sifs_hist_tlv_v {
u32 sifs_hist_status[0]; /* HTT_TX_PDEV_SIFS_BURST_HIST_STATS */
};
struct htt_tx_pdev_stats_tx_ppdu_stats_tlv_v {
u32 num_data_ppdus_legacy_su;
u32 num_data_ppdus_ac_su;
u32 num_data_ppdus_ax_su;
u32 num_data_ppdus_ac_su_txbf;
u32 num_data_ppdus_ax_su_txbf;
};
/* NOTE: Variable length TLV, use length spec to infer array size .
*
* Tried_mpdu_cnt_hist is the histogram of MPDUs tries per HWQ.
* The tries here is the count of the MPDUS within a PPDU that the
* HW had attempted to transmit on air, for the HWSCH Schedule
* command submitted by FW.It is not the retry attempts.
* The histogram bins are 0-29, 30-59, 60-89 and so on. The are
* 10 bins in this histogram. They are defined in FW using the
* following macros
* #define WAL_MAX_TRIED_MPDU_CNT_HISTOGRAM 9
* #define WAL_TRIED_MPDU_CNT_HISTOGRAM_INTERVAL 30
*/
struct htt_tx_pdev_stats_tried_mpdu_cnt_hist_tlv_v {
u32 hist_bin_size;
u32 tried_mpdu_cnt_hist[]; /* HTT_TX_PDEV_TRIED_MPDU_CNT_HIST */
};
/* == SOC ERROR STATS == */
/* =============== PDEV ERROR STATS ============== */
#define HTT_STATS_MAX_HW_INTR_NAME_LEN 8
struct htt_hw_stats_intr_misc_tlv {
/* Stored as little endian */
u8 hw_intr_name[HTT_STATS_MAX_HW_INTR_NAME_LEN];
u32 mask;
u32 count;
};
#define HTT_STATS_MAX_HW_MODULE_NAME_LEN 8
struct htt_hw_stats_wd_timeout_tlv {
/* Stored as little endian */
u8 hw_module_name[HTT_STATS_MAX_HW_MODULE_NAME_LEN];
u32 count;
};
struct htt_hw_stats_pdev_errs_tlv {
u32 mac_id__word; /* BIT [ 7 : 0] : mac_id */
u32 tx_abort;
u32 tx_abort_fail_count;
u32 rx_abort;
u32 rx_abort_fail_count;
u32 warm_reset;
u32 cold_reset;
u32 tx_flush;
u32 tx_glb_reset;
u32 tx_txq_reset;
u32 rx_timeout_reset;
};
struct htt_hw_stats_whal_tx_tlv {
u32 mac_id__word;
u32 last_unpause_ppdu_id;
u32 hwsch_unpause_wait_tqm_write;
u32 hwsch_dummy_tlv_skipped;
u32 hwsch_misaligned_offset_received;
u32 hwsch_reset_count;
u32 hwsch_dev_reset_war;
u32 hwsch_delayed_pause;
u32 hwsch_long_delayed_pause;
u32 sch_rx_ppdu_no_response;
u32 sch_selfgen_response;
u32 sch_rx_sifs_resp_trigger;
};
/* ============ PEER STATS ============ */
struct htt_msdu_flow_stats_tlv {
u32 last_update_timestamp;
u32 last_add_timestamp;
u32 last_remove_timestamp;
u32 total_processed_msdu_count;
u32 cur_msdu_count_in_flowq;
u32 sw_peer_id;
u32 tx_flow_no__tid_num__drop_rule;
u32 last_cycle_enqueue_count;
u32 last_cycle_dequeue_count;
u32 last_cycle_drop_count;
u32 current_drop_th;
};
#define MAX_HTT_TID_NAME 8
/* Tidq stats */
struct htt_tx_tid_stats_tlv {
/* Stored as little endian */
u8 tid_name[MAX_HTT_TID_NAME];
u32 sw_peer_id__tid_num;
u32 num_sched_pending__num_ppdu_in_hwq;
u32 tid_flags;
u32 hw_queued;
u32 hw_reaped;
u32 mpdus_hw_filter;
u32 qdepth_bytes;
u32 qdepth_num_msdu;
u32 qdepth_num_mpdu;
u32 last_scheduled_tsmp;
u32 pause_module_id;
u32 block_module_id;
u32 tid_tx_airtime;
};
/* Tidq stats */
struct htt_tx_tid_stats_v1_tlv {
/* Stored as little endian */
u8 tid_name[MAX_HTT_TID_NAME];
u32 sw_peer_id__tid_num;
u32 num_sched_pending__num_ppdu_in_hwq;
u32 tid_flags;
u32 max_qdepth_bytes;
u32 max_qdepth_n_msdus;
u32 rsvd;
u32 qdepth_bytes;
u32 qdepth_num_msdu;
u32 qdepth_num_mpdu;
u32 last_scheduled_tsmp;
u32 pause_module_id;
u32 block_module_id;
u32 tid_tx_airtime;
u32 allow_n_flags;
u32 sendn_frms_allowed;
};
struct htt_rx_tid_stats_tlv {
u32 sw_peer_id__tid_num;
u8 tid_name[MAX_HTT_TID_NAME];
u32 dup_in_reorder;
u32 dup_past_outside_window;
u32 dup_past_within_window;
u32 rxdesc_err_decrypt;
u32 tid_rx_airtime;
};
#define HTT_MAX_COUNTER_NAME 8
struct htt_counter_tlv {
u8 counter_name[HTT_MAX_COUNTER_NAME];
u32 count;
};
struct htt_peer_stats_cmn_tlv {
u32 ppdu_cnt;
u32 mpdu_cnt;
u32 msdu_cnt;
u32 pause_bitmap;
u32 block_bitmap;
u32 current_timestamp;
u32 peer_tx_airtime;
u32 peer_rx_airtime;
s32 rssi;
u32 peer_enqueued_count_low;
u32 peer_enqueued_count_high;
u32 peer_dequeued_count_low;
u32 peer_dequeued_count_high;
u32 peer_dropped_count_low;
u32 peer_dropped_count_high;
u32 ppdu_transmitted_bytes_low;
u32 ppdu_transmitted_bytes_high;
u32 peer_ttl_removed_count;
u32 inactive_time;
};
struct htt_peer_details_tlv {
u32 peer_type;
u32 sw_peer_id;
u32 vdev_pdev_ast_idx;
struct htt_mac_addr mac_addr;
u32 peer_flags;
u32 qpeer_flags;
};
enum htt_stats_param_type {
HTT_STATS_PREAM_OFDM,
HTT_STATS_PREAM_CCK,
HTT_STATS_PREAM_HT,
HTT_STATS_PREAM_VHT,
HTT_STATS_PREAM_HE,
HTT_STATS_PREAM_RSVD,
HTT_STATS_PREAM_RSVD1,
HTT_STATS_PREAM_COUNT,
};
#define HTT_TX_PEER_STATS_NUM_MCS_COUNTERS 12
#define HTT_TX_PEER_STATS_NUM_GI_COUNTERS 4
#define HTT_TX_PEER_STATS_NUM_DCM_COUNTERS 5
#define HTT_TX_PEER_STATS_NUM_BW_COUNTERS 4
#define HTT_TX_PEER_STATS_NUM_SPATIAL_STREAMS 8
#define HTT_TX_PEER_STATS_NUM_PREAMBLE_TYPES HTT_STATS_PREAM_COUNT
struct htt_tx_peer_rate_stats_tlv {
u32 tx_ldpc;
u32 rts_cnt;
u32 ack_rssi;
u32 tx_mcs[HTT_TX_PEER_STATS_NUM_MCS_COUNTERS];
u32 tx_su_mcs[HTT_TX_PEER_STATS_NUM_MCS_COUNTERS];
u32 tx_mu_mcs[HTT_TX_PEER_STATS_NUM_MCS_COUNTERS];
/* element 0,1, ...7 -> NSS 1,2, ...8 */
u32 tx_nss[HTT_TX_PEER_STATS_NUM_SPATIAL_STREAMS];
/* element 0: 20 MHz, 1: 40 MHz, 2: 80 MHz, 3: 160 and 80+80 MHz */
u32 tx_bw[HTT_TX_PEER_STATS_NUM_BW_COUNTERS];
u32 tx_stbc[HTT_TX_PEER_STATS_NUM_MCS_COUNTERS];
u32 tx_pream[HTT_TX_PEER_STATS_NUM_PREAMBLE_TYPES];
/* Counters to track number of tx packets in each GI
* (400us, 800us, 1600us & 3200us) in each mcs (0-11)
*/
u32 tx_gi[HTT_TX_PEER_STATS_NUM_GI_COUNTERS][HTT_TX_PEER_STATS_NUM_MCS_COUNTERS];
/* Counters to track packets in dcm mcs (MCS 0, 1, 3, 4) */
u32 tx_dcm[HTT_TX_PEER_STATS_NUM_DCM_COUNTERS];
};
#define HTT_RX_PEER_STATS_NUM_MCS_COUNTERS 12
#define HTT_RX_PEER_STATS_NUM_GI_COUNTERS 4
#define HTT_RX_PEER_STATS_NUM_DCM_COUNTERS 5
#define HTT_RX_PEER_STATS_NUM_BW_COUNTERS 4
#define HTT_RX_PEER_STATS_NUM_SPATIAL_STREAMS 8
#define HTT_RX_PEER_STATS_NUM_PREAMBLE_TYPES HTT_STATS_PREAM_COUNT
struct htt_rx_peer_rate_stats_tlv {
u32 nsts;
/* Number of rx ldpc packets */
u32 rx_ldpc;
/* Number of rx rts packets */
u32 rts_cnt;
u32 rssi_mgmt; /* units = dB above noise floor */
u32 rssi_data; /* units = dB above noise floor */
u32 rssi_comb; /* units = dB above noise floor */
u32 rx_mcs[HTT_RX_PEER_STATS_NUM_MCS_COUNTERS];
/* element 0,1, ...7 -> NSS 1,2, ...8 */
u32 rx_nss[HTT_RX_PEER_STATS_NUM_SPATIAL_STREAMS];
u32 rx_dcm[HTT_RX_PEER_STATS_NUM_DCM_COUNTERS];
u32 rx_stbc[HTT_RX_PEER_STATS_NUM_MCS_COUNTERS];
/* element 0: 20 MHz, 1: 40 MHz, 2: 80 MHz, 3: 160 and 80+80 MHz */
u32 rx_bw[HTT_RX_PEER_STATS_NUM_BW_COUNTERS];
u32 rx_pream[HTT_RX_PEER_STATS_NUM_PREAMBLE_TYPES];
/* units = dB above noise floor */
u8 rssi_chain[HTT_RX_PEER_STATS_NUM_SPATIAL_STREAMS]
[HTT_RX_PEER_STATS_NUM_BW_COUNTERS];
/* Counters to track number of rx packets in each GI in each mcs (0-11) */
u32 rx_gi[HTT_RX_PEER_STATS_NUM_GI_COUNTERS]
[HTT_RX_PEER_STATS_NUM_MCS_COUNTERS];
};
enum htt_peer_stats_req_mode {
HTT_PEER_STATS_REQ_MODE_NO_QUERY,
HTT_PEER_STATS_REQ_MODE_QUERY_TQM,
HTT_PEER_STATS_REQ_MODE_FLUSH_TQM,
};
enum htt_peer_stats_tlv_enum {
HTT_PEER_STATS_CMN_TLV = 0,
HTT_PEER_DETAILS_TLV = 1,
HTT_TX_PEER_RATE_STATS_TLV = 2,
HTT_RX_PEER_RATE_STATS_TLV = 3,
HTT_TX_TID_STATS_TLV = 4,
HTT_RX_TID_STATS_TLV = 5,
HTT_MSDU_FLOW_STATS_TLV = 6,
HTT_PEER_STATS_MAX_TLV = 31,
};
/* =========== MUMIMO HWQ stats =========== */
/* MU MIMO stats per hwQ */
struct htt_tx_hwq_mu_mimo_sch_stats_tlv {
u32 mu_mimo_sch_posted;
u32 mu_mimo_sch_failed;
u32 mu_mimo_ppdu_posted;
};
struct htt_tx_hwq_mu_mimo_mpdu_stats_tlv {
u32 mu_mimo_mpdus_queued_usr;
u32 mu_mimo_mpdus_tried_usr;
u32 mu_mimo_mpdus_failed_usr;
u32 mu_mimo_mpdus_requeued_usr;
u32 mu_mimo_err_no_ba_usr;
u32 mu_mimo_mpdu_underrun_usr;
u32 mu_mimo_ampdu_underrun_usr;
};
struct htt_tx_hwq_mu_mimo_cmn_stats_tlv {
u32 mac_id__hwq_id__word;
};
/* == TX HWQ STATS == */
struct htt_tx_hwq_stats_cmn_tlv {
u32 mac_id__hwq_id__word;
/* PPDU level stats */
u32 xretry;
u32 underrun_cnt;
u32 flush_cnt;
u32 filt_cnt;
u32 null_mpdu_bmap;
u32 user_ack_failure;
u32 ack_tlv_proc;
u32 sched_id_proc;
u32 null_mpdu_tx_count;
u32 mpdu_bmap_not_recvd;
/* Selfgen stats per hwQ */
u32 num_bar;
u32 rts;
u32 cts2self;
u32 qos_null;
/* MPDU level stats */
u32 mpdu_tried_cnt;
u32 mpdu_queued_cnt;
u32 mpdu_ack_fail_cnt;
u32 mpdu_filt_cnt;
u32 false_mpdu_ack_count;
u32 txq_timeout;
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_hwq_difs_latency_stats_tlv_v {
u32 hist_intvl;
/* histogram of ppdu post to hwsch - > cmd status received */
u32 difs_latency_hist[]; /* HTT_TX_HWQ_MAX_DIFS_LATENCY_BINS */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_hwq_cmd_result_stats_tlv_v {
/* Histogram of sched cmd result */
u32 cmd_result[0]; /* HTT_TX_HWQ_MAX_CMD_RESULT_STATS */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_hwq_cmd_stall_stats_tlv_v {
/* Histogram of various pause conitions */
u32 cmd_stall_status[0]; /* HTT_TX_HWQ_MAX_CMD_STALL_STATS */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_hwq_fes_result_stats_tlv_v {
/* Histogram of number of user fes result */
u32 fes_result[0]; /* HTT_TX_HWQ_MAX_FES_RESULT_STATS */
};
/* NOTE: Variable length TLV, use length spec to infer array size
*
* The hwq_tried_mpdu_cnt_hist is a histogram of MPDUs tries per HWQ.
* The tries here is the count of the MPDUS within a PPDU that the HW
* had attempted to transmit on air, for the HWSCH Schedule command
* submitted by FW in this HWQ .It is not the retry attempts. The
* histogram bins are 0-29, 30-59, 60-89 and so on. The are 10 bins
* in this histogram.
* they are defined in FW using the following macros
* #define WAL_MAX_TRIED_MPDU_CNT_HISTOGRAM 9
* #define WAL_TRIED_MPDU_CNT_HISTOGRAM_INTERVAL 30
*/
struct htt_tx_hwq_tried_mpdu_cnt_hist_tlv_v {
u32 hist_bin_size;
/* Histogram of number of mpdus on tried mpdu */
u32 tried_mpdu_cnt_hist[]; /* HTT_TX_HWQ_TRIED_MPDU_CNT_HIST */
};
/* NOTE: Variable length TLV, use length spec to infer array size
*
* The txop_used_cnt_hist is the histogram of txop per burst. After
* completing the burst, we identify the txop used in the burst and
* incr the corresponding bin.
* Each bin represents 1ms & we have 10 bins in this histogram.
* they are deined in FW using the following macros
* #define WAL_MAX_TXOP_USED_CNT_HISTOGRAM 10
* #define WAL_TXOP_USED_HISTOGRAM_INTERVAL 1000 ( 1 ms )
*/
struct htt_tx_hwq_txop_used_cnt_hist_tlv_v {
/* Histogram of txop used cnt */
u32 txop_used_cnt_hist[0]; /* HTT_TX_HWQ_TXOP_USED_CNT_HIST */
};
/* == TX SELFGEN STATS == */
struct htt_tx_selfgen_cmn_stats_tlv {
u32 mac_id__word;
u32 su_bar;
u32 rts;
u32 cts2self;
u32 qos_null;
u32 delayed_bar_1; /* MU user 1 */
u32 delayed_bar_2; /* MU user 2 */
u32 delayed_bar_3; /* MU user 3 */
u32 delayed_bar_4; /* MU user 4 */
u32 delayed_bar_5; /* MU user 5 */
u32 delayed_bar_6; /* MU user 6 */
u32 delayed_bar_7; /* MU user 7 */
};
struct htt_tx_selfgen_ac_stats_tlv {
/* 11AC */
u32 ac_su_ndpa;
u32 ac_su_ndp;
u32 ac_mu_mimo_ndpa;
u32 ac_mu_mimo_ndp;
u32 ac_mu_mimo_brpoll_1; /* MU user 1 */
u32 ac_mu_mimo_brpoll_2; /* MU user 2 */
u32 ac_mu_mimo_brpoll_3; /* MU user 3 */
};
struct htt_tx_selfgen_ax_stats_tlv {
/* 11AX */
u32 ax_su_ndpa;
u32 ax_su_ndp;
u32 ax_mu_mimo_ndpa;
u32 ax_mu_mimo_ndp;
u32 ax_mu_mimo_brpoll_1; /* MU user 1 */
u32 ax_mu_mimo_brpoll_2; /* MU user 2 */
u32 ax_mu_mimo_brpoll_3; /* MU user 3 */
u32 ax_mu_mimo_brpoll_4; /* MU user 4 */
u32 ax_mu_mimo_brpoll_5; /* MU user 5 */
u32 ax_mu_mimo_brpoll_6; /* MU user 6 */
u32 ax_mu_mimo_brpoll_7; /* MU user 7 */
u32 ax_basic_trigger;
u32 ax_bsr_trigger;
u32 ax_mu_bar_trigger;
u32 ax_mu_rts_trigger;
};
struct htt_tx_selfgen_ac_err_stats_tlv {
/* 11AC error stats */
u32 ac_su_ndp_err;
u32 ac_su_ndpa_err;
u32 ac_mu_mimo_ndpa_err;
u32 ac_mu_mimo_ndp_err;
u32 ac_mu_mimo_brp1_err;
u32 ac_mu_mimo_brp2_err;
u32 ac_mu_mimo_brp3_err;
};
struct htt_tx_selfgen_ax_err_stats_tlv {
/* 11AX error stats */
u32 ax_su_ndp_err;
u32 ax_su_ndpa_err;
u32 ax_mu_mimo_ndpa_err;
u32 ax_mu_mimo_ndp_err;
u32 ax_mu_mimo_brp1_err;
u32 ax_mu_mimo_brp2_err;
u32 ax_mu_mimo_brp3_err;
u32 ax_mu_mimo_brp4_err;
u32 ax_mu_mimo_brp5_err;
u32 ax_mu_mimo_brp6_err;
u32 ax_mu_mimo_brp7_err;
u32 ax_basic_trigger_err;
u32 ax_bsr_trigger_err;
u32 ax_mu_bar_trigger_err;
u32 ax_mu_rts_trigger_err;
};
/* == TX MU STATS == */
#define HTT_TX_PDEV_STATS_NUM_AC_MUMIMO_USER_STATS 4
#define HTT_TX_PDEV_STATS_NUM_AX_MUMIMO_USER_STATS 8
#define HTT_TX_PDEV_STATS_NUM_OFDMA_USER_STATS 74
struct htt_tx_pdev_mu_mimo_sch_stats_tlv {
/* mu-mimo sw sched cmd stats */
u32 mu_mimo_sch_posted;
u32 mu_mimo_sch_failed;
/* MU PPDU stats per hwQ */
u32 mu_mimo_ppdu_posted;
/*
* Counts the number of users in each transmission of
* the given TX mode.
*
* Index is the number of users - 1.
*/
u32 ac_mu_mimo_sch_nusers[HTT_TX_PDEV_STATS_NUM_AC_MUMIMO_USER_STATS];
u32 ax_mu_mimo_sch_nusers[HTT_TX_PDEV_STATS_NUM_AX_MUMIMO_USER_STATS];
u32 ax_ofdma_sch_nusers[HTT_TX_PDEV_STATS_NUM_OFDMA_USER_STATS];
};
struct htt_tx_pdev_mu_mimo_mpdu_stats_tlv {
u32 mu_mimo_mpdus_queued_usr;
u32 mu_mimo_mpdus_tried_usr;
u32 mu_mimo_mpdus_failed_usr;
u32 mu_mimo_mpdus_requeued_usr;
u32 mu_mimo_err_no_ba_usr;
u32 mu_mimo_mpdu_underrun_usr;
u32 mu_mimo_ampdu_underrun_usr;
u32 ax_mu_mimo_mpdus_queued_usr;
u32 ax_mu_mimo_mpdus_tried_usr;
u32 ax_mu_mimo_mpdus_failed_usr;
u32 ax_mu_mimo_mpdus_requeued_usr;
u32 ax_mu_mimo_err_no_ba_usr;
u32 ax_mu_mimo_mpdu_underrun_usr;
u32 ax_mu_mimo_ampdu_underrun_usr;
u32 ax_ofdma_mpdus_queued_usr;
u32 ax_ofdma_mpdus_tried_usr;
u32 ax_ofdma_mpdus_failed_usr;
u32 ax_ofdma_mpdus_requeued_usr;
u32 ax_ofdma_err_no_ba_usr;
u32 ax_ofdma_mpdu_underrun_usr;
u32 ax_ofdma_ampdu_underrun_usr;
};
#define HTT_STATS_TX_SCHED_MODE_MU_MIMO_AC 1
#define HTT_STATS_TX_SCHED_MODE_MU_MIMO_AX 2
#define HTT_STATS_TX_SCHED_MODE_MU_OFDMA_AX 3
struct htt_tx_pdev_mpdu_stats_tlv {
/* mpdu level stats */
u32 mpdus_queued_usr;
u32 mpdus_tried_usr;
u32 mpdus_failed_usr;
u32 mpdus_requeued_usr;
u32 err_no_ba_usr;
u32 mpdu_underrun_usr;
u32 ampdu_underrun_usr;
u32 user_index;
u32 tx_sched_mode; /* HTT_STATS_TX_SCHED_MODE_xxx */
};
/* == TX SCHED STATS == */
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_sched_txq_cmd_posted_tlv_v {
u32 sched_cmd_posted[0]; /* HTT_TX_PDEV_SCHED_TX_MODE_MAX */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_sched_txq_cmd_reaped_tlv_v {
u32 sched_cmd_reaped[0]; /* HTT_TX_PDEV_SCHED_TX_MODE_MAX */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_sched_txq_sched_order_su_tlv_v {
u32 sched_order_su[0]; /* HTT_TX_PDEV_NUM_SCHED_ORDER_LOG */
};
enum htt_sched_txq_sched_ineligibility_tlv_enum {
HTT_SCHED_TID_SKIP_SCHED_MASK_DISABLED = 0,
HTT_SCHED_TID_SKIP_NOTIFY_MPDU,
HTT_SCHED_TID_SKIP_MPDU_STATE_INVALID,
HTT_SCHED_TID_SKIP_SCHED_DISABLED,
HTT_SCHED_TID_SKIP_TQM_BYPASS_CMD_PENDING,
HTT_SCHED_TID_SKIP_SECOND_SU_SCHEDULE,
HTT_SCHED_TID_SKIP_CMD_SLOT_NOT_AVAIL,
HTT_SCHED_TID_SKIP_NO_ENQ,
HTT_SCHED_TID_SKIP_LOW_ENQ,
HTT_SCHED_TID_SKIP_PAUSED,
HTT_SCHED_TID_SKIP_UL,
HTT_SCHED_TID_REMOVE_PAUSED,
HTT_SCHED_TID_REMOVE_NO_ENQ,
HTT_SCHED_TID_REMOVE_UL,
HTT_SCHED_TID_QUERY,
HTT_SCHED_TID_SU_ONLY,
HTT_SCHED_TID_ELIGIBLE,
HTT_SCHED_INELIGIBILITY_MAX,
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_sched_txq_sched_ineligibility_tlv_v {
/* indexed by htt_sched_txq_sched_ineligibility_tlv_enum */
u32 sched_ineligibility[0];
};
struct htt_tx_pdev_stats_sched_per_txq_tlv {
u32 mac_id__txq_id__word;
u32 sched_policy;
u32 last_sched_cmd_posted_timestamp;
u32 last_sched_cmd_compl_timestamp;
u32 sched_2_tac_lwm_count;
u32 sched_2_tac_ring_full;
u32 sched_cmd_post_failure;
u32 num_active_tids;
u32 num_ps_schedules;
u32 sched_cmds_pending;
u32 num_tid_register;
u32 num_tid_unregister;
u32 num_qstats_queried;
u32 qstats_update_pending;
u32 last_qstats_query_timestamp;
u32 num_tqm_cmdq_full;
u32 num_de_sched_algo_trigger;
u32 num_rt_sched_algo_trigger;
u32 num_tqm_sched_algo_trigger;
u32 notify_sched;
u32 dur_based_sendn_term;
};
struct htt_stats_tx_sched_cmn_tlv {
/* BIT [ 7 : 0] :- mac_id
* BIT [31 : 8] :- reserved
*/
u32 mac_id__word;
/* Current timestamp */
u32 current_timestamp;
};
/* == TQM STATS == */
#define HTT_TX_TQM_MAX_GEN_MPDU_END_REASON 16
#define HTT_TX_TQM_MAX_LIST_MPDU_END_REASON 16
#define HTT_TX_TQM_MAX_LIST_MPDU_CNT_HISTOGRAM_BINS 16
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_tqm_gen_mpdu_stats_tlv_v {
u32 gen_mpdu_end_reason[0]; /* HTT_TX_TQM_MAX_GEN_MPDU_END_REASON */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_tqm_list_mpdu_stats_tlv_v {
u32 list_mpdu_end_reason[0]; /* HTT_TX_TQM_MAX_LIST_MPDU_END_REASON */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_tx_tqm_list_mpdu_cnt_tlv_v {
u32 list_mpdu_cnt_hist[0];
/* HTT_TX_TQM_MAX_LIST_MPDU_CNT_HISTOGRAM_BINS */
};
struct htt_tx_tqm_pdev_stats_tlv_v {
u32 msdu_count;
u32 mpdu_count;
u32 remove_msdu;
u32 remove_mpdu;
u32 remove_msdu_ttl;
u32 send_bar;
u32 bar_sync;
u32 notify_mpdu;
u32 sync_cmd;
u32 write_cmd;
u32 hwsch_trigger;
u32 ack_tlv_proc;
u32 gen_mpdu_cmd;
u32 gen_list_cmd;
u32 remove_mpdu_cmd;
u32 remove_mpdu_tried_cmd;
u32 mpdu_queue_stats_cmd;
u32 mpdu_head_info_cmd;
u32 msdu_flow_stats_cmd;
u32 remove_msdu_cmd;
u32 remove_msdu_ttl_cmd;
u32 flush_cache_cmd;
u32 update_mpduq_cmd;
u32 enqueue;
u32 enqueue_notify;
u32 notify_mpdu_at_head;
u32 notify_mpdu_state_valid;
/*
* On receiving TQM_FLOW_NOT_EMPTY_STATUS from TQM, (on MSDUs being enqueued
* the flow is non empty), if the number of MSDUs is greater than the threshold,
* notify is incremented. UDP_THRESH counters are for UDP MSDUs, and NONUDP are
* for non-UDP MSDUs.
* MSDUQ_SWNOTIFY_UDP_THRESH1 threshold - sched_udp_notify1 is incremented
* MSDUQ_SWNOTIFY_UDP_THRESH2 threshold - sched_udp_notify2 is incremented
* MSDUQ_SWNOTIFY_NONUDP_THRESH1 threshold - sched_nonudp_notify1 is incremented
* MSDUQ_SWNOTIFY_NONUDP_THRESH2 threshold - sched_nonudp_notify2 is incremented
*
* Notify signifies that we trigger the scheduler.
*/
u32 sched_udp_notify1;
u32 sched_udp_notify2;
u32 sched_nonudp_notify1;
u32 sched_nonudp_notify2;
};
struct htt_tx_tqm_cmn_stats_tlv {
u32 mac_id__word;
u32 max_cmdq_id;
u32 list_mpdu_cnt_hist_intvl;
/* Global stats */
u32 add_msdu;
u32 q_empty;
u32 q_not_empty;
u32 drop_notification;
u32 desc_threshold;
};
struct htt_tx_tqm_error_stats_tlv {
/* Error stats */
u32 q_empty_failure;
u32 q_not_empty_failure;
u32 add_msdu_failure;
};
/* == TQM CMDQ stats == */
struct htt_tx_tqm_cmdq_status_tlv {
u32 mac_id__cmdq_id__word;
u32 sync_cmd;
u32 write_cmd;
u32 gen_mpdu_cmd;
u32 mpdu_queue_stats_cmd;
u32 mpdu_head_info_cmd;
u32 msdu_flow_stats_cmd;
u32 remove_mpdu_cmd;
u32 remove_msdu_cmd;
u32 flush_cache_cmd;
u32 update_mpduq_cmd;
u32 update_msduq_cmd;
};
/* == TX-DE STATS == */
/* Structures for tx de stats */
struct htt_tx_de_eapol_packets_stats_tlv {
u32 m1_packets;
u32 m2_packets;
u32 m3_packets;
u32 m4_packets;
u32 g1_packets;
u32 g2_packets;
};
struct htt_tx_de_classify_failed_stats_tlv {
u32 ap_bss_peer_not_found;
u32 ap_bcast_mcast_no_peer;
u32 sta_delete_in_progress;
u32 ibss_no_bss_peer;
u32 invalid_vdev_type;
u32 invalid_ast_peer_entry;
u32 peer_entry_invalid;
u32 ethertype_not_ip;
u32 eapol_lookup_failed;
u32 qpeer_not_allow_data;
u32 fse_tid_override;
u32 ipv6_jumbogram_zero_length;
u32 qos_to_non_qos_in_prog;
};
struct htt_tx_de_classify_stats_tlv {
u32 arp_packets;
u32 igmp_packets;
u32 dhcp_packets;
u32 host_inspected;
u32 htt_included;
u32 htt_valid_mcs;
u32 htt_valid_nss;
u32 htt_valid_preamble_type;
u32 htt_valid_chainmask;
u32 htt_valid_guard_interval;
u32 htt_valid_retries;
u32 htt_valid_bw_info;
u32 htt_valid_power;
u32 htt_valid_key_flags;
u32 htt_valid_no_encryption;
u32 fse_entry_count;
u32 fse_priority_be;
u32 fse_priority_high;
u32 fse_priority_low;
u32 fse_traffic_ptrn_be;
u32 fse_traffic_ptrn_over_sub;
u32 fse_traffic_ptrn_bursty;
u32 fse_traffic_ptrn_interactive;
u32 fse_traffic_ptrn_periodic;
u32 fse_hwqueue_alloc;
u32 fse_hwqueue_created;
u32 fse_hwqueue_send_to_host;
u32 mcast_entry;
u32 bcast_entry;
u32 htt_update_peer_cache;
u32 htt_learning_frame;
u32 fse_invalid_peer;
/*
* mec_notify is HTT TX WBM multicast echo check notification
* from firmware to host. FW sends SA addresses to host for all
* multicast/broadcast packets received on STA side.
*/
u32 mec_notify;
};
struct htt_tx_de_classify_status_stats_tlv {
u32 eok;
u32 classify_done;
u32 lookup_failed;
u32 send_host_dhcp;
u32 send_host_mcast;
u32 send_host_unknown_dest;
u32 send_host;
u32 status_invalid;
};
struct htt_tx_de_enqueue_packets_stats_tlv {
u32 enqueued_pkts;
u32 to_tqm;
u32 to_tqm_bypass;
};
struct htt_tx_de_enqueue_discard_stats_tlv {
u32 discarded_pkts;
u32 local_frames;
u32 is_ext_msdu;
};
struct htt_tx_de_compl_stats_tlv {
u32 tcl_dummy_frame;
u32 tqm_dummy_frame;
u32 tqm_notify_frame;
u32 fw2wbm_enq;
u32 tqm_bypass_frame;
};
/*
* The htt_tx_de_fw2wbm_ring_full_hist_tlv is a histogram of time we waited
* for the fw2wbm ring buffer. we are requesting a buffer in FW2WBM release
* ring,which may fail, due to non availability of buffer. Hence we sleep for
* 200us & again request for it. This is a histogram of time we wait, with
* bin of 200ms & there are 10 bin (2 seconds max)
* They are defined by the following macros in FW
* #define ENTRIES_PER_BIN_COUNT 1000 // per bin 1000 * 200us = 200ms
* #define RING_FULL_BIN_ENTRIES (WAL_TX_DE_FW2WBM_ALLOC_TIMEOUT_COUNT /
* ENTRIES_PER_BIN_COUNT)
*/
struct htt_tx_de_fw2wbm_ring_full_hist_tlv {
u32 fw2wbm_ring_full_hist[0];
};
struct htt_tx_de_cmn_stats_tlv {
u32 mac_id__word;
/* Global Stats */
u32 tcl2fw_entry_count;
u32 not_to_fw;
u32 invalid_pdev_vdev_peer;
u32 tcl_res_invalid_addrx;
u32 wbm2fw_entry_count;
u32 invalid_pdev;
};
/* == RING-IF STATS == */
#define HTT_STATS_LOW_WM_BINS 5
#define HTT_STATS_HIGH_WM_BINS 5
struct htt_ring_if_stats_tlv {
u32 base_addr; /* DWORD aligned base memory address of the ring */
u32 elem_size;
u32 num_elems__prefetch_tail_idx;
u32 head_idx__tail_idx;
u32 shadow_head_idx__shadow_tail_idx;
u32 num_tail_incr;
u32 lwm_thresh__hwm_thresh;
u32 overrun_hit_count;
u32 underrun_hit_count;
u32 prod_blockwait_count;
u32 cons_blockwait_count;
u32 low_wm_hit_count[HTT_STATS_LOW_WM_BINS];
u32 high_wm_hit_count[HTT_STATS_HIGH_WM_BINS];
};
struct htt_ring_if_cmn_tlv {
u32 mac_id__word;
u32 num_records;
};
/* == SFM STATS == */
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_sfm_client_user_tlv_v {
/* Number of DWORDS used per user and per client */
u32 dwords_used_by_user_n[0];
};
struct htt_sfm_client_tlv {
/* Client ID */
u32 client_id;
/* Minimum number of buffers */
u32 buf_min;
/* Maximum number of buffers */
u32 buf_max;
/* Number of Busy buffers */
u32 buf_busy;
/* Number of Allocated buffers */
u32 buf_alloc;
/* Number of Available/Usable buffers */
u32 buf_avail;
/* Number of users */
u32 num_users;
};
struct htt_sfm_cmn_tlv {
u32 mac_id__word;
/* Indicates the total number of 128 byte buffers
* in the CMEM that are available for buffer sharing
*/
u32 buf_total;
/* Indicates for certain client or all the clients
* there is no dowrd saved in SFM, refer to SFM_R1_MEM_EMPTY
*/
u32 mem_empty;
/* DEALLOCATE_BUFFERS, refer to register SFM_R0_DEALLOCATE_BUFFERS */
u32 deallocate_bufs;
/* Number of Records */
u32 num_records;
};
/* == SRNG STATS == */
struct htt_sring_stats_tlv {
u32 mac_id__ring_id__arena__ep;
u32 base_addr_lsb; /* DWORD aligned base memory address of the ring */
u32 base_addr_msb;
u32 ring_size;
u32 elem_size;
u32 num_avail_words__num_valid_words;
u32 head_ptr__tail_ptr;
u32 consumer_empty__producer_full;
u32 prefetch_count__internal_tail_ptr;
};
struct htt_sring_cmn_tlv {
u32 num_records;
};
/* == PDEV TX RATE CTRL STATS == */
#define HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS 12
#define HTT_TX_PDEV_STATS_NUM_GI_COUNTERS 4
#define HTT_TX_PDEV_STATS_NUM_DCM_COUNTERS 5
#define HTT_TX_PDEV_STATS_NUM_BW_COUNTERS 4
#define HTT_TX_PDEV_STATS_NUM_SPATIAL_STREAMS 8
#define HTT_TX_PDEV_STATS_NUM_PREAMBLE_TYPES HTT_STATS_PREAM_COUNT
#define HTT_TX_PDEV_STATS_NUM_LEGACY_CCK_STATS 4
#define HTT_TX_PDEV_STATS_NUM_LEGACY_OFDM_STATS 8
#define HTT_TX_PDEV_STATS_NUM_LTF 4
#define HTT_TX_NUM_OF_SOUNDING_STATS_WORDS \
(HTT_TX_PDEV_STATS_NUM_BW_COUNTERS * \
HTT_TX_PDEV_STATS_NUM_AX_MUMIMO_USER_STATS)
struct htt_tx_pdev_rate_stats_tlv {
u32 mac_id__word;
u32 tx_ldpc;
u32 rts_cnt;
/* RSSI value of last ack packet (units = dB above noise floor) */
u32 ack_rssi;
u32 tx_mcs[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 tx_su_mcs[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 tx_mu_mcs[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
/* element 0,1, ...7 -> NSS 1,2, ...8 */
u32 tx_nss[HTT_TX_PDEV_STATS_NUM_SPATIAL_STREAMS];
/* element 0: 20 MHz, 1: 40 MHz, 2: 80 MHz, 3: 160 and 80+80 MHz */
u32 tx_bw[HTT_TX_PDEV_STATS_NUM_BW_COUNTERS];
u32 tx_stbc[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 tx_pream[HTT_TX_PDEV_STATS_NUM_PREAMBLE_TYPES];
/* Counters to track number of tx packets
* in each GI (400us, 800us, 1600us & 3200us) in each mcs (0-11)
*/
u32 tx_gi[HTT_TX_PDEV_STATS_NUM_GI_COUNTERS][HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
/* Counters to track packets in dcm mcs (MCS 0, 1, 3, 4) */
u32 tx_dcm[HTT_TX_PDEV_STATS_NUM_DCM_COUNTERS];
/* Number of CTS-acknowledged RTS packets */
u32 rts_success;
/*
* Counters for legacy 11a and 11b transmissions.
*
* The index corresponds to:
*
* CCK: 0: 1 Mbps, 1: 2 Mbps, 2: 5.5 Mbps, 3: 11 Mbps
*
* OFDM: 0: 6 Mbps, 1: 9 Mbps, 2: 12 Mbps, 3: 18 Mbps,
* 4: 24 Mbps, 5: 36 Mbps, 6: 48 Mbps, 7: 54 Mbps
*/
u32 tx_legacy_cck_rate[HTT_TX_PDEV_STATS_NUM_LEGACY_CCK_STATS];
u32 tx_legacy_ofdm_rate[HTT_TX_PDEV_STATS_NUM_LEGACY_OFDM_STATS];
u32 ac_mu_mimo_tx_ldpc;
u32 ax_mu_mimo_tx_ldpc;
u32 ofdma_tx_ldpc;
/*
* Counters for 11ax HE LTF selection during TX.
*
* The index corresponds to:
*
* 0: unused, 1: 1x LTF, 2: 2x LTF, 3: 4x LTF
*/
u32 tx_he_ltf[HTT_TX_PDEV_STATS_NUM_LTF];
u32 ac_mu_mimo_tx_mcs[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 ax_mu_mimo_tx_mcs[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 ofdma_tx_mcs[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 ac_mu_mimo_tx_nss[HTT_TX_PDEV_STATS_NUM_SPATIAL_STREAMS];
u32 ax_mu_mimo_tx_nss[HTT_TX_PDEV_STATS_NUM_SPATIAL_STREAMS];
u32 ofdma_tx_nss[HTT_TX_PDEV_STATS_NUM_SPATIAL_STREAMS];
u32 ac_mu_mimo_tx_bw[HTT_TX_PDEV_STATS_NUM_BW_COUNTERS];
u32 ax_mu_mimo_tx_bw[HTT_TX_PDEV_STATS_NUM_BW_COUNTERS];
u32 ofdma_tx_bw[HTT_TX_PDEV_STATS_NUM_BW_COUNTERS];
u32 ac_mu_mimo_tx_gi[HTT_TX_PDEV_STATS_NUM_GI_COUNTERS]
[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 ax_mu_mimo_tx_gi[HTT_TX_PDEV_STATS_NUM_GI_COUNTERS]
[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 ofdma_tx_gi[HTT_TX_PDEV_STATS_NUM_GI_COUNTERS]
[HTT_TX_PDEV_STATS_NUM_MCS_COUNTERS];
};
/* == PDEV RX RATE CTRL STATS == */
#define HTT_RX_PDEV_STATS_NUM_LEGACY_CCK_STATS 4
#define HTT_RX_PDEV_STATS_NUM_LEGACY_OFDM_STATS 8
#define HTT_RX_PDEV_STATS_NUM_MCS_COUNTERS 12
#define HTT_RX_PDEV_STATS_NUM_GI_COUNTERS 4
#define HTT_RX_PDEV_STATS_NUM_DCM_COUNTERS 5
#define HTT_RX_PDEV_STATS_NUM_BW_COUNTERS 4
#define HTT_RX_PDEV_STATS_NUM_SPATIAL_STREAMS 8
#define HTT_RX_PDEV_STATS_NUM_PREAMBLE_TYPES HTT_STATS_PREAM_COUNT
#define HTT_RX_PDEV_MAX_OFDMA_NUM_USER 8
#define HTT_RX_PDEV_STATS_RXEVM_MAX_PILOTS_PER_NSS 16
struct htt_rx_pdev_rate_stats_tlv {
u32 mac_id__word;
u32 nsts;
u32 rx_ldpc;
u32 rts_cnt;
u32 rssi_mgmt; /* units = dB above noise floor */
u32 rssi_data; /* units = dB above noise floor */
u32 rssi_comb; /* units = dB above noise floor */
u32 rx_mcs[HTT_RX_PDEV_STATS_NUM_MCS_COUNTERS];
/* element 0,1, ...7 -> NSS 1,2, ...8 */
u32 rx_nss[HTT_RX_PDEV_STATS_NUM_SPATIAL_STREAMS];
u32 rx_dcm[HTT_RX_PDEV_STATS_NUM_DCM_COUNTERS];
u32 rx_stbc[HTT_RX_PDEV_STATS_NUM_MCS_COUNTERS];
/* element 0: 20 MHz, 1: 40 MHz, 2: 80 MHz, 3: 160 and 80+80 MHz */
u32 rx_bw[HTT_RX_PDEV_STATS_NUM_BW_COUNTERS];
u32 rx_pream[HTT_RX_PDEV_STATS_NUM_PREAMBLE_TYPES];
u8 rssi_chain[HTT_RX_PDEV_STATS_NUM_SPATIAL_STREAMS]
[HTT_RX_PDEV_STATS_NUM_BW_COUNTERS];
/* units = dB above noise floor */
/* Counters to track number of rx packets
* in each GI in each mcs (0-11)
*/
u32 rx_gi[HTT_RX_PDEV_STATS_NUM_GI_COUNTERS][HTT_RX_PDEV_STATS_NUM_MCS_COUNTERS];
s32 rssi_in_dbm; /* rx Signal Strength value in dBm unit */
u32 rx_11ax_su_ext;
u32 rx_11ac_mumimo;
u32 rx_11ax_mumimo;
u32 rx_11ax_ofdma;
u32 txbf;
u32 rx_legacy_cck_rate[HTT_RX_PDEV_STATS_NUM_LEGACY_CCK_STATS];
u32 rx_legacy_ofdm_rate[HTT_RX_PDEV_STATS_NUM_LEGACY_OFDM_STATS];
u32 rx_active_dur_us_low;
u32 rx_active_dur_us_high;
u32 rx_11ax_ul_ofdma;
u32 ul_ofdma_rx_mcs[HTT_RX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 ul_ofdma_rx_gi[HTT_TX_PDEV_STATS_NUM_GI_COUNTERS]
[HTT_RX_PDEV_STATS_NUM_MCS_COUNTERS];
u32 ul_ofdma_rx_nss[HTT_TX_PDEV_STATS_NUM_SPATIAL_STREAMS];
u32 ul_ofdma_rx_bw[HTT_TX_PDEV_STATS_NUM_BW_COUNTERS];
u32 ul_ofdma_rx_stbc;
u32 ul_ofdma_rx_ldpc;
/* record the stats for each user index */
u32 rx_ulofdma_non_data_ppdu[HTT_RX_PDEV_MAX_OFDMA_NUM_USER]; /* ppdu level */
u32 rx_ulofdma_data_ppdu[HTT_RX_PDEV_MAX_OFDMA_NUM_USER]; /* ppdu level */
u32 rx_ulofdma_mpdu_ok[HTT_RX_PDEV_MAX_OFDMA_NUM_USER]; /* mpdu level */
u32 rx_ulofdma_mpdu_fail[HTT_RX_PDEV_MAX_OFDMA_NUM_USER]; /* mpdu level */
u32 nss_count;
u32 pilot_count;
/* RxEVM stats in dB */
s32 rx_pilot_evm_db[HTT_RX_PDEV_STATS_NUM_SPATIAL_STREAMS]
[HTT_RX_PDEV_STATS_RXEVM_MAX_PILOTS_PER_NSS];
/* rx_pilot_evm_db_mean:
* EVM mean across pilots, computed as
* mean(10*log10(rx_pilot_evm_linear)) = mean(rx_pilot_evm_db)
*/
s32 rx_pilot_evm_db_mean[HTT_RX_PDEV_STATS_NUM_SPATIAL_STREAMS];
s8 rx_ul_fd_rssi[HTT_RX_PDEV_STATS_NUM_SPATIAL_STREAMS]
[HTT_RX_PDEV_MAX_OFDMA_NUM_USER]; /* dBm units */
/* per_chain_rssi_pkt_type:
* This field shows what type of rx frame the per-chain RSSI was computed
* on, by recording the frame type and sub-type as bit-fields within this
* field:
* BIT [3 : 0] :- IEEE80211_FC0_TYPE
* BIT [7 : 4] :- IEEE80211_FC0_SUBTYPE
* BIT [31 : 8] :- Reserved
*/
u32 per_chain_rssi_pkt_type;
s8 rx_per_chain_rssi_in_dbm[HTT_RX_PDEV_STATS_NUM_SPATIAL_STREAMS]
[HTT_RX_PDEV_STATS_NUM_BW_COUNTERS];
};
/* == RX PDEV/SOC STATS == */
struct htt_rx_soc_fw_stats_tlv {
u32 fw_reo_ring_data_msdu;
u32 fw_to_host_data_msdu_bcmc;
u32 fw_to_host_data_msdu_uc;
u32 ofld_remote_data_buf_recycle_cnt;
u32 ofld_remote_free_buf_indication_cnt;
u32 ofld_buf_to_host_data_msdu_uc;
u32 reo_fw_ring_to_host_data_msdu_uc;
u32 wbm_sw_ring_reap;
u32 wbm_forward_to_host_cnt;
u32 wbm_target_recycle_cnt;
u32 target_refill_ring_recycle_cnt;
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_rx_soc_fw_refill_ring_empty_tlv_v {
u32 refill_ring_empty_cnt[0]; /* HTT_RX_STATS_REFILL_MAX_RING */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_rx_soc_fw_refill_ring_num_refill_tlv_v {
u32 refill_ring_num_refill[0]; /* HTT_RX_STATS_REFILL_MAX_RING */
};
/* RXDMA error code from WBM released packets */
enum htt_rx_rxdma_error_code_enum {
HTT_RX_RXDMA_OVERFLOW_ERR = 0,
HTT_RX_RXDMA_MPDU_LENGTH_ERR = 1,
HTT_RX_RXDMA_FCS_ERR = 2,
HTT_RX_RXDMA_DECRYPT_ERR = 3,
HTT_RX_RXDMA_TKIP_MIC_ERR = 4,
HTT_RX_RXDMA_UNECRYPTED_ERR = 5,
HTT_RX_RXDMA_MSDU_LEN_ERR = 6,
HTT_RX_RXDMA_MSDU_LIMIT_ERR = 7,
HTT_RX_RXDMA_WIFI_PARSE_ERR = 8,
HTT_RX_RXDMA_AMSDU_PARSE_ERR = 9,
HTT_RX_RXDMA_SA_TIMEOUT_ERR = 10,
HTT_RX_RXDMA_DA_TIMEOUT_ERR = 11,
HTT_RX_RXDMA_FLOW_TIMEOUT_ERR = 12,
HTT_RX_RXDMA_FLUSH_REQUEST = 13,
HTT_RX_RXDMA_ERR_CODE_RVSD0 = 14,
HTT_RX_RXDMA_ERR_CODE_RVSD1 = 15,
/* This MAX_ERR_CODE should not be used in any host/target messages,
* so that even though it is defined within a host/target interface
* definition header file, it isn't actually part of the host/target
* interface, and thus can be modified.
*/
HTT_RX_RXDMA_MAX_ERR_CODE
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_rx_soc_fw_refill_ring_num_rxdma_err_tlv_v {
u32 rxdma_err[0]; /* HTT_RX_RXDMA_MAX_ERR_CODE */
};
/* REO error code from WBM released packets */
enum htt_rx_reo_error_code_enum {
HTT_RX_REO_QUEUE_DESC_ADDR_ZERO = 0,
HTT_RX_REO_QUEUE_DESC_NOT_VALID = 1,
HTT_RX_AMPDU_IN_NON_BA = 2,
HTT_RX_NON_BA_DUPLICATE = 3,
HTT_RX_BA_DUPLICATE = 4,
HTT_RX_REGULAR_FRAME_2K_JUMP = 5,
HTT_RX_BAR_FRAME_2K_JUMP = 6,
HTT_RX_REGULAR_FRAME_OOR = 7,
HTT_RX_BAR_FRAME_OOR = 8,
HTT_RX_BAR_FRAME_NO_BA_SESSION = 9,
HTT_RX_BAR_FRAME_SN_EQUALS_SSN = 10,
HTT_RX_PN_CHECK_FAILED = 11,
HTT_RX_2K_ERROR_HANDLING_FLAG_SET = 12,
HTT_RX_PN_ERROR_HANDLING_FLAG_SET = 13,
HTT_RX_QUEUE_DESCRIPTOR_BLOCKED_SET = 14,
HTT_RX_REO_ERR_CODE_RVSD = 15,
/* This MAX_ERR_CODE should not be used in any host/target messages,
* so that even though it is defined within a host/target interface
* definition header file, it isn't actually part of the host/target
* interface, and thus can be modified.
*/
HTT_RX_REO_MAX_ERR_CODE
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_rx_soc_fw_refill_ring_num_reo_err_tlv_v {
u32 reo_err[0]; /* HTT_RX_REO_MAX_ERR_CODE */
};
/* == RX PDEV STATS == */
#define HTT_STATS_SUBTYPE_MAX 16
struct htt_rx_pdev_fw_stats_tlv {
u32 mac_id__word;
u32 ppdu_recvd;
u32 mpdu_cnt_fcs_ok;
u32 mpdu_cnt_fcs_err;
u32 tcp_msdu_cnt;
u32 tcp_ack_msdu_cnt;
u32 udp_msdu_cnt;
u32 other_msdu_cnt;
u32 fw_ring_mpdu_ind;
u32 fw_ring_mgmt_subtype[HTT_STATS_SUBTYPE_MAX];
u32 fw_ring_ctrl_subtype[HTT_STATS_SUBTYPE_MAX];
u32 fw_ring_mcast_data_msdu;
u32 fw_ring_bcast_data_msdu;
u32 fw_ring_ucast_data_msdu;
u32 fw_ring_null_data_msdu;
u32 fw_ring_mpdu_drop;
u32 ofld_local_data_ind_cnt;
u32 ofld_local_data_buf_recycle_cnt;
u32 drx_local_data_ind_cnt;
u32 drx_local_data_buf_recycle_cnt;
u32 local_nondata_ind_cnt;
u32 local_nondata_buf_recycle_cnt;
u32 fw_status_buf_ring_refill_cnt;
u32 fw_status_buf_ring_empty_cnt;
u32 fw_pkt_buf_ring_refill_cnt;
u32 fw_pkt_buf_ring_empty_cnt;
u32 fw_link_buf_ring_refill_cnt;
u32 fw_link_buf_ring_empty_cnt;
u32 host_pkt_buf_ring_refill_cnt;
u32 host_pkt_buf_ring_empty_cnt;
u32 mon_pkt_buf_ring_refill_cnt;
u32 mon_pkt_buf_ring_empty_cnt;
u32 mon_status_buf_ring_refill_cnt;
u32 mon_status_buf_ring_empty_cnt;
u32 mon_desc_buf_ring_refill_cnt;
u32 mon_desc_buf_ring_empty_cnt;
u32 mon_dest_ring_update_cnt;
u32 mon_dest_ring_full_cnt;
u32 rx_suspend_cnt;
u32 rx_suspend_fail_cnt;
u32 rx_resume_cnt;
u32 rx_resume_fail_cnt;
u32 rx_ring_switch_cnt;
u32 rx_ring_restore_cnt;
u32 rx_flush_cnt;
u32 rx_recovery_reset_cnt;
};
#define HTT_STATS_PHY_ERR_MAX 43
struct htt_rx_pdev_fw_stats_phy_err_tlv {
u32 mac_id__word;
u32 total_phy_err_cnt;
/* Counts of different types of phy errs
* The mapping of PHY error types to phy_err array elements is HW dependent.
* The only currently-supported mapping is shown below:
*
* 0 phyrx_err_phy_off Reception aborted due to receiving a PHY_OFF TLV
* 1 phyrx_err_synth_off
* 2 phyrx_err_ofdma_timing
* 3 phyrx_err_ofdma_signal_parity
* 4 phyrx_err_ofdma_rate_illegal
* 5 phyrx_err_ofdma_length_illegal
* 6 phyrx_err_ofdma_restart
* 7 phyrx_err_ofdma_service
* 8 phyrx_err_ppdu_ofdma_power_drop
* 9 phyrx_err_cck_blokker
* 10 phyrx_err_cck_timing
* 11 phyrx_err_cck_header_crc
* 12 phyrx_err_cck_rate_illegal
* 13 phyrx_err_cck_length_illegal
* 14 phyrx_err_cck_restart
* 15 phyrx_err_cck_service
* 16 phyrx_err_cck_power_drop
* 17 phyrx_err_ht_crc_err
* 18 phyrx_err_ht_length_illegal
* 19 phyrx_err_ht_rate_illegal
* 20 phyrx_err_ht_zlf
* 21 phyrx_err_false_radar_ext
* 22 phyrx_err_green_field
* 23 phyrx_err_bw_gt_dyn_bw
* 24 phyrx_err_leg_ht_mismatch
* 25 phyrx_err_vht_crc_error
* 26 phyrx_err_vht_siga_unsupported
* 27 phyrx_err_vht_lsig_len_invalid
* 28 phyrx_err_vht_ndp_or_zlf
* 29 phyrx_err_vht_nsym_lt_zero
* 30 phyrx_err_vht_rx_extra_symbol_mismatch
* 31 phyrx_err_vht_rx_skip_group_id0
* 32 phyrx_err_vht_rx_skip_group_id1to62
* 33 phyrx_err_vht_rx_skip_group_id63
* 34 phyrx_err_ofdm_ldpc_decoder_disabled
* 35 phyrx_err_defer_nap
* 36 phyrx_err_fdomain_timeout
* 37 phyrx_err_lsig_rel_check
* 38 phyrx_err_bt_collision
* 39 phyrx_err_unsupported_mu_feedback
* 40 phyrx_err_ppdu_tx_interrupt_rx
* 41 phyrx_err_unsupported_cbf
* 42 phyrx_err_other
*/
u32 phy_err[HTT_STATS_PHY_ERR_MAX];
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_rx_pdev_fw_ring_mpdu_err_tlv_v {
/* Num error MPDU for each RxDMA error type */
u32 fw_ring_mpdu_err[0]; /* HTT_RX_STATS_RXDMA_MAX_ERR */
};
/* NOTE: Variable length TLV, use length spec to infer array size */
struct htt_rx_pdev_fw_mpdu_drop_tlv_v {
/* Num MPDU dropped */
u32 fw_mpdu_drop[0]; /* HTT_RX_STATS_FW_DROP_REASON_MAX */
};
#define HTT_PDEV_CCA_STATS_TX_FRAME_INFO_PRESENT (0x1)
#define HTT_PDEV_CCA_STATS_RX_FRAME_INFO_PRESENT (0x2)
#define HTT_PDEV_CCA_STATS_RX_CLEAR_INFO_PRESENT (0x4)
#define HTT_PDEV_CCA_STATS_MY_RX_FRAME_INFO_PRESENT (0x8)
#define HTT_PDEV_CCA_STATS_USEC_CNT_INFO_PRESENT (0x10)
#define HTT_PDEV_CCA_STATS_MED_RX_IDLE_INFO_PRESENT (0x20)
#define HTT_PDEV_CCA_STATS_MED_TX_IDLE_GLOBAL_INFO_PRESENT (0x40)
#define HTT_PDEV_CCA_STATS_CCA_OBBS_USEC_INFO_PRESENT (0x80)
struct htt_pdev_stats_cca_counters_tlv {
/* Below values are obtained from the HW Cycles counter registers */
u32 tx_frame_usec;
u32 rx_frame_usec;
u32 rx_clear_usec;
u32 my_rx_frame_usec;
u32 usec_cnt;
u32 med_rx_idle_usec;
u32 med_tx_idle_global_usec;
u32 cca_obss_usec;
};
struct htt_pdev_cca_stats_hist_v1_tlv {
u32 chan_num;
/* num of CCA records (Num of htt_pdev_stats_cca_counters_tlv)*/
u32 num_records;
u32 valid_cca_counters_bitmap;
u32 collection_interval;
/* This will be followed by an array which contains the CCA stats
* collected in the last N intervals,
* if the indication is for last N intervals CCA stats.
* Then the pdev_cca_stats[0] element contains the oldest CCA stats
* and pdev_cca_stats[N-1] will have the most recent CCA stats.
* htt_pdev_stats_cca_counters_tlv cca_hist_tlv[1];
*/
};
struct htt_pdev_stats_twt_session_tlv {
u32 vdev_id;
struct htt_mac_addr peer_mac;
u32 flow_id_flags;
/* TWT_DIALOG_ID_UNAVAILABLE is used
* when TWT session is not initiated by host
*/
u32 dialog_id;
u32 wake_dura_us;
u32 wake_intvl_us;
u32 sp_offset_us;
};
struct htt_pdev_stats_twt_sessions_tlv {
u32 pdev_id;
u32 num_sessions;
struct htt_pdev_stats_twt_session_tlv twt_session[];
};
enum htt_rx_reo_resource_sample_id_enum {
/* Global link descriptor queued in REO */
HTT_RX_REO_RESOURCE_GLOBAL_LINK_DESC_COUNT_0 = 0,
HTT_RX_REO_RESOURCE_GLOBAL_LINK_DESC_COUNT_1 = 1,
HTT_RX_REO_RESOURCE_GLOBAL_LINK_DESC_COUNT_2 = 2,
/*Number of queue descriptors of this aging group */
HTT_RX_REO_RESOURCE_BUFFERS_USED_AC0 = 3,
HTT_RX_REO_RESOURCE_BUFFERS_USED_AC1 = 4,
HTT_RX_REO_RESOURCE_BUFFERS_USED_AC2 = 5,
HTT_RX_REO_RESOURCE_BUFFERS_USED_AC3 = 6,
/* Total number of MSDUs buffered in AC */
HTT_RX_REO_RESOURCE_AGING_NUM_QUEUES_AC0 = 7,
HTT_RX_REO_RESOURCE_AGING_NUM_QUEUES_AC1 = 8,
HTT_RX_REO_RESOURCE_AGING_NUM_QUEUES_AC2 = 9,
HTT_RX_REO_RESOURCE_AGING_NUM_QUEUES_AC3 = 10,
HTT_RX_REO_RESOURCE_STATS_MAX = 16
};
struct htt_rx_reo_resource_stats_tlv_v {
/* Variable based on the Number of records. HTT_RX_REO_RESOURCE_STATS_MAX */
u32 sample_id;
u32 total_max;
u32 total_avg;
u32 total_sample;
u32 non_zeros_avg;
u32 non_zeros_sample;
u32 last_non_zeros_max;
u32 last_non_zeros_min;
u32 last_non_zeros_avg;
u32 last_non_zeros_sample;
};
/* == TX SOUNDING STATS == */
enum htt_txbf_sound_steer_modes {
HTT_IMPLICIT_TXBF_STEER_STATS = 0,
HTT_EXPLICIT_TXBF_SU_SIFS_STEER_STATS = 1,
HTT_EXPLICIT_TXBF_SU_RBO_STEER_STATS = 2,
HTT_EXPLICIT_TXBF_MU_SIFS_STEER_STATS = 3,
HTT_EXPLICIT_TXBF_MU_RBO_STEER_STATS = 4,
HTT_TXBF_MAX_NUM_OF_MODES = 5
};
enum htt_stats_sounding_tx_mode {
HTT_TX_AC_SOUNDING_MODE = 0,
HTT_TX_AX_SOUNDING_MODE = 1,
};
struct htt_tx_sounding_stats_tlv {
u32 tx_sounding_mode; /* HTT_TX_XX_SOUNDING_MODE */
/* Counts number of soundings for all steering modes in each bw */
u32 cbf_20[HTT_TXBF_MAX_NUM_OF_MODES];
u32 cbf_40[HTT_TXBF_MAX_NUM_OF_MODES];
u32 cbf_80[HTT_TXBF_MAX_NUM_OF_MODES];
u32 cbf_160[HTT_TXBF_MAX_NUM_OF_MODES];
/*
* The sounding array is a 2-D array stored as an 1-D array of
* u32. The stats for a particular user/bw combination is
* referenced with the following:
*
* sounding[(user* max_bw) + bw]
*
* ... where max_bw == 4 for 160mhz
*/
u32 sounding[HTT_TX_NUM_OF_SOUNDING_STATS_WORDS];
};
struct htt_pdev_obss_pd_stats_tlv {
u32 num_obss_tx_ppdu_success;
u32 num_obss_tx_ppdu_failure;
u32 num_sr_tx_transmissions;
u32 num_spatial_reuse_opportunities;
u32 num_non_srg_opportunities;
u32 num_non_srg_ppdu_tried;
u32 num_non_srg_ppdu_success;
u32 num_srg_opportunities;
u32 num_srg_ppdu_tried;
u32 num_srg_ppdu_success;
u32 num_psr_opportunities;
u32 num_psr_ppdu_tried;
u32 num_psr_ppdu_success;
};
struct htt_ring_backpressure_stats_tlv {
u32 pdev_id;
u32 current_head_idx;
u32 current_tail_idx;
u32 num_htt_msgs_sent;
/* Time in milliseconds for which the ring has been in
* its current backpressure condition
*/
u32 backpressure_time_ms;
/* backpressure_hist - histogram showing how many times
* different degrees of backpressure duration occurred:
* Index 0 indicates the number of times ring was
* continuously in backpressure state for 100 - 200ms.
* Index 1 indicates the number of times ring was
* continuously in backpressure state for 200 - 300ms.
* Index 2 indicates the number of times ring was
* continuously in backpressure state for 300 - 400ms.
* Index 3 indicates the number of times ring was
* continuously in backpressure state for 400 - 500ms.
* Index 4 indicates the number of times ring was
* continuously in backpressure state beyond 500ms.
*/
u32 backpressure_hist[5];
};
#ifdef CONFIG_ATH11K_DEBUGFS
void ath11k_debugfs_htt_stats_init(struct ath11k *ar);
void ath11k_debugfs_htt_ext_stats_handler(struct ath11k_base *ab,
struct sk_buff *skb);
int ath11k_debugfs_htt_stats_req(struct ath11k *ar);
#else /* CONFIG_ATH11K_DEBUGFS */
static inline void ath11k_debugfs_htt_stats_init(struct ath11k *ar)
{
}
static inline void ath11k_debugfs_htt_ext_stats_handler(struct ath11k_base *ab,
struct sk_buff *skb)
{
}
static inline int ath11k_debugfs_htt_stats_req(struct ath11k *ar)
{
return 0;
}
#endif /* CONFIG_ATH11K_DEBUGFS */
#endif
|