Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
freebsd
GitHub Repository: freebsd/freebsd-src
Path: blob/main/sys/contrib/dev/athk/ath12k/hal.c
48375 views
1
// SPDX-License-Identifier: BSD-3-Clause-Clear
2
/*
3
* Copyright (c) 2018-2021 The Linux Foundation. All rights reserved.
4
* Copyright (c) 2021-2022 Qualcomm Innovation Center, Inc. All rights reserved.
5
*/
6
#include <linux/dma-mapping.h>
7
#include "hal_tx.h"
8
#include "hal_rx.h"
9
#include "debug.h"
10
#include "hal_desc.h"
11
#include "hif.h"
12
13
static const struct hal_srng_config hw_srng_config_template[] = {
14
/* TODO: max_rings can populated by querying HW capabilities */
15
[HAL_REO_DST] = {
16
.start_ring_id = HAL_SRNG_RING_ID_REO2SW1,
17
.max_rings = 8,
18
.entry_size = sizeof(struct hal_reo_dest_ring) >> 2,
19
.mac_type = ATH12K_HAL_SRNG_UMAC,
20
.ring_dir = HAL_SRNG_DIR_DST,
21
.max_size = HAL_REO_REO2SW1_RING_BASE_MSB_RING_SIZE,
22
},
23
[HAL_REO_EXCEPTION] = {
24
/* Designating REO2SW0 ring as exception ring.
25
* Any of theREO2SW rings can be used as exception ring.
26
*/
27
.start_ring_id = HAL_SRNG_RING_ID_REO2SW0,
28
.max_rings = 1,
29
.entry_size = sizeof(struct hal_reo_dest_ring) >> 2,
30
.mac_type = ATH12K_HAL_SRNG_UMAC,
31
.ring_dir = HAL_SRNG_DIR_DST,
32
.max_size = HAL_REO_REO2SW0_RING_BASE_MSB_RING_SIZE,
33
},
34
[HAL_REO_REINJECT] = {
35
.start_ring_id = HAL_SRNG_RING_ID_SW2REO,
36
.max_rings = 4,
37
.entry_size = sizeof(struct hal_reo_entrance_ring) >> 2,
38
.mac_type = ATH12K_HAL_SRNG_UMAC,
39
.ring_dir = HAL_SRNG_DIR_SRC,
40
.max_size = HAL_REO_SW2REO_RING_BASE_MSB_RING_SIZE,
41
},
42
[HAL_REO_CMD] = {
43
.start_ring_id = HAL_SRNG_RING_ID_REO_CMD,
44
.max_rings = 1,
45
.entry_size = (sizeof(struct hal_tlv_64_hdr) +
46
sizeof(struct hal_reo_get_queue_stats)) >> 2,
47
.mac_type = ATH12K_HAL_SRNG_UMAC,
48
.ring_dir = HAL_SRNG_DIR_SRC,
49
.max_size = HAL_REO_CMD_RING_BASE_MSB_RING_SIZE,
50
},
51
[HAL_REO_STATUS] = {
52
.start_ring_id = HAL_SRNG_RING_ID_REO_STATUS,
53
.max_rings = 1,
54
.entry_size = (sizeof(struct hal_tlv_64_hdr) +
55
sizeof(struct hal_reo_get_queue_stats_status)) >> 2,
56
.mac_type = ATH12K_HAL_SRNG_UMAC,
57
.ring_dir = HAL_SRNG_DIR_DST,
58
.max_size = HAL_REO_STATUS_RING_BASE_MSB_RING_SIZE,
59
},
60
[HAL_TCL_DATA] = {
61
.start_ring_id = HAL_SRNG_RING_ID_SW2TCL1,
62
.max_rings = 6,
63
.entry_size = sizeof(struct hal_tcl_data_cmd) >> 2,
64
.mac_type = ATH12K_HAL_SRNG_UMAC,
65
.ring_dir = HAL_SRNG_DIR_SRC,
66
.max_size = HAL_SW2TCL1_RING_BASE_MSB_RING_SIZE,
67
},
68
[HAL_TCL_CMD] = {
69
.start_ring_id = HAL_SRNG_RING_ID_SW2TCL_CMD,
70
.max_rings = 1,
71
.entry_size = sizeof(struct hal_tcl_gse_cmd) >> 2,
72
.mac_type = ATH12K_HAL_SRNG_UMAC,
73
.ring_dir = HAL_SRNG_DIR_SRC,
74
.max_size = HAL_SW2TCL1_CMD_RING_BASE_MSB_RING_SIZE,
75
},
76
[HAL_TCL_STATUS] = {
77
.start_ring_id = HAL_SRNG_RING_ID_TCL_STATUS,
78
.max_rings = 1,
79
.entry_size = (sizeof(struct hal_tlv_hdr) +
80
sizeof(struct hal_tcl_status_ring)) >> 2,
81
.mac_type = ATH12K_HAL_SRNG_UMAC,
82
.ring_dir = HAL_SRNG_DIR_DST,
83
.max_size = HAL_TCL_STATUS_RING_BASE_MSB_RING_SIZE,
84
},
85
[HAL_CE_SRC] = {
86
.start_ring_id = HAL_SRNG_RING_ID_CE0_SRC,
87
.max_rings = 16,
88
.entry_size = sizeof(struct hal_ce_srng_src_desc) >> 2,
89
.mac_type = ATH12K_HAL_SRNG_UMAC,
90
.ring_dir = HAL_SRNG_DIR_SRC,
91
.max_size = HAL_CE_SRC_RING_BASE_MSB_RING_SIZE,
92
},
93
[HAL_CE_DST] = {
94
.start_ring_id = HAL_SRNG_RING_ID_CE0_DST,
95
.max_rings = 16,
96
.entry_size = sizeof(struct hal_ce_srng_dest_desc) >> 2,
97
.mac_type = ATH12K_HAL_SRNG_UMAC,
98
.ring_dir = HAL_SRNG_DIR_SRC,
99
.max_size = HAL_CE_DST_RING_BASE_MSB_RING_SIZE,
100
},
101
[HAL_CE_DST_STATUS] = {
102
.start_ring_id = HAL_SRNG_RING_ID_CE0_DST_STATUS,
103
.max_rings = 16,
104
.entry_size = sizeof(struct hal_ce_srng_dst_status_desc) >> 2,
105
.mac_type = ATH12K_HAL_SRNG_UMAC,
106
.ring_dir = HAL_SRNG_DIR_DST,
107
.max_size = HAL_CE_DST_STATUS_RING_BASE_MSB_RING_SIZE,
108
},
109
[HAL_WBM_IDLE_LINK] = {
110
.start_ring_id = HAL_SRNG_RING_ID_WBM_IDLE_LINK,
111
.max_rings = 1,
112
.entry_size = sizeof(struct hal_wbm_link_desc) >> 2,
113
.mac_type = ATH12K_HAL_SRNG_UMAC,
114
.ring_dir = HAL_SRNG_DIR_SRC,
115
.max_size = HAL_WBM_IDLE_LINK_RING_BASE_MSB_RING_SIZE,
116
},
117
[HAL_SW2WBM_RELEASE] = {
118
.start_ring_id = HAL_SRNG_RING_ID_WBM_SW0_RELEASE,
119
.max_rings = 2,
120
.entry_size = sizeof(struct hal_wbm_release_ring) >> 2,
121
.mac_type = ATH12K_HAL_SRNG_UMAC,
122
.ring_dir = HAL_SRNG_DIR_SRC,
123
.max_size = HAL_SW2WBM_RELEASE_RING_BASE_MSB_RING_SIZE,
124
},
125
[HAL_WBM2SW_RELEASE] = {
126
.start_ring_id = HAL_SRNG_RING_ID_WBM2SW0_RELEASE,
127
.max_rings = 8,
128
.entry_size = sizeof(struct hal_wbm_release_ring) >> 2,
129
.mac_type = ATH12K_HAL_SRNG_UMAC,
130
.ring_dir = HAL_SRNG_DIR_DST,
131
.max_size = HAL_WBM2SW_RELEASE_RING_BASE_MSB_RING_SIZE,
132
},
133
[HAL_RXDMA_BUF] = {
134
.start_ring_id = HAL_SRNG_SW2RXDMA_BUF0,
135
.max_rings = 1,
136
.entry_size = sizeof(struct hal_wbm_buffer_ring) >> 2,
137
.mac_type = ATH12K_HAL_SRNG_DMAC,
138
.ring_dir = HAL_SRNG_DIR_SRC,
139
.max_size = HAL_RXDMA_RING_MAX_SIZE_BE,
140
},
141
[HAL_RXDMA_DST] = {
142
.start_ring_id = HAL_SRNG_RING_ID_WMAC1_RXDMA2SW0,
143
.max_rings = 0,
144
.entry_size = 0,
145
.mac_type = ATH12K_HAL_SRNG_PMAC,
146
.ring_dir = HAL_SRNG_DIR_DST,
147
.max_size = HAL_RXDMA_RING_MAX_SIZE_BE,
148
},
149
[HAL_RXDMA_MONITOR_BUF] = {
150
.start_ring_id = HAL_SRNG_SW2RXMON_BUF0,
151
.max_rings = 1,
152
.entry_size = sizeof(struct hal_mon_buf_ring) >> 2,
153
.mac_type = ATH12K_HAL_SRNG_PMAC,
154
.ring_dir = HAL_SRNG_DIR_SRC,
155
.max_size = HAL_RXDMA_RING_MAX_SIZE_BE,
156
},
157
[HAL_RXDMA_MONITOR_STATUS] = { 0, },
158
[HAL_RXDMA_MONITOR_DESC] = { 0, },
159
[HAL_RXDMA_DIR_BUF] = {
160
.start_ring_id = HAL_SRNG_RING_ID_RXDMA_DIR_BUF,
161
.max_rings = 2,
162
.entry_size = 8 >> 2, /* TODO: Define the struct */
163
.mac_type = ATH12K_HAL_SRNG_PMAC,
164
.ring_dir = HAL_SRNG_DIR_SRC,
165
.max_size = HAL_RXDMA_RING_MAX_SIZE_BE,
166
},
167
[HAL_PPE2TCL] = {
168
.start_ring_id = HAL_SRNG_RING_ID_PPE2TCL1,
169
.max_rings = 1,
170
.entry_size = sizeof(struct hal_tcl_entrance_from_ppe_ring) >> 2,
171
.mac_type = ATH12K_HAL_SRNG_PMAC,
172
.ring_dir = HAL_SRNG_DIR_SRC,
173
.max_size = HAL_SW2TCL1_RING_BASE_MSB_RING_SIZE,
174
},
175
[HAL_PPE_RELEASE] = {
176
.start_ring_id = HAL_SRNG_RING_ID_WBM_PPE_RELEASE,
177
.max_rings = 1,
178
.entry_size = sizeof(struct hal_wbm_release_ring) >> 2,
179
.mac_type = ATH12K_HAL_SRNG_PMAC,
180
.ring_dir = HAL_SRNG_DIR_SRC,
181
.max_size = HAL_WBM2PPE_RELEASE_RING_BASE_MSB_RING_SIZE,
182
},
183
[HAL_TX_MONITOR_BUF] = {
184
.start_ring_id = HAL_SRNG_SW2TXMON_BUF0,
185
.max_rings = 1,
186
.entry_size = sizeof(struct hal_mon_buf_ring) >> 2,
187
.mac_type = ATH12K_HAL_SRNG_PMAC,
188
.ring_dir = HAL_SRNG_DIR_SRC,
189
.max_size = HAL_RXDMA_RING_MAX_SIZE_BE,
190
},
191
[HAL_RXDMA_MONITOR_DST] = {
192
.start_ring_id = HAL_SRNG_RING_ID_WMAC1_SW2RXMON_BUF0,
193
.max_rings = 1,
194
.entry_size = sizeof(struct hal_mon_dest_desc) >> 2,
195
.mac_type = ATH12K_HAL_SRNG_PMAC,
196
.ring_dir = HAL_SRNG_DIR_DST,
197
.max_size = HAL_RXDMA_RING_MAX_SIZE_BE,
198
},
199
[HAL_TX_MONITOR_DST] = {
200
.start_ring_id = HAL_SRNG_RING_ID_WMAC1_TXMON2SW0_BUF0,
201
.max_rings = 1,
202
.entry_size = sizeof(struct hal_mon_dest_desc) >> 2,
203
.mac_type = ATH12K_HAL_SRNG_PMAC,
204
.ring_dir = HAL_SRNG_DIR_DST,
205
.max_size = HAL_RXDMA_RING_MAX_SIZE_BE,
206
}
207
};
208
209
static const struct ath12k_hal_tcl_to_wbm_rbm_map
210
ath12k_hal_qcn9274_tcl_to_wbm_rbm_map[DP_TCL_NUM_RING_MAX] = {
211
{
212
.wbm_ring_num = 0,
213
.rbm_id = HAL_RX_BUF_RBM_SW0_BM,
214
},
215
{
216
.wbm_ring_num = 1,
217
.rbm_id = HAL_RX_BUF_RBM_SW1_BM,
218
},
219
{
220
.wbm_ring_num = 2,
221
.rbm_id = HAL_RX_BUF_RBM_SW2_BM,
222
},
223
{
224
.wbm_ring_num = 4,
225
.rbm_id = HAL_RX_BUF_RBM_SW4_BM,
226
}
227
};
228
229
static const struct ath12k_hal_tcl_to_wbm_rbm_map
230
ath12k_hal_wcn7850_tcl_to_wbm_rbm_map[DP_TCL_NUM_RING_MAX] = {
231
{
232
.wbm_ring_num = 0,
233
.rbm_id = HAL_RX_BUF_RBM_SW0_BM,
234
},
235
{
236
.wbm_ring_num = 2,
237
.rbm_id = HAL_RX_BUF_RBM_SW2_BM,
238
},
239
{
240
.wbm_ring_num = 4,
241
.rbm_id = HAL_RX_BUF_RBM_SW4_BM,
242
},
243
};
244
245
static unsigned int ath12k_hal_reo1_ring_id_offset(struct ath12k_base *ab)
246
{
247
return HAL_REO1_RING_ID(ab) - HAL_REO1_RING_BASE_LSB(ab);
248
}
249
250
static unsigned int ath12k_hal_reo1_ring_msi1_base_lsb_offset(struct ath12k_base *ab)
251
{
252
return HAL_REO1_RING_MSI1_BASE_LSB(ab) - HAL_REO1_RING_BASE_LSB(ab);
253
}
254
255
static unsigned int ath12k_hal_reo1_ring_msi1_base_msb_offset(struct ath12k_base *ab)
256
{
257
return HAL_REO1_RING_MSI1_BASE_MSB(ab) - HAL_REO1_RING_BASE_LSB(ab);
258
}
259
260
static unsigned int ath12k_hal_reo1_ring_msi1_data_offset(struct ath12k_base *ab)
261
{
262
return HAL_REO1_RING_MSI1_DATA(ab) - HAL_REO1_RING_BASE_LSB(ab);
263
}
264
265
static unsigned int ath12k_hal_reo1_ring_base_msb_offset(struct ath12k_base *ab)
266
{
267
return HAL_REO1_RING_BASE_MSB(ab) - HAL_REO1_RING_BASE_LSB(ab);
268
}
269
270
static unsigned int ath12k_hal_reo1_ring_producer_int_setup_offset(struct ath12k_base *ab)
271
{
272
return HAL_REO1_RING_PRODUCER_INT_SETUP(ab) - HAL_REO1_RING_BASE_LSB(ab);
273
}
274
275
static unsigned int ath12k_hal_reo1_ring_hp_addr_lsb_offset(struct ath12k_base *ab)
276
{
277
return HAL_REO1_RING_HP_ADDR_LSB(ab) - HAL_REO1_RING_BASE_LSB(ab);
278
}
279
280
static unsigned int ath12k_hal_reo1_ring_hp_addr_msb_offset(struct ath12k_base *ab)
281
{
282
return HAL_REO1_RING_HP_ADDR_MSB(ab) - HAL_REO1_RING_BASE_LSB(ab);
283
}
284
285
static unsigned int ath12k_hal_reo1_ring_misc_offset(struct ath12k_base *ab)
286
{
287
return HAL_REO1_RING_MISC(ab) - HAL_REO1_RING_BASE_LSB(ab);
288
}
289
290
static bool ath12k_hw_qcn9274_rx_desc_get_first_msdu(struct hal_rx_desc *desc)
291
{
292
return !!le16_get_bits(desc->u.qcn9274.msdu_end.info5,
293
RX_MSDU_END_INFO5_FIRST_MSDU);
294
}
295
296
static bool ath12k_hw_qcn9274_rx_desc_get_last_msdu(struct hal_rx_desc *desc)
297
{
298
return !!le16_get_bits(desc->u.qcn9274.msdu_end.info5,
299
RX_MSDU_END_INFO5_LAST_MSDU);
300
}
301
302
static u8 ath12k_hw_qcn9274_rx_desc_get_l3_pad_bytes(struct hal_rx_desc *desc)
303
{
304
return le16_get_bits(desc->u.qcn9274.msdu_end.info5,
305
RX_MSDU_END_INFO5_L3_HDR_PADDING);
306
}
307
308
static bool ath12k_hw_qcn9274_rx_desc_encrypt_valid(struct hal_rx_desc *desc)
309
{
310
return !!le32_get_bits(desc->u.qcn9274.mpdu_start.info4,
311
RX_MPDU_START_INFO4_ENCRYPT_INFO_VALID);
312
}
313
314
static u32 ath12k_hw_qcn9274_rx_desc_get_encrypt_type(struct hal_rx_desc *desc)
315
{
316
return le32_get_bits(desc->u.qcn9274.mpdu_start.info2,
317
RX_MPDU_START_INFO2_ENC_TYPE);
318
}
319
320
static u8 ath12k_hw_qcn9274_rx_desc_get_decap_type(struct hal_rx_desc *desc)
321
{
322
return le32_get_bits(desc->u.qcn9274.msdu_end.info11,
323
RX_MSDU_END_INFO11_DECAP_FORMAT);
324
}
325
326
static u8 ath12k_hw_qcn9274_rx_desc_get_mesh_ctl(struct hal_rx_desc *desc)
327
{
328
return le32_get_bits(desc->u.qcn9274.msdu_end.info11,
329
RX_MSDU_END_INFO11_MESH_CTRL_PRESENT);
330
}
331
332
static bool ath12k_hw_qcn9274_rx_desc_get_mpdu_seq_ctl_vld(struct hal_rx_desc *desc)
333
{
334
return !!le32_get_bits(desc->u.qcn9274.mpdu_start.info4,
335
RX_MPDU_START_INFO4_MPDU_SEQ_CTRL_VALID);
336
}
337
338
static bool ath12k_hw_qcn9274_rx_desc_get_mpdu_fc_valid(struct hal_rx_desc *desc)
339
{
340
return !!le32_get_bits(desc->u.qcn9274.mpdu_start.info4,
341
RX_MPDU_START_INFO4_MPDU_FCTRL_VALID);
342
}
343
344
static u16 ath12k_hw_qcn9274_rx_desc_get_mpdu_start_seq_no(struct hal_rx_desc *desc)
345
{
346
return le32_get_bits(desc->u.qcn9274.mpdu_start.info4,
347
RX_MPDU_START_INFO4_MPDU_SEQ_NUM);
348
}
349
350
static u16 ath12k_hw_qcn9274_rx_desc_get_msdu_len(struct hal_rx_desc *desc)
351
{
352
return le32_get_bits(desc->u.qcn9274.msdu_end.info10,
353
RX_MSDU_END_INFO10_MSDU_LENGTH);
354
}
355
356
static u8 ath12k_hw_qcn9274_rx_desc_get_msdu_sgi(struct hal_rx_desc *desc)
357
{
358
return le32_get_bits(desc->u.qcn9274.msdu_end.info12,
359
RX_MSDU_END_INFO12_SGI);
360
}
361
362
static u8 ath12k_hw_qcn9274_rx_desc_get_msdu_rate_mcs(struct hal_rx_desc *desc)
363
{
364
return le32_get_bits(desc->u.qcn9274.msdu_end.info12,
365
RX_MSDU_END_INFO12_RATE_MCS);
366
}
367
368
static u8 ath12k_hw_qcn9274_rx_desc_get_msdu_rx_bw(struct hal_rx_desc *desc)
369
{
370
return le32_get_bits(desc->u.qcn9274.msdu_end.info12,
371
RX_MSDU_END_INFO12_RECV_BW);
372
}
373
374
static u32 ath12k_hw_qcn9274_rx_desc_get_msdu_freq(struct hal_rx_desc *desc)
375
{
376
return __le32_to_cpu(desc->u.qcn9274.msdu_end.phy_meta_data);
377
}
378
379
static u8 ath12k_hw_qcn9274_rx_desc_get_msdu_pkt_type(struct hal_rx_desc *desc)
380
{
381
return le32_get_bits(desc->u.qcn9274.msdu_end.info12,
382
RX_MSDU_END_INFO12_PKT_TYPE);
383
}
384
385
static u8 ath12k_hw_qcn9274_rx_desc_get_msdu_nss(struct hal_rx_desc *desc)
386
{
387
return le32_get_bits(desc->u.qcn9274.msdu_end.info12,
388
RX_MSDU_END_INFO12_MIMO_SS_BITMAP);
389
}
390
391
static u8 ath12k_hw_qcn9274_rx_desc_get_mpdu_tid(struct hal_rx_desc *desc)
392
{
393
return le16_get_bits(desc->u.qcn9274.msdu_end.info5,
394
RX_MSDU_END_INFO5_TID);
395
}
396
397
static u16 ath12k_hw_qcn9274_rx_desc_get_mpdu_peer_id(struct hal_rx_desc *desc)
398
{
399
return __le16_to_cpu(desc->u.qcn9274.mpdu_start.sw_peer_id);
400
}
401
402
static void ath12k_hw_qcn9274_rx_desc_copy_end_tlv(struct hal_rx_desc *fdesc,
403
struct hal_rx_desc *ldesc)
404
{
405
memcpy(&fdesc->u.qcn9274.msdu_end, &ldesc->u.qcn9274.msdu_end,
406
sizeof(struct rx_msdu_end_qcn9274));
407
}
408
409
static u32 ath12k_hw_qcn9274_rx_desc_get_mpdu_ppdu_id(struct hal_rx_desc *desc)
410
{
411
return __le16_to_cpu(desc->u.qcn9274.mpdu_start.phy_ppdu_id);
412
}
413
414
static void ath12k_hw_qcn9274_rx_desc_set_msdu_len(struct hal_rx_desc *desc, u16 len)
415
{
416
u32 info = __le32_to_cpu(desc->u.qcn9274.msdu_end.info10);
417
418
info &= ~RX_MSDU_END_INFO10_MSDU_LENGTH;
419
info |= u32_encode_bits(len, RX_MSDU_END_INFO10_MSDU_LENGTH);
420
421
desc->u.qcn9274.msdu_end.info10 = __cpu_to_le32(info);
422
}
423
424
static u8 *ath12k_hw_qcn9274_rx_desc_get_msdu_payload(struct hal_rx_desc *desc)
425
{
426
return &desc->u.qcn9274.msdu_payload[0];
427
}
428
429
static u32 ath12k_hw_qcn9274_rx_desc_get_mpdu_start_offset(void)
430
{
431
return offsetof(struct hal_rx_desc_qcn9274, mpdu_start);
432
}
433
434
static u32 ath12k_hw_qcn9274_rx_desc_get_msdu_end_offset(void)
435
{
436
return offsetof(struct hal_rx_desc_qcn9274, msdu_end);
437
}
438
439
static bool ath12k_hw_qcn9274_rx_desc_mac_addr2_valid(struct hal_rx_desc *desc)
440
{
441
return __le32_to_cpu(desc->u.qcn9274.mpdu_start.info4) &
442
RX_MPDU_START_INFO4_MAC_ADDR2_VALID;
443
}
444
445
static u8 *ath12k_hw_qcn9274_rx_desc_mpdu_start_addr2(struct hal_rx_desc *desc)
446
{
447
return desc->u.qcn9274.mpdu_start.addr2;
448
}
449
450
static bool ath12k_hw_qcn9274_rx_desc_is_da_mcbc(struct hal_rx_desc *desc)
451
{
452
return __le16_to_cpu(desc->u.qcn9274.msdu_end.info5) &
453
RX_MSDU_END_INFO5_DA_IS_MCBC;
454
}
455
456
static void ath12k_hw_qcn9274_rx_desc_get_dot11_hdr(struct hal_rx_desc *desc,
457
struct ieee80211_hdr *hdr)
458
{
459
hdr->frame_control = desc->u.qcn9274.mpdu_start.frame_ctrl;
460
hdr->duration_id = desc->u.qcn9274.mpdu_start.duration;
461
ether_addr_copy(hdr->addr1, desc->u.qcn9274.mpdu_start.addr1);
462
ether_addr_copy(hdr->addr2, desc->u.qcn9274.mpdu_start.addr2);
463
ether_addr_copy(hdr->addr3, desc->u.qcn9274.mpdu_start.addr3);
464
if (__le32_to_cpu(desc->u.qcn9274.mpdu_start.info4) &
465
RX_MPDU_START_INFO4_MAC_ADDR4_VALID) {
466
ether_addr_copy(hdr->addr4, desc->u.qcn9274.mpdu_start.addr4);
467
}
468
hdr->seq_ctrl = desc->u.qcn9274.mpdu_start.seq_ctrl;
469
}
470
471
static void ath12k_hw_qcn9274_rx_desc_get_crypto_hdr(struct hal_rx_desc *desc,
472
u8 *crypto_hdr,
473
enum hal_encrypt_type enctype)
474
{
475
unsigned int key_id;
476
477
switch (enctype) {
478
case HAL_ENCRYPT_TYPE_OPEN:
479
return;
480
case HAL_ENCRYPT_TYPE_TKIP_NO_MIC:
481
case HAL_ENCRYPT_TYPE_TKIP_MIC:
482
crypto_hdr[0] =
483
HAL_RX_MPDU_INFO_PN_GET_BYTE2(desc->u.qcn9274.mpdu_start.pn[0]);
484
crypto_hdr[1] = 0;
485
crypto_hdr[2] =
486
HAL_RX_MPDU_INFO_PN_GET_BYTE1(desc->u.qcn9274.mpdu_start.pn[0]);
487
break;
488
case HAL_ENCRYPT_TYPE_CCMP_128:
489
case HAL_ENCRYPT_TYPE_CCMP_256:
490
case HAL_ENCRYPT_TYPE_GCMP_128:
491
case HAL_ENCRYPT_TYPE_AES_GCMP_256:
492
crypto_hdr[0] =
493
HAL_RX_MPDU_INFO_PN_GET_BYTE1(desc->u.qcn9274.mpdu_start.pn[0]);
494
crypto_hdr[1] =
495
HAL_RX_MPDU_INFO_PN_GET_BYTE2(desc->u.qcn9274.mpdu_start.pn[0]);
496
crypto_hdr[2] = 0;
497
break;
498
case HAL_ENCRYPT_TYPE_WEP_40:
499
case HAL_ENCRYPT_TYPE_WEP_104:
500
case HAL_ENCRYPT_TYPE_WEP_128:
501
case HAL_ENCRYPT_TYPE_WAPI_GCM_SM4:
502
case HAL_ENCRYPT_TYPE_WAPI:
503
return;
504
}
505
key_id = le32_get_bits(desc->u.qcn9274.mpdu_start.info5,
506
RX_MPDU_START_INFO5_KEY_ID);
507
crypto_hdr[3] = 0x20 | (key_id << 6);
508
crypto_hdr[4] = HAL_RX_MPDU_INFO_PN_GET_BYTE3(desc->u.qcn9274.mpdu_start.pn[0]);
509
crypto_hdr[5] = HAL_RX_MPDU_INFO_PN_GET_BYTE4(desc->u.qcn9274.mpdu_start.pn[0]);
510
crypto_hdr[6] = HAL_RX_MPDU_INFO_PN_GET_BYTE1(desc->u.qcn9274.mpdu_start.pn[1]);
511
crypto_hdr[7] = HAL_RX_MPDU_INFO_PN_GET_BYTE2(desc->u.qcn9274.mpdu_start.pn[1]);
512
}
513
514
static u16 ath12k_hw_qcn9274_rx_desc_get_mpdu_frame_ctl(struct hal_rx_desc *desc)
515
{
516
return __le16_to_cpu(desc->u.qcn9274.mpdu_start.frame_ctrl);
517
}
518
519
static int ath12k_hal_srng_create_config_qcn9274(struct ath12k_base *ab)
520
{
521
struct ath12k_hal *hal = &ab->hal;
522
struct hal_srng_config *s;
523
524
hal->srng_config = kmemdup(hw_srng_config_template,
525
sizeof(hw_srng_config_template),
526
GFP_KERNEL);
527
if (!hal->srng_config)
528
return -ENOMEM;
529
530
s = &hal->srng_config[HAL_REO_DST];
531
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO1_RING_BASE_LSB(ab);
532
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO1_RING_HP;
533
s->reg_size[0] = HAL_REO2_RING_BASE_LSB(ab) - HAL_REO1_RING_BASE_LSB(ab);
534
s->reg_size[1] = HAL_REO2_RING_HP - HAL_REO1_RING_HP;
535
536
s = &hal->srng_config[HAL_REO_EXCEPTION];
537
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_SW0_RING_BASE_LSB(ab);
538
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_SW0_RING_HP;
539
540
s = &hal->srng_config[HAL_REO_REINJECT];
541
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_SW2REO_RING_BASE_LSB(ab);
542
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_SW2REO_RING_HP;
543
s->reg_size[0] = HAL_SW2REO1_RING_BASE_LSB(ab) - HAL_SW2REO_RING_BASE_LSB(ab);
544
s->reg_size[1] = HAL_SW2REO1_RING_HP - HAL_SW2REO_RING_HP;
545
546
s = &hal->srng_config[HAL_REO_CMD];
547
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_CMD_RING_BASE_LSB(ab);
548
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_CMD_HP;
549
550
s = &hal->srng_config[HAL_REO_STATUS];
551
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_STATUS_RING_BASE_LSB(ab);
552
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_STATUS_HP;
553
554
s = &hal->srng_config[HAL_TCL_DATA];
555
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL1_RING_BASE_LSB;
556
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL1_RING_HP;
557
s->reg_size[0] = HAL_TCL2_RING_BASE_LSB - HAL_TCL1_RING_BASE_LSB;
558
s->reg_size[1] = HAL_TCL2_RING_HP - HAL_TCL1_RING_HP;
559
560
s = &hal->srng_config[HAL_TCL_CMD];
561
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_RING_BASE_LSB(ab);
562
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_RING_HP;
563
564
s = &hal->srng_config[HAL_TCL_STATUS];
565
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_STATUS_RING_BASE_LSB(ab);
566
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_STATUS_RING_HP;
567
568
s = &hal->srng_config[HAL_CE_SRC];
569
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_CE0_SRC_REG + HAL_CE_DST_RING_BASE_LSB;
570
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_CE0_SRC_REG + HAL_CE_DST_RING_HP;
571
s->reg_size[0] = HAL_SEQ_WCSS_UMAC_CE1_SRC_REG -
572
HAL_SEQ_WCSS_UMAC_CE0_SRC_REG;
573
s->reg_size[1] = HAL_SEQ_WCSS_UMAC_CE1_SRC_REG -
574
HAL_SEQ_WCSS_UMAC_CE0_SRC_REG;
575
576
s = &hal->srng_config[HAL_CE_DST];
577
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG + HAL_CE_DST_RING_BASE_LSB;
578
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG + HAL_CE_DST_RING_HP;
579
s->reg_size[0] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG -
580
HAL_SEQ_WCSS_UMAC_CE0_DST_REG;
581
s->reg_size[1] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG -
582
HAL_SEQ_WCSS_UMAC_CE0_DST_REG;
583
584
s = &hal->srng_config[HAL_CE_DST_STATUS];
585
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG +
586
HAL_CE_DST_STATUS_RING_BASE_LSB;
587
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG + HAL_CE_DST_STATUS_RING_HP;
588
s->reg_size[0] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG -
589
HAL_SEQ_WCSS_UMAC_CE0_DST_REG;
590
s->reg_size[1] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG -
591
HAL_SEQ_WCSS_UMAC_CE0_DST_REG;
592
593
s = &hal->srng_config[HAL_WBM_IDLE_LINK];
594
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_IDLE_LINK_RING_BASE_LSB(ab);
595
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_IDLE_LINK_RING_HP;
596
597
s = &hal->srng_config[HAL_SW2WBM_RELEASE];
598
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG +
599
HAL_WBM_SW_RELEASE_RING_BASE_LSB(ab);
600
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_SW_RELEASE_RING_HP;
601
s->reg_size[0] = HAL_WBM_SW1_RELEASE_RING_BASE_LSB(ab) -
602
HAL_WBM_SW_RELEASE_RING_BASE_LSB(ab);
603
s->reg_size[1] = HAL_WBM_SW1_RELEASE_RING_HP - HAL_WBM_SW_RELEASE_RING_HP;
604
605
s = &hal->srng_config[HAL_WBM2SW_RELEASE];
606
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM0_RELEASE_RING_BASE_LSB(ab);
607
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM0_RELEASE_RING_HP;
608
s->reg_size[0] = HAL_WBM1_RELEASE_RING_BASE_LSB(ab) -
609
HAL_WBM0_RELEASE_RING_BASE_LSB(ab);
610
s->reg_size[1] = HAL_WBM1_RELEASE_RING_HP - HAL_WBM0_RELEASE_RING_HP;
611
612
/* Some LMAC rings are not accessed from the host:
613
* RXDMA_BUG, RXDMA_DST, RXDMA_MONITOR_BUF, RXDMA_MONITOR_STATUS,
614
* RXDMA_MONITOR_DST, RXDMA_MONITOR_DESC, RXDMA_DIR_BUF_SRC,
615
* RXDMA_RX_MONITOR_BUF, TX_MONITOR_BUF, TX_MONITOR_DST, SW2RXDMA
616
*/
617
s = &hal->srng_config[HAL_PPE2TCL];
618
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_PPE2TCL1_RING_BASE_LSB;
619
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_PPE2TCL1_RING_HP;
620
621
s = &hal->srng_config[HAL_PPE_RELEASE];
622
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG +
623
HAL_WBM_PPE_RELEASE_RING_BASE_LSB(ab);
624
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_PPE_RELEASE_RING_HP;
625
626
return 0;
627
}
628
629
static bool ath12k_hw_qcn9274_dp_rx_h_msdu_done(struct hal_rx_desc *desc)
630
{
631
return !!le32_get_bits(desc->u.qcn9274.msdu_end.info14,
632
RX_MSDU_END_INFO14_MSDU_DONE);
633
}
634
635
static bool ath12k_hw_qcn9274_dp_rx_h_l4_cksum_fail(struct hal_rx_desc *desc)
636
{
637
return !!le32_get_bits(desc->u.qcn9274.msdu_end.info13,
638
RX_MSDU_END_INFO13_TCP_UDP_CKSUM_FAIL);
639
}
640
641
static bool ath12k_hw_qcn9274_dp_rx_h_ip_cksum_fail(struct hal_rx_desc *desc)
642
{
643
return !!le32_get_bits(desc->u.qcn9274.msdu_end.info13,
644
RX_MSDU_END_INFO13_IP_CKSUM_FAIL);
645
}
646
647
static bool ath12k_hw_qcn9274_dp_rx_h_is_decrypted(struct hal_rx_desc *desc)
648
{
649
return (le32_get_bits(desc->u.qcn9274.msdu_end.info14,
650
RX_MSDU_END_INFO14_DECRYPT_STATUS_CODE) ==
651
RX_DESC_DECRYPT_STATUS_CODE_OK);
652
}
653
654
static u32 ath12k_hw_qcn9274_dp_rx_h_mpdu_err(struct hal_rx_desc *desc)
655
{
656
u32 info = __le32_to_cpu(desc->u.qcn9274.msdu_end.info13);
657
u32 errmap = 0;
658
659
if (info & RX_MSDU_END_INFO13_FCS_ERR)
660
errmap |= HAL_RX_MPDU_ERR_FCS;
661
662
if (info & RX_MSDU_END_INFO13_DECRYPT_ERR)
663
errmap |= HAL_RX_MPDU_ERR_DECRYPT;
664
665
if (info & RX_MSDU_END_INFO13_TKIP_MIC_ERR)
666
errmap |= HAL_RX_MPDU_ERR_TKIP_MIC;
667
668
if (info & RX_MSDU_END_INFO13_A_MSDU_ERROR)
669
errmap |= HAL_RX_MPDU_ERR_AMSDU_ERR;
670
671
if (info & RX_MSDU_END_INFO13_OVERFLOW_ERR)
672
errmap |= HAL_RX_MPDU_ERR_OVERFLOW;
673
674
if (info & RX_MSDU_END_INFO13_MSDU_LEN_ERR)
675
errmap |= HAL_RX_MPDU_ERR_MSDU_LEN;
676
677
if (info & RX_MSDU_END_INFO13_MPDU_LEN_ERR)
678
errmap |= HAL_RX_MPDU_ERR_MPDU_LEN;
679
680
return errmap;
681
}
682
683
const struct hal_ops hal_qcn9274_ops = {
684
.rx_desc_get_first_msdu = ath12k_hw_qcn9274_rx_desc_get_first_msdu,
685
.rx_desc_get_last_msdu = ath12k_hw_qcn9274_rx_desc_get_last_msdu,
686
.rx_desc_get_l3_pad_bytes = ath12k_hw_qcn9274_rx_desc_get_l3_pad_bytes,
687
.rx_desc_encrypt_valid = ath12k_hw_qcn9274_rx_desc_encrypt_valid,
688
.rx_desc_get_encrypt_type = ath12k_hw_qcn9274_rx_desc_get_encrypt_type,
689
.rx_desc_get_decap_type = ath12k_hw_qcn9274_rx_desc_get_decap_type,
690
.rx_desc_get_mesh_ctl = ath12k_hw_qcn9274_rx_desc_get_mesh_ctl,
691
.rx_desc_get_mpdu_seq_ctl_vld = ath12k_hw_qcn9274_rx_desc_get_mpdu_seq_ctl_vld,
692
.rx_desc_get_mpdu_fc_valid = ath12k_hw_qcn9274_rx_desc_get_mpdu_fc_valid,
693
.rx_desc_get_mpdu_start_seq_no = ath12k_hw_qcn9274_rx_desc_get_mpdu_start_seq_no,
694
.rx_desc_get_msdu_len = ath12k_hw_qcn9274_rx_desc_get_msdu_len,
695
.rx_desc_get_msdu_sgi = ath12k_hw_qcn9274_rx_desc_get_msdu_sgi,
696
.rx_desc_get_msdu_rate_mcs = ath12k_hw_qcn9274_rx_desc_get_msdu_rate_mcs,
697
.rx_desc_get_msdu_rx_bw = ath12k_hw_qcn9274_rx_desc_get_msdu_rx_bw,
698
.rx_desc_get_msdu_freq = ath12k_hw_qcn9274_rx_desc_get_msdu_freq,
699
.rx_desc_get_msdu_pkt_type = ath12k_hw_qcn9274_rx_desc_get_msdu_pkt_type,
700
.rx_desc_get_msdu_nss = ath12k_hw_qcn9274_rx_desc_get_msdu_nss,
701
.rx_desc_get_mpdu_tid = ath12k_hw_qcn9274_rx_desc_get_mpdu_tid,
702
.rx_desc_get_mpdu_peer_id = ath12k_hw_qcn9274_rx_desc_get_mpdu_peer_id,
703
.rx_desc_copy_end_tlv = ath12k_hw_qcn9274_rx_desc_copy_end_tlv,
704
.rx_desc_get_mpdu_ppdu_id = ath12k_hw_qcn9274_rx_desc_get_mpdu_ppdu_id,
705
.rx_desc_set_msdu_len = ath12k_hw_qcn9274_rx_desc_set_msdu_len,
706
.rx_desc_get_msdu_payload = ath12k_hw_qcn9274_rx_desc_get_msdu_payload,
707
.rx_desc_get_mpdu_start_offset = ath12k_hw_qcn9274_rx_desc_get_mpdu_start_offset,
708
.rx_desc_get_msdu_end_offset = ath12k_hw_qcn9274_rx_desc_get_msdu_end_offset,
709
.rx_desc_mac_addr2_valid = ath12k_hw_qcn9274_rx_desc_mac_addr2_valid,
710
.rx_desc_mpdu_start_addr2 = ath12k_hw_qcn9274_rx_desc_mpdu_start_addr2,
711
.rx_desc_is_da_mcbc = ath12k_hw_qcn9274_rx_desc_is_da_mcbc,
712
.rx_desc_get_dot11_hdr = ath12k_hw_qcn9274_rx_desc_get_dot11_hdr,
713
.rx_desc_get_crypto_header = ath12k_hw_qcn9274_rx_desc_get_crypto_hdr,
714
.rx_desc_get_mpdu_frame_ctl = ath12k_hw_qcn9274_rx_desc_get_mpdu_frame_ctl,
715
.create_srng_config = ath12k_hal_srng_create_config_qcn9274,
716
.tcl_to_wbm_rbm_map = ath12k_hal_qcn9274_tcl_to_wbm_rbm_map,
717
.dp_rx_h_msdu_done = ath12k_hw_qcn9274_dp_rx_h_msdu_done,
718
.dp_rx_h_l4_cksum_fail = ath12k_hw_qcn9274_dp_rx_h_l4_cksum_fail,
719
.dp_rx_h_ip_cksum_fail = ath12k_hw_qcn9274_dp_rx_h_ip_cksum_fail,
720
.dp_rx_h_is_decrypted = ath12k_hw_qcn9274_dp_rx_h_is_decrypted,
721
.dp_rx_h_mpdu_err = ath12k_hw_qcn9274_dp_rx_h_mpdu_err,
722
};
723
724
static bool ath12k_hw_wcn7850_rx_desc_get_first_msdu(struct hal_rx_desc *desc)
725
{
726
return !!le16_get_bits(desc->u.wcn7850.msdu_end.info5,
727
RX_MSDU_END_INFO5_FIRST_MSDU);
728
}
729
730
static bool ath12k_hw_wcn7850_rx_desc_get_last_msdu(struct hal_rx_desc *desc)
731
{
732
return !!le16_get_bits(desc->u.wcn7850.msdu_end.info5,
733
RX_MSDU_END_INFO5_LAST_MSDU);
734
}
735
736
static u8 ath12k_hw_wcn7850_rx_desc_get_l3_pad_bytes(struct hal_rx_desc *desc)
737
{
738
return le16_get_bits(desc->u.wcn7850.msdu_end.info5,
739
RX_MSDU_END_INFO5_L3_HDR_PADDING);
740
}
741
742
static bool ath12k_hw_wcn7850_rx_desc_encrypt_valid(struct hal_rx_desc *desc)
743
{
744
return !!le32_get_bits(desc->u.wcn7850.mpdu_start.info4,
745
RX_MPDU_START_INFO4_ENCRYPT_INFO_VALID);
746
}
747
748
static u32 ath12k_hw_wcn7850_rx_desc_get_encrypt_type(struct hal_rx_desc *desc)
749
{
750
return le32_get_bits(desc->u.wcn7850.mpdu_start.info2,
751
RX_MPDU_START_INFO2_ENC_TYPE);
752
}
753
754
static u8 ath12k_hw_wcn7850_rx_desc_get_decap_type(struct hal_rx_desc *desc)
755
{
756
return le32_get_bits(desc->u.wcn7850.msdu_end.info11,
757
RX_MSDU_END_INFO11_DECAP_FORMAT);
758
}
759
760
static u8 ath12k_hw_wcn7850_rx_desc_get_mesh_ctl(struct hal_rx_desc *desc)
761
{
762
return le32_get_bits(desc->u.wcn7850.msdu_end.info11,
763
RX_MSDU_END_INFO11_MESH_CTRL_PRESENT);
764
}
765
766
static bool ath12k_hw_wcn7850_rx_desc_get_mpdu_seq_ctl_vld(struct hal_rx_desc *desc)
767
{
768
return !!le32_get_bits(desc->u.wcn7850.mpdu_start.info4,
769
RX_MPDU_START_INFO4_MPDU_SEQ_CTRL_VALID);
770
}
771
772
static bool ath12k_hw_wcn7850_rx_desc_get_mpdu_fc_valid(struct hal_rx_desc *desc)
773
{
774
return !!le32_get_bits(desc->u.wcn7850.mpdu_start.info4,
775
RX_MPDU_START_INFO4_MPDU_FCTRL_VALID);
776
}
777
778
static u16 ath12k_hw_wcn7850_rx_desc_get_mpdu_start_seq_no(struct hal_rx_desc *desc)
779
{
780
return le32_get_bits(desc->u.wcn7850.mpdu_start.info4,
781
RX_MPDU_START_INFO4_MPDU_SEQ_NUM);
782
}
783
784
static u16 ath12k_hw_wcn7850_rx_desc_get_msdu_len(struct hal_rx_desc *desc)
785
{
786
return le32_get_bits(desc->u.wcn7850.msdu_end.info10,
787
RX_MSDU_END_INFO10_MSDU_LENGTH);
788
}
789
790
static u8 ath12k_hw_wcn7850_rx_desc_get_msdu_sgi(struct hal_rx_desc *desc)
791
{
792
return le32_get_bits(desc->u.wcn7850.msdu_end.info12,
793
RX_MSDU_END_INFO12_SGI);
794
}
795
796
static u8 ath12k_hw_wcn7850_rx_desc_get_msdu_rate_mcs(struct hal_rx_desc *desc)
797
{
798
return le32_get_bits(desc->u.wcn7850.msdu_end.info12,
799
RX_MSDU_END_INFO12_RATE_MCS);
800
}
801
802
static u8 ath12k_hw_wcn7850_rx_desc_get_msdu_rx_bw(struct hal_rx_desc *desc)
803
{
804
return le32_get_bits(desc->u.wcn7850.msdu_end.info12,
805
RX_MSDU_END_INFO12_RECV_BW);
806
}
807
808
static u32 ath12k_hw_wcn7850_rx_desc_get_msdu_freq(struct hal_rx_desc *desc)
809
{
810
return __le32_to_cpu(desc->u.wcn7850.msdu_end.phy_meta_data);
811
}
812
813
static u8 ath12k_hw_wcn7850_rx_desc_get_msdu_pkt_type(struct hal_rx_desc *desc)
814
{
815
return le32_get_bits(desc->u.wcn7850.msdu_end.info12,
816
RX_MSDU_END_INFO12_PKT_TYPE);
817
}
818
819
static u8 ath12k_hw_wcn7850_rx_desc_get_msdu_nss(struct hal_rx_desc *desc)
820
{
821
return le32_get_bits(desc->u.wcn7850.msdu_end.info12,
822
RX_MSDU_END_INFO12_MIMO_SS_BITMAP);
823
}
824
825
static u8 ath12k_hw_wcn7850_rx_desc_get_mpdu_tid(struct hal_rx_desc *desc)
826
{
827
return le16_get_bits(desc->u.wcn7850.msdu_end.info5,
828
RX_MSDU_END_INFO5_TID);
829
}
830
831
static u16 ath12k_hw_wcn7850_rx_desc_get_mpdu_peer_id(struct hal_rx_desc *desc)
832
{
833
return __le16_to_cpu(desc->u.wcn7850.mpdu_start.sw_peer_id);
834
}
835
836
static void ath12k_hw_wcn7850_rx_desc_copy_end_tlv(struct hal_rx_desc *fdesc,
837
struct hal_rx_desc *ldesc)
838
{
839
memcpy(&fdesc->u.wcn7850.msdu_end, &ldesc->u.wcn7850.msdu_end,
840
sizeof(struct rx_msdu_end_qcn9274));
841
}
842
843
static u32 ath12k_hw_wcn7850_rx_desc_get_mpdu_start_tag(struct hal_rx_desc *desc)
844
{
845
return le64_get_bits(desc->u.wcn7850.mpdu_start_tag,
846
HAL_TLV_HDR_TAG);
847
}
848
849
static u32 ath12k_hw_wcn7850_rx_desc_get_mpdu_ppdu_id(struct hal_rx_desc *desc)
850
{
851
return __le16_to_cpu(desc->u.wcn7850.mpdu_start.phy_ppdu_id);
852
}
853
854
static void ath12k_hw_wcn7850_rx_desc_set_msdu_len(struct hal_rx_desc *desc, u16 len)
855
{
856
u32 info = __le32_to_cpu(desc->u.wcn7850.msdu_end.info10);
857
858
info &= ~RX_MSDU_END_INFO10_MSDU_LENGTH;
859
info |= u32_encode_bits(len, RX_MSDU_END_INFO10_MSDU_LENGTH);
860
861
desc->u.wcn7850.msdu_end.info10 = __cpu_to_le32(info);
862
}
863
864
static u8 *ath12k_hw_wcn7850_rx_desc_get_msdu_payload(struct hal_rx_desc *desc)
865
{
866
return &desc->u.wcn7850.msdu_payload[0];
867
}
868
869
static u32 ath12k_hw_wcn7850_rx_desc_get_mpdu_start_offset(void)
870
{
871
return offsetof(struct hal_rx_desc_wcn7850, mpdu_start_tag);
872
}
873
874
static u32 ath12k_hw_wcn7850_rx_desc_get_msdu_end_offset(void)
875
{
876
return offsetof(struct hal_rx_desc_wcn7850, msdu_end_tag);
877
}
878
879
static bool ath12k_hw_wcn7850_rx_desc_mac_addr2_valid(struct hal_rx_desc *desc)
880
{
881
return __le32_to_cpu(desc->u.wcn7850.mpdu_start.info4) &
882
RX_MPDU_START_INFO4_MAC_ADDR2_VALID;
883
}
884
885
static u8 *ath12k_hw_wcn7850_rx_desc_mpdu_start_addr2(struct hal_rx_desc *desc)
886
{
887
return desc->u.wcn7850.mpdu_start.addr2;
888
}
889
890
static bool ath12k_hw_wcn7850_rx_desc_is_da_mcbc(struct hal_rx_desc *desc)
891
{
892
return __le16_to_cpu(desc->u.wcn7850.msdu_end.info5) &
893
RX_MSDU_END_INFO5_DA_IS_MCBC;
894
}
895
896
static void ath12k_hw_wcn7850_rx_desc_get_dot11_hdr(struct hal_rx_desc *desc,
897
struct ieee80211_hdr *hdr)
898
{
899
hdr->frame_control = desc->u.wcn7850.mpdu_start.frame_ctrl;
900
hdr->duration_id = desc->u.wcn7850.mpdu_start.duration;
901
ether_addr_copy(hdr->addr1, desc->u.wcn7850.mpdu_start.addr1);
902
ether_addr_copy(hdr->addr2, desc->u.wcn7850.mpdu_start.addr2);
903
ether_addr_copy(hdr->addr3, desc->u.wcn7850.mpdu_start.addr3);
904
if (__le32_to_cpu(desc->u.wcn7850.mpdu_start.info4) &
905
RX_MPDU_START_INFO4_MAC_ADDR4_VALID) {
906
ether_addr_copy(hdr->addr4, desc->u.wcn7850.mpdu_start.addr4);
907
}
908
hdr->seq_ctrl = desc->u.wcn7850.mpdu_start.seq_ctrl;
909
}
910
911
static void ath12k_hw_wcn7850_rx_desc_get_crypto_hdr(struct hal_rx_desc *desc,
912
u8 *crypto_hdr,
913
enum hal_encrypt_type enctype)
914
{
915
unsigned int key_id;
916
917
switch (enctype) {
918
case HAL_ENCRYPT_TYPE_OPEN:
919
return;
920
case HAL_ENCRYPT_TYPE_TKIP_NO_MIC:
921
case HAL_ENCRYPT_TYPE_TKIP_MIC:
922
crypto_hdr[0] =
923
HAL_RX_MPDU_INFO_PN_GET_BYTE2(desc->u.wcn7850.mpdu_start.pn[0]);
924
crypto_hdr[1] = 0;
925
crypto_hdr[2] =
926
HAL_RX_MPDU_INFO_PN_GET_BYTE1(desc->u.wcn7850.mpdu_start.pn[0]);
927
break;
928
case HAL_ENCRYPT_TYPE_CCMP_128:
929
case HAL_ENCRYPT_TYPE_CCMP_256:
930
case HAL_ENCRYPT_TYPE_GCMP_128:
931
case HAL_ENCRYPT_TYPE_AES_GCMP_256:
932
crypto_hdr[0] =
933
HAL_RX_MPDU_INFO_PN_GET_BYTE1(desc->u.wcn7850.mpdu_start.pn[0]);
934
crypto_hdr[1] =
935
HAL_RX_MPDU_INFO_PN_GET_BYTE2(desc->u.wcn7850.mpdu_start.pn[0]);
936
crypto_hdr[2] = 0;
937
break;
938
case HAL_ENCRYPT_TYPE_WEP_40:
939
case HAL_ENCRYPT_TYPE_WEP_104:
940
case HAL_ENCRYPT_TYPE_WEP_128:
941
case HAL_ENCRYPT_TYPE_WAPI_GCM_SM4:
942
case HAL_ENCRYPT_TYPE_WAPI:
943
return;
944
}
945
key_id = u32_get_bits(__le32_to_cpu(desc->u.wcn7850.mpdu_start.info5),
946
RX_MPDU_START_INFO5_KEY_ID);
947
crypto_hdr[3] = 0x20 | (key_id << 6);
948
crypto_hdr[4] = HAL_RX_MPDU_INFO_PN_GET_BYTE3(desc->u.wcn7850.mpdu_start.pn[0]);
949
crypto_hdr[5] = HAL_RX_MPDU_INFO_PN_GET_BYTE4(desc->u.wcn7850.mpdu_start.pn[0]);
950
crypto_hdr[6] = HAL_RX_MPDU_INFO_PN_GET_BYTE1(desc->u.wcn7850.mpdu_start.pn[1]);
951
crypto_hdr[7] = HAL_RX_MPDU_INFO_PN_GET_BYTE2(desc->u.wcn7850.mpdu_start.pn[1]);
952
}
953
954
static u16 ath12k_hw_wcn7850_rx_desc_get_mpdu_frame_ctl(struct hal_rx_desc *desc)
955
{
956
return __le16_to_cpu(desc->u.wcn7850.mpdu_start.frame_ctrl);
957
}
958
959
static int ath12k_hal_srng_create_config_wcn7850(struct ath12k_base *ab)
960
{
961
struct ath12k_hal *hal = &ab->hal;
962
struct hal_srng_config *s;
963
964
hal->srng_config = kmemdup(hw_srng_config_template,
965
sizeof(hw_srng_config_template),
966
GFP_KERNEL);
967
if (!hal->srng_config)
968
return -ENOMEM;
969
970
s = &hal->srng_config[HAL_REO_DST];
971
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO1_RING_BASE_LSB(ab);
972
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO1_RING_HP;
973
s->reg_size[0] = HAL_REO2_RING_BASE_LSB(ab) - HAL_REO1_RING_BASE_LSB(ab);
974
s->reg_size[1] = HAL_REO2_RING_HP - HAL_REO1_RING_HP;
975
976
s = &hal->srng_config[HAL_REO_EXCEPTION];
977
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_SW0_RING_BASE_LSB(ab);
978
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_SW0_RING_HP;
979
980
s = &hal->srng_config[HAL_REO_REINJECT];
981
s->max_rings = 1;
982
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_SW2REO_RING_BASE_LSB(ab);
983
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_SW2REO_RING_HP;
984
985
s = &hal->srng_config[HAL_REO_CMD];
986
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_CMD_RING_BASE_LSB(ab);
987
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_CMD_HP;
988
989
s = &hal->srng_config[HAL_REO_STATUS];
990
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_STATUS_RING_BASE_LSB(ab);
991
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_REO_REG + HAL_REO_STATUS_HP;
992
993
s = &hal->srng_config[HAL_TCL_DATA];
994
s->max_rings = 5;
995
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL1_RING_BASE_LSB;
996
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL1_RING_HP;
997
s->reg_size[0] = HAL_TCL2_RING_BASE_LSB - HAL_TCL1_RING_BASE_LSB;
998
s->reg_size[1] = HAL_TCL2_RING_HP - HAL_TCL1_RING_HP;
999
1000
s = &hal->srng_config[HAL_TCL_CMD];
1001
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_RING_BASE_LSB(ab);
1002
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_RING_HP;
1003
1004
s = &hal->srng_config[HAL_TCL_STATUS];
1005
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_STATUS_RING_BASE_LSB(ab);
1006
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_TCL_REG + HAL_TCL_STATUS_RING_HP;
1007
1008
s = &hal->srng_config[HAL_CE_SRC];
1009
s->max_rings = 12;
1010
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_CE0_SRC_REG + HAL_CE_DST_RING_BASE_LSB;
1011
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_CE0_SRC_REG + HAL_CE_DST_RING_HP;
1012
s->reg_size[0] = HAL_SEQ_WCSS_UMAC_CE1_SRC_REG -
1013
HAL_SEQ_WCSS_UMAC_CE0_SRC_REG;
1014
s->reg_size[1] = HAL_SEQ_WCSS_UMAC_CE1_SRC_REG -
1015
HAL_SEQ_WCSS_UMAC_CE0_SRC_REG;
1016
1017
s = &hal->srng_config[HAL_CE_DST];
1018
s->max_rings = 12;
1019
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG + HAL_CE_DST_RING_BASE_LSB;
1020
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG + HAL_CE_DST_RING_HP;
1021
s->reg_size[0] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG -
1022
HAL_SEQ_WCSS_UMAC_CE0_DST_REG;
1023
s->reg_size[1] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG -
1024
HAL_SEQ_WCSS_UMAC_CE0_DST_REG;
1025
1026
s = &hal->srng_config[HAL_CE_DST_STATUS];
1027
s->max_rings = 12;
1028
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG +
1029
HAL_CE_DST_STATUS_RING_BASE_LSB;
1030
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_CE0_DST_REG + HAL_CE_DST_STATUS_RING_HP;
1031
s->reg_size[0] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG -
1032
HAL_SEQ_WCSS_UMAC_CE0_DST_REG;
1033
s->reg_size[1] = HAL_SEQ_WCSS_UMAC_CE1_DST_REG -
1034
HAL_SEQ_WCSS_UMAC_CE0_DST_REG;
1035
1036
s = &hal->srng_config[HAL_WBM_IDLE_LINK];
1037
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_IDLE_LINK_RING_BASE_LSB(ab);
1038
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_IDLE_LINK_RING_HP;
1039
1040
s = &hal->srng_config[HAL_SW2WBM_RELEASE];
1041
s->max_rings = 1;
1042
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG +
1043
HAL_WBM_SW_RELEASE_RING_BASE_LSB(ab);
1044
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_SW_RELEASE_RING_HP;
1045
1046
s = &hal->srng_config[HAL_WBM2SW_RELEASE];
1047
s->reg_start[0] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM0_RELEASE_RING_BASE_LSB(ab);
1048
s->reg_start[1] = HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM0_RELEASE_RING_HP;
1049
s->reg_size[0] = HAL_WBM1_RELEASE_RING_BASE_LSB(ab) -
1050
HAL_WBM0_RELEASE_RING_BASE_LSB(ab);
1051
s->reg_size[1] = HAL_WBM1_RELEASE_RING_HP - HAL_WBM0_RELEASE_RING_HP;
1052
1053
s = &hal->srng_config[HAL_RXDMA_BUF];
1054
s->max_rings = 2;
1055
s->mac_type = ATH12K_HAL_SRNG_PMAC;
1056
1057
s = &hal->srng_config[HAL_RXDMA_DST];
1058
s->max_rings = 1;
1059
s->entry_size = sizeof(struct hal_reo_entrance_ring) >> 2;
1060
1061
/* below rings are not used */
1062
s = &hal->srng_config[HAL_RXDMA_DIR_BUF];
1063
s->max_rings = 0;
1064
1065
s = &hal->srng_config[HAL_PPE2TCL];
1066
s->max_rings = 0;
1067
1068
s = &hal->srng_config[HAL_PPE_RELEASE];
1069
s->max_rings = 0;
1070
1071
s = &hal->srng_config[HAL_TX_MONITOR_BUF];
1072
s->max_rings = 0;
1073
1074
s = &hal->srng_config[HAL_TX_MONITOR_DST];
1075
s->max_rings = 0;
1076
1077
s = &hal->srng_config[HAL_PPE2TCL];
1078
s->max_rings = 0;
1079
1080
return 0;
1081
}
1082
1083
static bool ath12k_hw_wcn7850_dp_rx_h_msdu_done(struct hal_rx_desc *desc)
1084
{
1085
return !!le32_get_bits(desc->u.wcn7850.msdu_end.info14,
1086
RX_MSDU_END_INFO14_MSDU_DONE);
1087
}
1088
1089
static bool ath12k_hw_wcn7850_dp_rx_h_l4_cksum_fail(struct hal_rx_desc *desc)
1090
{
1091
return !!le32_get_bits(desc->u.wcn7850.msdu_end.info13,
1092
RX_MSDU_END_INFO13_TCP_UDP_CKSUM_FAIL);
1093
}
1094
1095
static bool ath12k_hw_wcn7850_dp_rx_h_ip_cksum_fail(struct hal_rx_desc *desc)
1096
{
1097
return !!le32_get_bits(desc->u.wcn7850.msdu_end.info13,
1098
RX_MSDU_END_INFO13_IP_CKSUM_FAIL);
1099
}
1100
1101
static bool ath12k_hw_wcn7850_dp_rx_h_is_decrypted(struct hal_rx_desc *desc)
1102
{
1103
return (le32_get_bits(desc->u.wcn7850.msdu_end.info14,
1104
RX_MSDU_END_INFO14_DECRYPT_STATUS_CODE) ==
1105
RX_DESC_DECRYPT_STATUS_CODE_OK);
1106
}
1107
1108
static u32 ath12k_hw_wcn7850_dp_rx_h_mpdu_err(struct hal_rx_desc *desc)
1109
{
1110
u32 info = __le32_to_cpu(desc->u.wcn7850.msdu_end.info13);
1111
u32 errmap = 0;
1112
1113
if (info & RX_MSDU_END_INFO13_FCS_ERR)
1114
errmap |= HAL_RX_MPDU_ERR_FCS;
1115
1116
if (info & RX_MSDU_END_INFO13_DECRYPT_ERR)
1117
errmap |= HAL_RX_MPDU_ERR_DECRYPT;
1118
1119
if (info & RX_MSDU_END_INFO13_TKIP_MIC_ERR)
1120
errmap |= HAL_RX_MPDU_ERR_TKIP_MIC;
1121
1122
if (info & RX_MSDU_END_INFO13_A_MSDU_ERROR)
1123
errmap |= HAL_RX_MPDU_ERR_AMSDU_ERR;
1124
1125
if (info & RX_MSDU_END_INFO13_OVERFLOW_ERR)
1126
errmap |= HAL_RX_MPDU_ERR_OVERFLOW;
1127
1128
if (info & RX_MSDU_END_INFO13_MSDU_LEN_ERR)
1129
errmap |= HAL_RX_MPDU_ERR_MSDU_LEN;
1130
1131
if (info & RX_MSDU_END_INFO13_MPDU_LEN_ERR)
1132
errmap |= HAL_RX_MPDU_ERR_MPDU_LEN;
1133
1134
return errmap;
1135
}
1136
1137
const struct hal_ops hal_wcn7850_ops = {
1138
.rx_desc_get_first_msdu = ath12k_hw_wcn7850_rx_desc_get_first_msdu,
1139
.rx_desc_get_last_msdu = ath12k_hw_wcn7850_rx_desc_get_last_msdu,
1140
.rx_desc_get_l3_pad_bytes = ath12k_hw_wcn7850_rx_desc_get_l3_pad_bytes,
1141
.rx_desc_encrypt_valid = ath12k_hw_wcn7850_rx_desc_encrypt_valid,
1142
.rx_desc_get_encrypt_type = ath12k_hw_wcn7850_rx_desc_get_encrypt_type,
1143
.rx_desc_get_decap_type = ath12k_hw_wcn7850_rx_desc_get_decap_type,
1144
.rx_desc_get_mesh_ctl = ath12k_hw_wcn7850_rx_desc_get_mesh_ctl,
1145
.rx_desc_get_mpdu_seq_ctl_vld = ath12k_hw_wcn7850_rx_desc_get_mpdu_seq_ctl_vld,
1146
.rx_desc_get_mpdu_fc_valid = ath12k_hw_wcn7850_rx_desc_get_mpdu_fc_valid,
1147
.rx_desc_get_mpdu_start_seq_no = ath12k_hw_wcn7850_rx_desc_get_mpdu_start_seq_no,
1148
.rx_desc_get_msdu_len = ath12k_hw_wcn7850_rx_desc_get_msdu_len,
1149
.rx_desc_get_msdu_sgi = ath12k_hw_wcn7850_rx_desc_get_msdu_sgi,
1150
.rx_desc_get_msdu_rate_mcs = ath12k_hw_wcn7850_rx_desc_get_msdu_rate_mcs,
1151
.rx_desc_get_msdu_rx_bw = ath12k_hw_wcn7850_rx_desc_get_msdu_rx_bw,
1152
.rx_desc_get_msdu_freq = ath12k_hw_wcn7850_rx_desc_get_msdu_freq,
1153
.rx_desc_get_msdu_pkt_type = ath12k_hw_wcn7850_rx_desc_get_msdu_pkt_type,
1154
.rx_desc_get_msdu_nss = ath12k_hw_wcn7850_rx_desc_get_msdu_nss,
1155
.rx_desc_get_mpdu_tid = ath12k_hw_wcn7850_rx_desc_get_mpdu_tid,
1156
.rx_desc_get_mpdu_peer_id = ath12k_hw_wcn7850_rx_desc_get_mpdu_peer_id,
1157
.rx_desc_copy_end_tlv = ath12k_hw_wcn7850_rx_desc_copy_end_tlv,
1158
.rx_desc_get_mpdu_start_tag = ath12k_hw_wcn7850_rx_desc_get_mpdu_start_tag,
1159
.rx_desc_get_mpdu_ppdu_id = ath12k_hw_wcn7850_rx_desc_get_mpdu_ppdu_id,
1160
.rx_desc_set_msdu_len = ath12k_hw_wcn7850_rx_desc_set_msdu_len,
1161
.rx_desc_get_msdu_payload = ath12k_hw_wcn7850_rx_desc_get_msdu_payload,
1162
.rx_desc_get_mpdu_start_offset = ath12k_hw_wcn7850_rx_desc_get_mpdu_start_offset,
1163
.rx_desc_get_msdu_end_offset = ath12k_hw_wcn7850_rx_desc_get_msdu_end_offset,
1164
.rx_desc_mac_addr2_valid = ath12k_hw_wcn7850_rx_desc_mac_addr2_valid,
1165
.rx_desc_mpdu_start_addr2 = ath12k_hw_wcn7850_rx_desc_mpdu_start_addr2,
1166
.rx_desc_is_da_mcbc = ath12k_hw_wcn7850_rx_desc_is_da_mcbc,
1167
.rx_desc_get_dot11_hdr = ath12k_hw_wcn7850_rx_desc_get_dot11_hdr,
1168
.rx_desc_get_crypto_header = ath12k_hw_wcn7850_rx_desc_get_crypto_hdr,
1169
.rx_desc_get_mpdu_frame_ctl = ath12k_hw_wcn7850_rx_desc_get_mpdu_frame_ctl,
1170
.create_srng_config = ath12k_hal_srng_create_config_wcn7850,
1171
.tcl_to_wbm_rbm_map = ath12k_hal_wcn7850_tcl_to_wbm_rbm_map,
1172
.dp_rx_h_msdu_done = ath12k_hw_wcn7850_dp_rx_h_msdu_done,
1173
.dp_rx_h_l4_cksum_fail = ath12k_hw_wcn7850_dp_rx_h_l4_cksum_fail,
1174
.dp_rx_h_ip_cksum_fail = ath12k_hw_wcn7850_dp_rx_h_ip_cksum_fail,
1175
.dp_rx_h_is_decrypted = ath12k_hw_wcn7850_dp_rx_h_is_decrypted,
1176
.dp_rx_h_mpdu_err = ath12k_hw_wcn7850_dp_rx_h_mpdu_err,
1177
};
1178
1179
static int ath12k_hal_alloc_cont_rdp(struct ath12k_base *ab)
1180
{
1181
struct ath12k_hal *hal = &ab->hal;
1182
size_t size;
1183
1184
size = sizeof(u32) * HAL_SRNG_RING_ID_MAX;
1185
hal->rdp.vaddr = dma_alloc_coherent(ab->dev, size, &hal->rdp.paddr,
1186
GFP_KERNEL);
1187
if (!hal->rdp.vaddr)
1188
return -ENOMEM;
1189
1190
return 0;
1191
}
1192
1193
static void ath12k_hal_free_cont_rdp(struct ath12k_base *ab)
1194
{
1195
struct ath12k_hal *hal = &ab->hal;
1196
size_t size;
1197
1198
if (!hal->rdp.vaddr)
1199
return;
1200
1201
size = sizeof(u32) * HAL_SRNG_RING_ID_MAX;
1202
dma_free_coherent(ab->dev, size,
1203
hal->rdp.vaddr, hal->rdp.paddr);
1204
hal->rdp.vaddr = NULL;
1205
}
1206
1207
static int ath12k_hal_alloc_cont_wrp(struct ath12k_base *ab)
1208
{
1209
struct ath12k_hal *hal = &ab->hal;
1210
size_t size;
1211
1212
size = sizeof(u32) * (HAL_SRNG_NUM_PMAC_RINGS + HAL_SRNG_NUM_DMAC_RINGS);
1213
hal->wrp.vaddr = dma_alloc_coherent(ab->dev, size, &hal->wrp.paddr,
1214
GFP_KERNEL);
1215
if (!hal->wrp.vaddr)
1216
return -ENOMEM;
1217
1218
return 0;
1219
}
1220
1221
static void ath12k_hal_free_cont_wrp(struct ath12k_base *ab)
1222
{
1223
struct ath12k_hal *hal = &ab->hal;
1224
size_t size;
1225
1226
if (!hal->wrp.vaddr)
1227
return;
1228
1229
size = sizeof(u32) * (HAL_SRNG_NUM_PMAC_RINGS + HAL_SRNG_NUM_DMAC_RINGS);
1230
dma_free_coherent(ab->dev, size,
1231
hal->wrp.vaddr, hal->wrp.paddr);
1232
hal->wrp.vaddr = NULL;
1233
}
1234
1235
static void ath12k_hal_ce_dst_setup(struct ath12k_base *ab,
1236
struct hal_srng *srng, int ring_num)
1237
{
1238
struct hal_srng_config *srng_config = &ab->hal.srng_config[HAL_CE_DST];
1239
u32 addr;
1240
u32 val;
1241
1242
addr = HAL_CE_DST_RING_CTRL +
1243
srng_config->reg_start[HAL_SRNG_REG_GRP_R0] +
1244
ring_num * srng_config->reg_size[HAL_SRNG_REG_GRP_R0];
1245
1246
val = ath12k_hif_read32(ab, addr);
1247
val &= ~HAL_CE_DST_R0_DEST_CTRL_MAX_LEN;
1248
val |= u32_encode_bits(srng->u.dst_ring.max_buffer_length,
1249
HAL_CE_DST_R0_DEST_CTRL_MAX_LEN);
1250
ath12k_hif_write32(ab, addr, val);
1251
}
1252
1253
static void ath12k_hal_srng_dst_hw_init(struct ath12k_base *ab,
1254
struct hal_srng *srng)
1255
{
1256
struct ath12k_hal *hal = &ab->hal;
1257
u32 val;
1258
u64 hp_addr;
1259
u32 reg_base;
1260
1261
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R0];
1262
1263
if (srng->flags & HAL_SRNG_FLAGS_MSI_INTR) {
1264
ath12k_hif_write32(ab, reg_base +
1265
ath12k_hal_reo1_ring_msi1_base_lsb_offset(ab),
1266
srng->msi_addr);
1267
1268
val = u32_encode_bits(((u64)srng->msi_addr >> HAL_ADDR_MSB_REG_SHIFT),
1269
HAL_REO1_RING_MSI1_BASE_MSB_ADDR) |
1270
HAL_REO1_RING_MSI1_BASE_MSB_MSI1_ENABLE;
1271
ath12k_hif_write32(ab, reg_base +
1272
ath12k_hal_reo1_ring_msi1_base_msb_offset(ab), val);
1273
1274
ath12k_hif_write32(ab,
1275
reg_base + ath12k_hal_reo1_ring_msi1_data_offset(ab),
1276
srng->msi_data);
1277
}
1278
1279
ath12k_hif_write32(ab, reg_base, srng->ring_base_paddr);
1280
1281
val = u32_encode_bits(((u64)srng->ring_base_paddr >> HAL_ADDR_MSB_REG_SHIFT),
1282
HAL_REO1_RING_BASE_MSB_RING_BASE_ADDR_MSB) |
1283
u32_encode_bits((srng->entry_size * srng->num_entries),
1284
HAL_REO1_RING_BASE_MSB_RING_SIZE);
1285
ath12k_hif_write32(ab, reg_base + ath12k_hal_reo1_ring_base_msb_offset(ab), val);
1286
1287
val = u32_encode_bits(srng->ring_id, HAL_REO1_RING_ID_RING_ID) |
1288
u32_encode_bits(srng->entry_size, HAL_REO1_RING_ID_ENTRY_SIZE);
1289
ath12k_hif_write32(ab, reg_base + ath12k_hal_reo1_ring_id_offset(ab), val);
1290
1291
/* interrupt setup */
1292
val = u32_encode_bits((srng->intr_timer_thres_us >> 3),
1293
HAL_REO1_RING_PRDR_INT_SETUP_INTR_TMR_THOLD);
1294
1295
val |= u32_encode_bits((srng->intr_batch_cntr_thres_entries * srng->entry_size),
1296
HAL_REO1_RING_PRDR_INT_SETUP_BATCH_COUNTER_THOLD);
1297
1298
ath12k_hif_write32(ab,
1299
reg_base + ath12k_hal_reo1_ring_producer_int_setup_offset(ab),
1300
val);
1301
1302
hp_addr = hal->rdp.paddr +
1303
((unsigned long)srng->u.dst_ring.hp_addr -
1304
(unsigned long)hal->rdp.vaddr);
1305
ath12k_hif_write32(ab, reg_base + ath12k_hal_reo1_ring_hp_addr_lsb_offset(ab),
1306
hp_addr & HAL_ADDR_LSB_REG_MASK);
1307
ath12k_hif_write32(ab, reg_base + ath12k_hal_reo1_ring_hp_addr_msb_offset(ab),
1308
hp_addr >> HAL_ADDR_MSB_REG_SHIFT);
1309
1310
/* Initialize head and tail pointers to indicate ring is empty */
1311
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R2];
1312
ath12k_hif_write32(ab, reg_base, 0);
1313
ath12k_hif_write32(ab, reg_base + HAL_REO1_RING_TP_OFFSET, 0);
1314
*srng->u.dst_ring.hp_addr = 0;
1315
1316
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R0];
1317
val = 0;
1318
if (srng->flags & HAL_SRNG_FLAGS_DATA_TLV_SWAP)
1319
val |= HAL_REO1_RING_MISC_DATA_TLV_SWAP;
1320
if (srng->flags & HAL_SRNG_FLAGS_RING_PTR_SWAP)
1321
val |= HAL_REO1_RING_MISC_HOST_FW_SWAP;
1322
if (srng->flags & HAL_SRNG_FLAGS_MSI_SWAP)
1323
val |= HAL_REO1_RING_MISC_MSI_SWAP;
1324
val |= HAL_REO1_RING_MISC_SRNG_ENABLE;
1325
1326
ath12k_hif_write32(ab, reg_base + ath12k_hal_reo1_ring_misc_offset(ab), val);
1327
}
1328
1329
static void ath12k_hal_srng_src_hw_init(struct ath12k_base *ab,
1330
struct hal_srng *srng)
1331
{
1332
struct ath12k_hal *hal = &ab->hal;
1333
u32 val;
1334
u64 tp_addr;
1335
u32 reg_base;
1336
1337
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R0];
1338
1339
if (srng->flags & HAL_SRNG_FLAGS_MSI_INTR) {
1340
ath12k_hif_write32(ab, reg_base +
1341
HAL_TCL1_RING_MSI1_BASE_LSB_OFFSET(ab),
1342
srng->msi_addr);
1343
1344
val = u32_encode_bits(((u64)srng->msi_addr >> HAL_ADDR_MSB_REG_SHIFT),
1345
HAL_TCL1_RING_MSI1_BASE_MSB_ADDR) |
1346
HAL_TCL1_RING_MSI1_BASE_MSB_MSI1_ENABLE;
1347
ath12k_hif_write32(ab, reg_base +
1348
HAL_TCL1_RING_MSI1_BASE_MSB_OFFSET(ab),
1349
val);
1350
1351
ath12k_hif_write32(ab, reg_base +
1352
HAL_TCL1_RING_MSI1_DATA_OFFSET(ab),
1353
srng->msi_data);
1354
}
1355
1356
ath12k_hif_write32(ab, reg_base, srng->ring_base_paddr);
1357
1358
val = u32_encode_bits(((u64)srng->ring_base_paddr >> HAL_ADDR_MSB_REG_SHIFT),
1359
HAL_TCL1_RING_BASE_MSB_RING_BASE_ADDR_MSB) |
1360
u32_encode_bits((srng->entry_size * srng->num_entries),
1361
HAL_TCL1_RING_BASE_MSB_RING_SIZE);
1362
ath12k_hif_write32(ab, reg_base + HAL_TCL1_RING_BASE_MSB_OFFSET, val);
1363
1364
val = u32_encode_bits(srng->entry_size, HAL_REO1_RING_ID_ENTRY_SIZE);
1365
ath12k_hif_write32(ab, reg_base + HAL_TCL1_RING_ID_OFFSET(ab), val);
1366
1367
val = u32_encode_bits(srng->intr_timer_thres_us,
1368
HAL_TCL1_RING_CONSR_INT_SETUP_IX0_INTR_TMR_THOLD);
1369
1370
val |= u32_encode_bits((srng->intr_batch_cntr_thres_entries * srng->entry_size),
1371
HAL_TCL1_RING_CONSR_INT_SETUP_IX0_BATCH_COUNTER_THOLD);
1372
1373
ath12k_hif_write32(ab,
1374
reg_base + HAL_TCL1_RING_CONSR_INT_SETUP_IX0_OFFSET(ab),
1375
val);
1376
1377
val = 0;
1378
if (srng->flags & HAL_SRNG_FLAGS_LOW_THRESH_INTR_EN) {
1379
val |= u32_encode_bits(srng->u.src_ring.low_threshold,
1380
HAL_TCL1_RING_CONSR_INT_SETUP_IX1_LOW_THOLD);
1381
}
1382
ath12k_hif_write32(ab,
1383
reg_base + HAL_TCL1_RING_CONSR_INT_SETUP_IX1_OFFSET(ab),
1384
val);
1385
1386
if (srng->ring_id != HAL_SRNG_RING_ID_WBM_IDLE_LINK) {
1387
tp_addr = hal->rdp.paddr +
1388
((unsigned long)srng->u.src_ring.tp_addr -
1389
(unsigned long)hal->rdp.vaddr);
1390
ath12k_hif_write32(ab,
1391
reg_base + HAL_TCL1_RING_TP_ADDR_LSB_OFFSET(ab),
1392
tp_addr & HAL_ADDR_LSB_REG_MASK);
1393
ath12k_hif_write32(ab,
1394
reg_base + HAL_TCL1_RING_TP_ADDR_MSB_OFFSET(ab),
1395
tp_addr >> HAL_ADDR_MSB_REG_SHIFT);
1396
}
1397
1398
/* Initialize head and tail pointers to indicate ring is empty */
1399
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R2];
1400
ath12k_hif_write32(ab, reg_base, 0);
1401
ath12k_hif_write32(ab, reg_base + HAL_TCL1_RING_TP_OFFSET, 0);
1402
*srng->u.src_ring.tp_addr = 0;
1403
1404
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R0];
1405
val = 0;
1406
if (srng->flags & HAL_SRNG_FLAGS_DATA_TLV_SWAP)
1407
val |= HAL_TCL1_RING_MISC_DATA_TLV_SWAP;
1408
if (srng->flags & HAL_SRNG_FLAGS_RING_PTR_SWAP)
1409
val |= HAL_TCL1_RING_MISC_HOST_FW_SWAP;
1410
if (srng->flags & HAL_SRNG_FLAGS_MSI_SWAP)
1411
val |= HAL_TCL1_RING_MISC_MSI_SWAP;
1412
1413
/* Loop count is not used for SRC rings */
1414
val |= HAL_TCL1_RING_MISC_MSI_LOOPCNT_DISABLE;
1415
1416
val |= HAL_TCL1_RING_MISC_SRNG_ENABLE;
1417
1418
if (srng->ring_id == HAL_SRNG_RING_ID_WBM_IDLE_LINK)
1419
val |= HAL_TCL1_RING_MISC_MSI_RING_ID_DISABLE;
1420
1421
ath12k_hif_write32(ab, reg_base + HAL_TCL1_RING_MISC_OFFSET(ab), val);
1422
}
1423
1424
static void ath12k_hal_srng_hw_init(struct ath12k_base *ab,
1425
struct hal_srng *srng)
1426
{
1427
if (srng->ring_dir == HAL_SRNG_DIR_SRC)
1428
ath12k_hal_srng_src_hw_init(ab, srng);
1429
else
1430
ath12k_hal_srng_dst_hw_init(ab, srng);
1431
}
1432
1433
static int ath12k_hal_srng_get_ring_id(struct ath12k_base *ab,
1434
enum hal_ring_type type,
1435
int ring_num, int mac_id)
1436
{
1437
struct hal_srng_config *srng_config = &ab->hal.srng_config[type];
1438
int ring_id;
1439
1440
if (ring_num >= srng_config->max_rings) {
1441
ath12k_warn(ab, "invalid ring number :%d\n", ring_num);
1442
return -EINVAL;
1443
}
1444
1445
ring_id = srng_config->start_ring_id + ring_num;
1446
if (srng_config->mac_type == ATH12K_HAL_SRNG_PMAC)
1447
ring_id += mac_id * HAL_SRNG_RINGS_PER_PMAC;
1448
1449
if (WARN_ON(ring_id >= HAL_SRNG_RING_ID_MAX))
1450
return -EINVAL;
1451
1452
return ring_id;
1453
}
1454
1455
int ath12k_hal_srng_get_entrysize(struct ath12k_base *ab, u32 ring_type)
1456
{
1457
struct hal_srng_config *srng_config;
1458
1459
if (WARN_ON(ring_type >= HAL_MAX_RING_TYPES))
1460
return -EINVAL;
1461
1462
srng_config = &ab->hal.srng_config[ring_type];
1463
1464
return (srng_config->entry_size << 2);
1465
}
1466
1467
int ath12k_hal_srng_get_max_entries(struct ath12k_base *ab, u32 ring_type)
1468
{
1469
struct hal_srng_config *srng_config;
1470
1471
if (WARN_ON(ring_type >= HAL_MAX_RING_TYPES))
1472
return -EINVAL;
1473
1474
srng_config = &ab->hal.srng_config[ring_type];
1475
1476
return (srng_config->max_size / srng_config->entry_size);
1477
}
1478
1479
void ath12k_hal_srng_get_params(struct ath12k_base *ab, struct hal_srng *srng,
1480
struct hal_srng_params *params)
1481
{
1482
params->ring_base_paddr = srng->ring_base_paddr;
1483
params->ring_base_vaddr = srng->ring_base_vaddr;
1484
params->num_entries = srng->num_entries;
1485
params->intr_timer_thres_us = srng->intr_timer_thres_us;
1486
params->intr_batch_cntr_thres_entries =
1487
srng->intr_batch_cntr_thres_entries;
1488
params->low_threshold = srng->u.src_ring.low_threshold;
1489
params->msi_addr = srng->msi_addr;
1490
params->msi2_addr = srng->msi2_addr;
1491
params->msi_data = srng->msi_data;
1492
params->msi2_data = srng->msi2_data;
1493
params->flags = srng->flags;
1494
}
1495
1496
dma_addr_t ath12k_hal_srng_get_hp_addr(struct ath12k_base *ab,
1497
struct hal_srng *srng)
1498
{
1499
if (!(srng->flags & HAL_SRNG_FLAGS_LMAC_RING))
1500
return 0;
1501
1502
if (srng->ring_dir == HAL_SRNG_DIR_SRC)
1503
return ab->hal.wrp.paddr +
1504
((unsigned long)srng->u.src_ring.hp_addr -
1505
(unsigned long)ab->hal.wrp.vaddr);
1506
else
1507
return ab->hal.rdp.paddr +
1508
((unsigned long)srng->u.dst_ring.hp_addr -
1509
(unsigned long)ab->hal.rdp.vaddr);
1510
}
1511
1512
dma_addr_t ath12k_hal_srng_get_tp_addr(struct ath12k_base *ab,
1513
struct hal_srng *srng)
1514
{
1515
if (!(srng->flags & HAL_SRNG_FLAGS_LMAC_RING))
1516
return 0;
1517
1518
if (srng->ring_dir == HAL_SRNG_DIR_SRC)
1519
return ab->hal.rdp.paddr +
1520
((unsigned long)srng->u.src_ring.tp_addr -
1521
(unsigned long)ab->hal.rdp.vaddr);
1522
else
1523
return ab->hal.wrp.paddr +
1524
((unsigned long)srng->u.dst_ring.tp_addr -
1525
(unsigned long)ab->hal.wrp.vaddr);
1526
}
1527
1528
u32 ath12k_hal_ce_get_desc_size(enum hal_ce_desc type)
1529
{
1530
switch (type) {
1531
case HAL_CE_DESC_SRC:
1532
return sizeof(struct hal_ce_srng_src_desc);
1533
case HAL_CE_DESC_DST:
1534
return sizeof(struct hal_ce_srng_dest_desc);
1535
case HAL_CE_DESC_DST_STATUS:
1536
return sizeof(struct hal_ce_srng_dst_status_desc);
1537
}
1538
1539
return 0;
1540
}
1541
1542
void ath12k_hal_ce_src_set_desc(struct hal_ce_srng_src_desc *desc, dma_addr_t paddr,
1543
u32 len, u32 id, u8 byte_swap_data)
1544
{
1545
desc->buffer_addr_low = cpu_to_le32(paddr & HAL_ADDR_LSB_REG_MASK);
1546
desc->buffer_addr_info =
1547
le32_encode_bits(((u64)paddr >> HAL_ADDR_MSB_REG_SHIFT),
1548
HAL_CE_SRC_DESC_ADDR_INFO_ADDR_HI) |
1549
le32_encode_bits(byte_swap_data,
1550
HAL_CE_SRC_DESC_ADDR_INFO_BYTE_SWAP) |
1551
le32_encode_bits(0, HAL_CE_SRC_DESC_ADDR_INFO_GATHER) |
1552
le32_encode_bits(len, HAL_CE_SRC_DESC_ADDR_INFO_LEN);
1553
desc->meta_info = le32_encode_bits(id, HAL_CE_SRC_DESC_META_INFO_DATA);
1554
}
1555
1556
void ath12k_hal_ce_dst_set_desc(struct hal_ce_srng_dest_desc *desc, dma_addr_t paddr)
1557
{
1558
desc->buffer_addr_low = cpu_to_le32(paddr & HAL_ADDR_LSB_REG_MASK);
1559
desc->buffer_addr_info =
1560
le32_encode_bits(((u64)paddr >> HAL_ADDR_MSB_REG_SHIFT),
1561
HAL_CE_DEST_DESC_ADDR_INFO_ADDR_HI);
1562
}
1563
1564
u32 ath12k_hal_ce_dst_status_get_length(struct hal_ce_srng_dst_status_desc *desc)
1565
{
1566
u32 len;
1567
1568
len = le32_get_bits(desc->flags, HAL_CE_DST_STATUS_DESC_FLAGS_LEN);
1569
desc->flags &= ~cpu_to_le32(HAL_CE_DST_STATUS_DESC_FLAGS_LEN);
1570
1571
return len;
1572
}
1573
1574
void ath12k_hal_set_link_desc_addr(struct hal_wbm_link_desc *desc, u32 cookie,
1575
dma_addr_t paddr)
1576
{
1577
desc->buf_addr_info.info0 = le32_encode_bits((paddr & HAL_ADDR_LSB_REG_MASK),
1578
BUFFER_ADDR_INFO0_ADDR);
1579
desc->buf_addr_info.info1 =
1580
le32_encode_bits(((u64)paddr >> HAL_ADDR_MSB_REG_SHIFT),
1581
BUFFER_ADDR_INFO1_ADDR) |
1582
le32_encode_bits(1, BUFFER_ADDR_INFO1_RET_BUF_MGR) |
1583
le32_encode_bits(cookie, BUFFER_ADDR_INFO1_SW_COOKIE);
1584
}
1585
1586
void *ath12k_hal_srng_dst_peek(struct ath12k_base *ab, struct hal_srng *srng)
1587
{
1588
lockdep_assert_held(&srng->lock);
1589
1590
if (srng->u.dst_ring.tp != srng->u.dst_ring.cached_hp)
1591
return (srng->ring_base_vaddr + srng->u.dst_ring.tp);
1592
1593
return NULL;
1594
}
1595
1596
void *ath12k_hal_srng_dst_get_next_entry(struct ath12k_base *ab,
1597
struct hal_srng *srng)
1598
{
1599
void *desc;
1600
1601
lockdep_assert_held(&srng->lock);
1602
1603
if (srng->u.dst_ring.tp == srng->u.dst_ring.cached_hp)
1604
return NULL;
1605
1606
desc = srng->ring_base_vaddr + srng->u.dst_ring.tp;
1607
1608
srng->u.dst_ring.tp = (srng->u.dst_ring.tp + srng->entry_size) %
1609
srng->ring_size;
1610
1611
return desc;
1612
}
1613
1614
int ath12k_hal_srng_dst_num_free(struct ath12k_base *ab, struct hal_srng *srng,
1615
bool sync_hw_ptr)
1616
{
1617
u32 tp, hp;
1618
1619
lockdep_assert_held(&srng->lock);
1620
1621
tp = srng->u.dst_ring.tp;
1622
1623
if (sync_hw_ptr) {
1624
hp = *srng->u.dst_ring.hp_addr;
1625
srng->u.dst_ring.cached_hp = hp;
1626
} else {
1627
hp = srng->u.dst_ring.cached_hp;
1628
}
1629
1630
if (hp >= tp)
1631
return (hp - tp) / srng->entry_size;
1632
else
1633
return (srng->ring_size - tp + hp) / srng->entry_size;
1634
}
1635
1636
/* Returns number of available entries in src ring */
1637
int ath12k_hal_srng_src_num_free(struct ath12k_base *ab, struct hal_srng *srng,
1638
bool sync_hw_ptr)
1639
{
1640
u32 tp, hp;
1641
1642
lockdep_assert_held(&srng->lock);
1643
1644
hp = srng->u.src_ring.hp;
1645
1646
if (sync_hw_ptr) {
1647
tp = *srng->u.src_ring.tp_addr;
1648
srng->u.src_ring.cached_tp = tp;
1649
} else {
1650
tp = srng->u.src_ring.cached_tp;
1651
}
1652
1653
if (tp > hp)
1654
return ((tp - hp) / srng->entry_size) - 1;
1655
else
1656
return ((srng->ring_size - hp + tp) / srng->entry_size) - 1;
1657
}
1658
1659
void *ath12k_hal_srng_src_get_next_entry(struct ath12k_base *ab,
1660
struct hal_srng *srng)
1661
{
1662
void *desc;
1663
u32 next_hp;
1664
1665
lockdep_assert_held(&srng->lock);
1666
1667
/* TODO: Using % is expensive, but we have to do this since size of some
1668
* SRNG rings is not power of 2 (due to descriptor sizes). Need to see
1669
* if separate function is defined for rings having power of 2 ring size
1670
* (TCL2SW, REO2SW, SW2RXDMA and CE rings) so that we can avoid the
1671
* overhead of % by using mask (with &).
1672
*/
1673
next_hp = (srng->u.src_ring.hp + srng->entry_size) % srng->ring_size;
1674
1675
if (next_hp == srng->u.src_ring.cached_tp)
1676
return NULL;
1677
1678
desc = srng->ring_base_vaddr + srng->u.src_ring.hp;
1679
srng->u.src_ring.hp = next_hp;
1680
1681
/* TODO: Reap functionality is not used by all rings. If particular
1682
* ring does not use reap functionality, we need not update reap_hp
1683
* with next_hp pointer. Need to make sure a separate function is used
1684
* before doing any optimization by removing below code updating
1685
* reap_hp.
1686
*/
1687
srng->u.src_ring.reap_hp = next_hp;
1688
1689
return desc;
1690
}
1691
1692
void *ath12k_hal_srng_src_reap_next(struct ath12k_base *ab,
1693
struct hal_srng *srng)
1694
{
1695
void *desc;
1696
u32 next_reap_hp;
1697
1698
lockdep_assert_held(&srng->lock);
1699
1700
next_reap_hp = (srng->u.src_ring.reap_hp + srng->entry_size) %
1701
srng->ring_size;
1702
1703
if (next_reap_hp == srng->u.src_ring.cached_tp)
1704
return NULL;
1705
1706
desc = srng->ring_base_vaddr + next_reap_hp;
1707
srng->u.src_ring.reap_hp = next_reap_hp;
1708
1709
return desc;
1710
}
1711
1712
void *ath12k_hal_srng_src_get_next_reaped(struct ath12k_base *ab,
1713
struct hal_srng *srng)
1714
{
1715
void *desc;
1716
1717
lockdep_assert_held(&srng->lock);
1718
1719
if (srng->u.src_ring.hp == srng->u.src_ring.reap_hp)
1720
return NULL;
1721
1722
desc = srng->ring_base_vaddr + srng->u.src_ring.hp;
1723
srng->u.src_ring.hp = (srng->u.src_ring.hp + srng->entry_size) %
1724
srng->ring_size;
1725
1726
return desc;
1727
}
1728
1729
void ath12k_hal_srng_access_begin(struct ath12k_base *ab, struct hal_srng *srng)
1730
{
1731
lockdep_assert_held(&srng->lock);
1732
1733
if (srng->ring_dir == HAL_SRNG_DIR_SRC)
1734
srng->u.src_ring.cached_tp =
1735
*(volatile u32 *)srng->u.src_ring.tp_addr;
1736
else
1737
srng->u.dst_ring.cached_hp = *srng->u.dst_ring.hp_addr;
1738
}
1739
1740
/* Update cached ring head/tail pointers to HW. ath12k_hal_srng_access_begin()
1741
* should have been called before this.
1742
*/
1743
void ath12k_hal_srng_access_end(struct ath12k_base *ab, struct hal_srng *srng)
1744
{
1745
lockdep_assert_held(&srng->lock);
1746
1747
/* TODO: See if we need a write memory barrier here */
1748
if (srng->flags & HAL_SRNG_FLAGS_LMAC_RING) {
1749
/* For LMAC rings, ring pointer updates are done through FW and
1750
* hence written to a shared memory location that is read by FW
1751
*/
1752
if (srng->ring_dir == HAL_SRNG_DIR_SRC) {
1753
srng->u.src_ring.last_tp =
1754
*(volatile u32 *)srng->u.src_ring.tp_addr;
1755
*srng->u.src_ring.hp_addr = srng->u.src_ring.hp;
1756
} else {
1757
srng->u.dst_ring.last_hp = *srng->u.dst_ring.hp_addr;
1758
*srng->u.dst_ring.tp_addr = srng->u.dst_ring.tp;
1759
}
1760
} else {
1761
if (srng->ring_dir == HAL_SRNG_DIR_SRC) {
1762
srng->u.src_ring.last_tp =
1763
*(volatile u32 *)srng->u.src_ring.tp_addr;
1764
ath12k_hif_write32(ab,
1765
(unsigned long)srng->u.src_ring.hp_addr -
1766
(unsigned long)ab->mem,
1767
srng->u.src_ring.hp);
1768
} else {
1769
srng->u.dst_ring.last_hp = *srng->u.dst_ring.hp_addr;
1770
ath12k_hif_write32(ab,
1771
(unsigned long)srng->u.dst_ring.tp_addr -
1772
(unsigned long)ab->mem,
1773
srng->u.dst_ring.tp);
1774
}
1775
}
1776
1777
srng->timestamp = jiffies;
1778
}
1779
1780
void ath12k_hal_setup_link_idle_list(struct ath12k_base *ab,
1781
struct hal_wbm_idle_scatter_list *sbuf,
1782
u32 nsbufs, u32 tot_link_desc,
1783
u32 end_offset)
1784
{
1785
struct ath12k_buffer_addr *link_addr;
1786
int i;
1787
u32 reg_scatter_buf_sz = HAL_WBM_IDLE_SCATTER_BUF_SIZE / 64;
1788
u32 val;
1789
1790
#if defined(__linux__)
1791
link_addr = (void *)sbuf[0].vaddr + HAL_WBM_IDLE_SCATTER_BUF_SIZE;
1792
#elif defined(__FreeBSD__)
1793
link_addr = (void *)((uintptr_t)sbuf[0].vaddr + HAL_WBM_IDLE_SCATTER_BUF_SIZE);
1794
#endif
1795
1796
for (i = 1; i < nsbufs; i++) {
1797
link_addr->info0 = cpu_to_le32(sbuf[i].paddr & HAL_ADDR_LSB_REG_MASK);
1798
1799
link_addr->info1 =
1800
le32_encode_bits((u64)sbuf[i].paddr >> HAL_ADDR_MSB_REG_SHIFT,
1801
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_39_32) |
1802
le32_encode_bits(BASE_ADDR_MATCH_TAG_VAL,
1803
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_MATCH_TAG);
1804
1805
#if defined(__linux__)
1806
link_addr = (void *)sbuf[i].vaddr +
1807
HAL_WBM_IDLE_SCATTER_BUF_SIZE;
1808
#elif defined(__FreeBSD__)
1809
link_addr = (void *)((uintptr_t)sbuf[i].vaddr +
1810
HAL_WBM_IDLE_SCATTER_BUF_SIZE);
1811
#endif
1812
}
1813
1814
val = u32_encode_bits(reg_scatter_buf_sz, HAL_WBM_SCATTER_BUFFER_SIZE) |
1815
u32_encode_bits(0x1, HAL_WBM_LINK_DESC_IDLE_LIST_MODE);
1816
1817
ath12k_hif_write32(ab,
1818
HAL_SEQ_WCSS_UMAC_WBM_REG +
1819
HAL_WBM_R0_IDLE_LIST_CONTROL_ADDR(ab),
1820
val);
1821
1822
val = u32_encode_bits(reg_scatter_buf_sz * nsbufs,
1823
HAL_WBM_SCATTER_RING_SIZE_OF_IDLE_LINK_DESC_LIST);
1824
ath12k_hif_write32(ab,
1825
HAL_SEQ_WCSS_UMAC_WBM_REG + HAL_WBM_R0_IDLE_LIST_SIZE_ADDR(ab),
1826
val);
1827
1828
val = u32_encode_bits(sbuf[0].paddr & HAL_ADDR_LSB_REG_MASK,
1829
BUFFER_ADDR_INFO0_ADDR);
1830
ath12k_hif_write32(ab,
1831
HAL_SEQ_WCSS_UMAC_WBM_REG +
1832
HAL_WBM_SCATTERED_RING_BASE_LSB(ab),
1833
val);
1834
1835
val = u32_encode_bits(BASE_ADDR_MATCH_TAG_VAL,
1836
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_MATCH_TAG) |
1837
u32_encode_bits((u64)sbuf[0].paddr >> HAL_ADDR_MSB_REG_SHIFT,
1838
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_39_32);
1839
ath12k_hif_write32(ab,
1840
HAL_SEQ_WCSS_UMAC_WBM_REG +
1841
HAL_WBM_SCATTERED_RING_BASE_MSB(ab),
1842
val);
1843
1844
/* Setup head and tail pointers for the idle list */
1845
val = u32_encode_bits(sbuf[nsbufs - 1].paddr, BUFFER_ADDR_INFO0_ADDR);
1846
ath12k_hif_write32(ab,
1847
HAL_SEQ_WCSS_UMAC_WBM_REG +
1848
HAL_WBM_SCATTERED_DESC_PTR_HEAD_INFO_IX0(ab),
1849
val);
1850
1851
val = u32_encode_bits(((u64)sbuf[nsbufs - 1].paddr >> HAL_ADDR_MSB_REG_SHIFT),
1852
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_39_32) |
1853
u32_encode_bits((end_offset >> 2),
1854
HAL_WBM_SCATTERED_DESC_HEAD_P_OFFSET_IX1);
1855
ath12k_hif_write32(ab,
1856
HAL_SEQ_WCSS_UMAC_WBM_REG +
1857
HAL_WBM_SCATTERED_DESC_PTR_HEAD_INFO_IX1(ab),
1858
val);
1859
1860
val = u32_encode_bits(sbuf[0].paddr, BUFFER_ADDR_INFO0_ADDR);
1861
ath12k_hif_write32(ab,
1862
HAL_SEQ_WCSS_UMAC_WBM_REG +
1863
HAL_WBM_SCATTERED_DESC_PTR_HEAD_INFO_IX0(ab),
1864
val);
1865
1866
val = u32_encode_bits(sbuf[0].paddr, BUFFER_ADDR_INFO0_ADDR);
1867
ath12k_hif_write32(ab,
1868
HAL_SEQ_WCSS_UMAC_WBM_REG +
1869
HAL_WBM_SCATTERED_DESC_PTR_TAIL_INFO_IX0(ab),
1870
val);
1871
1872
val = u32_encode_bits(((u64)sbuf[0].paddr >> HAL_ADDR_MSB_REG_SHIFT),
1873
HAL_WBM_SCATTERED_DESC_MSB_BASE_ADDR_39_32) |
1874
u32_encode_bits(0, HAL_WBM_SCATTERED_DESC_TAIL_P_OFFSET_IX1);
1875
ath12k_hif_write32(ab,
1876
HAL_SEQ_WCSS_UMAC_WBM_REG +
1877
HAL_WBM_SCATTERED_DESC_PTR_TAIL_INFO_IX1(ab),
1878
val);
1879
1880
val = 2 * tot_link_desc;
1881
ath12k_hif_write32(ab,
1882
HAL_SEQ_WCSS_UMAC_WBM_REG +
1883
HAL_WBM_SCATTERED_DESC_PTR_HP_ADDR(ab),
1884
val);
1885
1886
/* Enable the SRNG */
1887
val = u32_encode_bits(1, HAL_WBM_IDLE_LINK_RING_MISC_SRNG_ENABLE) |
1888
u32_encode_bits(1, HAL_WBM_IDLE_LINK_RING_MISC_RIND_ID_DISABLE);
1889
ath12k_hif_write32(ab,
1890
HAL_SEQ_WCSS_UMAC_WBM_REG +
1891
HAL_WBM_IDLE_LINK_RING_MISC_ADDR(ab),
1892
val);
1893
}
1894
1895
int ath12k_hal_srng_setup(struct ath12k_base *ab, enum hal_ring_type type,
1896
int ring_num, int mac_id,
1897
struct hal_srng_params *params)
1898
{
1899
struct ath12k_hal *hal = &ab->hal;
1900
struct hal_srng_config *srng_config = &ab->hal.srng_config[type];
1901
struct hal_srng *srng;
1902
int ring_id;
1903
u32 idx;
1904
int i;
1905
u32 reg_base;
1906
1907
ring_id = ath12k_hal_srng_get_ring_id(ab, type, ring_num, mac_id);
1908
if (ring_id < 0)
1909
return ring_id;
1910
1911
srng = &hal->srng_list[ring_id];
1912
1913
srng->ring_id = ring_id;
1914
srng->ring_dir = srng_config->ring_dir;
1915
srng->ring_base_paddr = params->ring_base_paddr;
1916
srng->ring_base_vaddr = params->ring_base_vaddr;
1917
srng->entry_size = srng_config->entry_size;
1918
srng->num_entries = params->num_entries;
1919
srng->ring_size = srng->entry_size * srng->num_entries;
1920
srng->intr_batch_cntr_thres_entries =
1921
params->intr_batch_cntr_thres_entries;
1922
srng->intr_timer_thres_us = params->intr_timer_thres_us;
1923
srng->flags = params->flags;
1924
srng->msi_addr = params->msi_addr;
1925
srng->msi2_addr = params->msi2_addr;
1926
srng->msi_data = params->msi_data;
1927
srng->msi2_data = params->msi2_data;
1928
srng->initialized = 1;
1929
spin_lock_init(&srng->lock);
1930
lockdep_set_class(&srng->lock, &srng->lock_key);
1931
1932
for (i = 0; i < HAL_SRNG_NUM_REG_GRP; i++) {
1933
srng->hwreg_base[i] = srng_config->reg_start[i] +
1934
(ring_num * srng_config->reg_size[i]);
1935
}
1936
1937
memset(srng->ring_base_vaddr, 0,
1938
(srng->entry_size * srng->num_entries) << 2);
1939
1940
reg_base = srng->hwreg_base[HAL_SRNG_REG_GRP_R2];
1941
1942
if (srng->ring_dir == HAL_SRNG_DIR_SRC) {
1943
srng->u.src_ring.hp = 0;
1944
srng->u.src_ring.cached_tp = 0;
1945
srng->u.src_ring.reap_hp = srng->ring_size - srng->entry_size;
1946
srng->u.src_ring.tp_addr = (void *)(hal->rdp.vaddr + ring_id);
1947
srng->u.src_ring.low_threshold = params->low_threshold *
1948
srng->entry_size;
1949
if (srng_config->mac_type == ATH12K_HAL_SRNG_UMAC) {
1950
if (!ab->hw_params->supports_shadow_regs)
1951
srng->u.src_ring.hp_addr =
1952
(u32 *)((unsigned long)ab->mem + reg_base);
1953
else
1954
ath12k_dbg(ab, ATH12K_DBG_HAL,
1955
"hal type %d ring_num %d reg_base 0x%x shadow 0x%lx\n",
1956
type, ring_num,
1957
reg_base,
1958
(unsigned long)srng->u.src_ring.hp_addr -
1959
(unsigned long)ab->mem);
1960
} else {
1961
idx = ring_id - HAL_SRNG_RING_ID_DMAC_CMN_ID_START;
1962
srng->u.src_ring.hp_addr = (void *)(hal->wrp.vaddr +
1963
idx);
1964
srng->flags |= HAL_SRNG_FLAGS_LMAC_RING;
1965
}
1966
} else {
1967
/* During initialization loop count in all the descriptors
1968
* will be set to zero, and HW will set it to 1 on completing
1969
* descriptor update in first loop, and increments it by 1 on
1970
* subsequent loops (loop count wraps around after reaching
1971
* 0xffff). The 'loop_cnt' in SW ring state is the expected
1972
* loop count in descriptors updated by HW (to be processed
1973
* by SW).
1974
*/
1975
srng->u.dst_ring.loop_cnt = 1;
1976
srng->u.dst_ring.tp = 0;
1977
srng->u.dst_ring.cached_hp = 0;
1978
srng->u.dst_ring.hp_addr = (void *)(hal->rdp.vaddr + ring_id);
1979
if (srng_config->mac_type == ATH12K_HAL_SRNG_UMAC) {
1980
if (!ab->hw_params->supports_shadow_regs)
1981
srng->u.dst_ring.tp_addr =
1982
(u32 *)((unsigned long)ab->mem + reg_base +
1983
(HAL_REO1_RING_TP - HAL_REO1_RING_HP));
1984
else
1985
ath12k_dbg(ab, ATH12K_DBG_HAL,
1986
"type %d ring_num %d target_reg 0x%x shadow 0x%lx\n",
1987
type, ring_num,
1988
reg_base + HAL_REO1_RING_TP - HAL_REO1_RING_HP,
1989
(unsigned long)srng->u.dst_ring.tp_addr -
1990
(unsigned long)ab->mem);
1991
} else {
1992
/* For PMAC & DMAC rings, tail pointer updates will be done
1993
* through FW by writing to a shared memory location
1994
*/
1995
idx = ring_id - HAL_SRNG_RING_ID_DMAC_CMN_ID_START;
1996
srng->u.dst_ring.tp_addr = (void *)(hal->wrp.vaddr +
1997
idx);
1998
srng->flags |= HAL_SRNG_FLAGS_LMAC_RING;
1999
}
2000
}
2001
2002
if (srng_config->mac_type != ATH12K_HAL_SRNG_UMAC)
2003
return ring_id;
2004
2005
ath12k_hal_srng_hw_init(ab, srng);
2006
2007
if (type == HAL_CE_DST) {
2008
srng->u.dst_ring.max_buffer_length = params->max_buffer_len;
2009
ath12k_hal_ce_dst_setup(ab, srng, ring_num);
2010
}
2011
2012
return ring_id;
2013
}
2014
2015
static void ath12k_hal_srng_update_hp_tp_addr(struct ath12k_base *ab,
2016
int shadow_cfg_idx,
2017
enum hal_ring_type ring_type,
2018
int ring_num)
2019
{
2020
struct hal_srng *srng;
2021
struct ath12k_hal *hal = &ab->hal;
2022
int ring_id;
2023
struct hal_srng_config *srng_config = &hal->srng_config[ring_type];
2024
2025
ring_id = ath12k_hal_srng_get_ring_id(ab, ring_type, ring_num, 0);
2026
if (ring_id < 0)
2027
return;
2028
2029
srng = &hal->srng_list[ring_id];
2030
2031
if (srng_config->ring_dir == HAL_SRNG_DIR_DST)
2032
srng->u.dst_ring.tp_addr = (u32 *)(HAL_SHADOW_REG(shadow_cfg_idx) +
2033
(unsigned long)ab->mem);
2034
else
2035
srng->u.src_ring.hp_addr = (u32 *)(HAL_SHADOW_REG(shadow_cfg_idx) +
2036
(unsigned long)ab->mem);
2037
}
2038
2039
int ath12k_hal_srng_update_shadow_config(struct ath12k_base *ab,
2040
enum hal_ring_type ring_type,
2041
int ring_num)
2042
{
2043
struct ath12k_hal *hal = &ab->hal;
2044
struct hal_srng_config *srng_config = &hal->srng_config[ring_type];
2045
int shadow_cfg_idx = hal->num_shadow_reg_configured;
2046
u32 target_reg;
2047
2048
if (shadow_cfg_idx >= HAL_SHADOW_NUM_REGS)
2049
return -EINVAL;
2050
2051
hal->num_shadow_reg_configured++;
2052
2053
target_reg = srng_config->reg_start[HAL_HP_OFFSET_IN_REG_START];
2054
target_reg += srng_config->reg_size[HAL_HP_OFFSET_IN_REG_START] *
2055
ring_num;
2056
2057
/* For destination ring, shadow the TP */
2058
if (srng_config->ring_dir == HAL_SRNG_DIR_DST)
2059
target_reg += HAL_OFFSET_FROM_HP_TO_TP;
2060
2061
hal->shadow_reg_addr[shadow_cfg_idx] = target_reg;
2062
2063
/* update hp/tp addr to hal structure*/
2064
ath12k_hal_srng_update_hp_tp_addr(ab, shadow_cfg_idx, ring_type,
2065
ring_num);
2066
2067
ath12k_dbg(ab, ATH12K_DBG_HAL,
2068
"target_reg %x, shadow reg 0x%x shadow_idx 0x%x, ring_type %d, ring num %d",
2069
target_reg,
2070
HAL_SHADOW_REG(shadow_cfg_idx),
2071
shadow_cfg_idx,
2072
ring_type, ring_num);
2073
2074
return 0;
2075
}
2076
2077
void ath12k_hal_srng_shadow_config(struct ath12k_base *ab)
2078
{
2079
struct ath12k_hal *hal = &ab->hal;
2080
int ring_type, ring_num;
2081
2082
/* update all the non-CE srngs. */
2083
for (ring_type = 0; ring_type < HAL_MAX_RING_TYPES; ring_type++) {
2084
struct hal_srng_config *srng_config = &hal->srng_config[ring_type];
2085
2086
if (ring_type == HAL_CE_SRC ||
2087
ring_type == HAL_CE_DST ||
2088
ring_type == HAL_CE_DST_STATUS)
2089
continue;
2090
2091
if (srng_config->mac_type == ATH12K_HAL_SRNG_DMAC ||
2092
srng_config->mac_type == ATH12K_HAL_SRNG_PMAC)
2093
continue;
2094
2095
for (ring_num = 0; ring_num < srng_config->max_rings; ring_num++)
2096
ath12k_hal_srng_update_shadow_config(ab, ring_type, ring_num);
2097
}
2098
}
2099
2100
void ath12k_hal_srng_get_shadow_config(struct ath12k_base *ab,
2101
u32 **cfg, u32 *len)
2102
{
2103
struct ath12k_hal *hal = &ab->hal;
2104
2105
*len = hal->num_shadow_reg_configured;
2106
*cfg = hal->shadow_reg_addr;
2107
}
2108
2109
void ath12k_hal_srng_shadow_update_hp_tp(struct ath12k_base *ab,
2110
struct hal_srng *srng)
2111
{
2112
lockdep_assert_held(&srng->lock);
2113
2114
/* check whether the ring is empty. Update the shadow
2115
* HP only when then ring isn't' empty.
2116
*/
2117
if (srng->ring_dir == HAL_SRNG_DIR_SRC &&
2118
*srng->u.src_ring.tp_addr != srng->u.src_ring.hp)
2119
ath12k_hal_srng_access_end(ab, srng);
2120
}
2121
2122
static void ath12k_hal_register_srng_lock_keys(struct ath12k_base *ab)
2123
{
2124
#if defined(__linux__)
2125
struct ath12k_hal *hal = &ab->hal;
2126
#endif
2127
u32 ring_id;
2128
2129
for (ring_id = 0; ring_id < HAL_SRNG_RING_ID_MAX; ring_id++)
2130
lockdep_register_key(&hal->srng_list[ring_id].lock_key);
2131
}
2132
2133
static void ath12k_hal_unregister_srng_lock_keys(struct ath12k_base *ab)
2134
{
2135
#if defined(__linux__)
2136
struct ath12k_hal *hal = &ab->hal;
2137
#endif
2138
u32 ring_id;
2139
2140
for (ring_id = 0; ring_id < HAL_SRNG_RING_ID_MAX; ring_id++)
2141
lockdep_unregister_key(&hal->srng_list[ring_id].lock_key);
2142
}
2143
2144
int ath12k_hal_srng_init(struct ath12k_base *ab)
2145
{
2146
struct ath12k_hal *hal = &ab->hal;
2147
int ret;
2148
2149
memset(hal, 0, sizeof(*hal));
2150
2151
ret = ab->hw_params->hal_ops->create_srng_config(ab);
2152
if (ret)
2153
goto err_hal;
2154
2155
ret = ath12k_hal_alloc_cont_rdp(ab);
2156
if (ret)
2157
goto err_hal;
2158
2159
ret = ath12k_hal_alloc_cont_wrp(ab);
2160
if (ret)
2161
goto err_free_cont_rdp;
2162
2163
ath12k_hal_register_srng_lock_keys(ab);
2164
2165
return 0;
2166
2167
err_free_cont_rdp:
2168
ath12k_hal_free_cont_rdp(ab);
2169
2170
err_hal:
2171
return ret;
2172
}
2173
2174
void ath12k_hal_srng_deinit(struct ath12k_base *ab)
2175
{
2176
struct ath12k_hal *hal = &ab->hal;
2177
2178
ath12k_hal_unregister_srng_lock_keys(ab);
2179
ath12k_hal_free_cont_rdp(ab);
2180
ath12k_hal_free_cont_wrp(ab);
2181
kfree(hal->srng_config);
2182
hal->srng_config = NULL;
2183
}
2184
2185
void ath12k_hal_dump_srng_stats(struct ath12k_base *ab)
2186
{
2187
struct hal_srng *srng;
2188
struct ath12k_ext_irq_grp *irq_grp;
2189
struct ath12k_ce_pipe *ce_pipe;
2190
int i;
2191
2192
ath12k_err(ab, "Last interrupt received for each CE:\n");
2193
for (i = 0; i < ab->hw_params->ce_count; i++) {
2194
ce_pipe = &ab->ce.ce_pipe[i];
2195
2196
if (ath12k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR)
2197
continue;
2198
2199
ath12k_err(ab, "CE_id %d pipe_num %d %ums before\n",
2200
i, ce_pipe->pipe_num,
2201
jiffies_to_msecs(jiffies - ce_pipe->timestamp));
2202
}
2203
2204
ath12k_err(ab, "\nLast interrupt received for each group:\n");
2205
for (i = 0; i < ATH12K_EXT_IRQ_GRP_NUM_MAX; i++) {
2206
irq_grp = &ab->ext_irq_grp[i];
2207
ath12k_err(ab, "group_id %d %ums before\n",
2208
irq_grp->grp_id,
2209
jiffies_to_msecs(jiffies - irq_grp->timestamp));
2210
}
2211
2212
for (i = 0; i < HAL_SRNG_RING_ID_MAX; i++) {
2213
srng = &ab->hal.srng_list[i];
2214
2215
if (!srng->initialized)
2216
continue;
2217
2218
if (srng->ring_dir == HAL_SRNG_DIR_SRC)
2219
ath12k_err(ab,
2220
"src srng id %u hp %u, reap_hp %u, cur tp %u, cached tp %u last tp %u napi processed before %ums\n",
2221
srng->ring_id, srng->u.src_ring.hp,
2222
srng->u.src_ring.reap_hp,
2223
*srng->u.src_ring.tp_addr, srng->u.src_ring.cached_tp,
2224
srng->u.src_ring.last_tp,
2225
jiffies_to_msecs(jiffies - srng->timestamp));
2226
else if (srng->ring_dir == HAL_SRNG_DIR_DST)
2227
ath12k_err(ab,
2228
"dst srng id %u tp %u, cur hp %u, cached hp %u last hp %u napi processed before %ums\n",
2229
srng->ring_id, srng->u.dst_ring.tp,
2230
*srng->u.dst_ring.hp_addr,
2231
srng->u.dst_ring.cached_hp,
2232
srng->u.dst_ring.last_hp,
2233
jiffies_to_msecs(jiffies - srng->timestamp));
2234
}
2235
}
2236
2237