blob: d7ce1ac9adc9eb8928219c3b83dc9f4be5a769ff [file] [log] [blame]
Arend van Spriel5b435de2011-10-05 13:19:03 +02001/*
2 * Copyright (c) 2010 Broadcom Corporation
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
Joe Perches8505a7e2011-11-13 11:41:04 -080016
17#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18
Arend van Spriel5b435de2011-10-05 13:19:03 +020019#include <linux/slab.h>
Arend van Spriel5b435de2011-10-05 13:19:03 +020020#include <linux/delay.h>
21#include <linux/pci.h>
Seth Forsheee041f652012-11-15 08:07:56 -060022#include <net/cfg80211.h>
23#include <net/mac80211.h>
Arend van Spriel5b435de2011-10-05 13:19:03 +020024
25#include <brcmu_utils.h>
26#include <aiutils.h>
27#include "types.h"
Seth Forsheee041f652012-11-15 08:07:56 -060028#include "main.h"
Arend van Spriel5b435de2011-10-05 13:19:03 +020029#include "dma.h"
Alwin Beukers23038212011-10-18 14:02:58 +020030#include "soc.h"
Seth Forsheee041f652012-11-15 08:07:56 -060031#include "scb.h"
32#include "ampdu.h"
Arend van Spriel5b435de2011-10-05 13:19:03 +020033
34/*
Arend van Spriele81da652011-12-08 15:06:53 -080035 * dma register field offset calculation
36 */
37#define DMA64REGOFFS(field) offsetof(struct dma64regs, field)
38#define DMA64TXREGOFFS(di, field) (di->d64txregbase + DMA64REGOFFS(field))
39#define DMA64RXREGOFFS(di, field) (di->d64rxregbase + DMA64REGOFFS(field))
40
41/*
Arend van Spriel5b435de2011-10-05 13:19:03 +020042 * DMA hardware requires each descriptor ring to be 8kB aligned, and fit within
43 * a contiguous 8kB physical address.
44 */
45#define D64RINGALIGN_BITS 13
46#define D64MAXRINGSZ (1 << D64RINGALIGN_BITS)
47#define D64RINGALIGN (1 << D64RINGALIGN_BITS)
48
49#define D64MAXDD (D64MAXRINGSZ / sizeof(struct dma64desc))
50
51/* transmit channel control */
52#define D64_XC_XE 0x00000001 /* transmit enable */
53#define D64_XC_SE 0x00000002 /* transmit suspend request */
54#define D64_XC_LE 0x00000004 /* loopback enable */
55#define D64_XC_FL 0x00000010 /* flush request */
56#define D64_XC_PD 0x00000800 /* parity check disable */
57#define D64_XC_AE 0x00030000 /* address extension bits */
58#define D64_XC_AE_SHIFT 16
59
60/* transmit descriptor table pointer */
61#define D64_XP_LD_MASK 0x00000fff /* last valid descriptor */
62
63/* transmit channel status */
64#define D64_XS0_CD_MASK 0x00001fff /* current descriptor pointer */
65#define D64_XS0_XS_MASK 0xf0000000 /* transmit state */
66#define D64_XS0_XS_SHIFT 28
67#define D64_XS0_XS_DISABLED 0x00000000 /* disabled */
68#define D64_XS0_XS_ACTIVE 0x10000000 /* active */
69#define D64_XS0_XS_IDLE 0x20000000 /* idle wait */
70#define D64_XS0_XS_STOPPED 0x30000000 /* stopped */
71#define D64_XS0_XS_SUSP 0x40000000 /* suspend pending */
72
73#define D64_XS1_AD_MASK 0x00001fff /* active descriptor */
74#define D64_XS1_XE_MASK 0xf0000000 /* transmit errors */
75#define D64_XS1_XE_SHIFT 28
76#define D64_XS1_XE_NOERR 0x00000000 /* no error */
77#define D64_XS1_XE_DPE 0x10000000 /* descriptor protocol error */
78#define D64_XS1_XE_DFU 0x20000000 /* data fifo underrun */
79#define D64_XS1_XE_DTE 0x30000000 /* data transfer error */
80#define D64_XS1_XE_DESRE 0x40000000 /* descriptor read error */
81#define D64_XS1_XE_COREE 0x50000000 /* core error */
82
83/* receive channel control */
84/* receive enable */
85#define D64_RC_RE 0x00000001
86/* receive frame offset */
87#define D64_RC_RO_MASK 0x000000fe
88#define D64_RC_RO_SHIFT 1
89/* direct fifo receive (pio) mode */
90#define D64_RC_FM 0x00000100
91/* separate rx header descriptor enable */
92#define D64_RC_SH 0x00000200
93/* overflow continue */
94#define D64_RC_OC 0x00000400
95/* parity check disable */
96#define D64_RC_PD 0x00000800
97/* address extension bits */
98#define D64_RC_AE 0x00030000
99#define D64_RC_AE_SHIFT 16
100
101/* flags for dma controller */
102/* partity enable */
103#define DMA_CTRL_PEN (1 << 0)
104/* rx overflow continue */
105#define DMA_CTRL_ROC (1 << 1)
106/* allow rx scatter to multiple descriptors */
107#define DMA_CTRL_RXMULTI (1 << 2)
108/* Unframed Rx/Tx data */
109#define DMA_CTRL_UNFRAMED (1 << 3)
110
111/* receive descriptor table pointer */
112#define D64_RP_LD_MASK 0x00000fff /* last valid descriptor */
113
114/* receive channel status */
115#define D64_RS0_CD_MASK 0x00001fff /* current descriptor pointer */
116#define D64_RS0_RS_MASK 0xf0000000 /* receive state */
117#define D64_RS0_RS_SHIFT 28
118#define D64_RS0_RS_DISABLED 0x00000000 /* disabled */
119#define D64_RS0_RS_ACTIVE 0x10000000 /* active */
120#define D64_RS0_RS_IDLE 0x20000000 /* idle wait */
121#define D64_RS0_RS_STOPPED 0x30000000 /* stopped */
122#define D64_RS0_RS_SUSP 0x40000000 /* suspend pending */
123
124#define D64_RS1_AD_MASK 0x0001ffff /* active descriptor */
125#define D64_RS1_RE_MASK 0xf0000000 /* receive errors */
126#define D64_RS1_RE_SHIFT 28
127#define D64_RS1_RE_NOERR 0x00000000 /* no error */
128#define D64_RS1_RE_DPO 0x10000000 /* descriptor protocol error */
129#define D64_RS1_RE_DFU 0x20000000 /* data fifo overflow */
130#define D64_RS1_RE_DTE 0x30000000 /* data transfer error */
131#define D64_RS1_RE_DESRE 0x40000000 /* descriptor read error */
132#define D64_RS1_RE_COREE 0x50000000 /* core error */
133
134/* fifoaddr */
135#define D64_FA_OFF_MASK 0xffff /* offset */
136#define D64_FA_SEL_MASK 0xf0000 /* select */
137#define D64_FA_SEL_SHIFT 16
138#define D64_FA_SEL_XDD 0x00000 /* transmit dma data */
139#define D64_FA_SEL_XDP 0x10000 /* transmit dma pointers */
140#define D64_FA_SEL_RDD 0x40000 /* receive dma data */
141#define D64_FA_SEL_RDP 0x50000 /* receive dma pointers */
142#define D64_FA_SEL_XFD 0x80000 /* transmit fifo data */
143#define D64_FA_SEL_XFP 0x90000 /* transmit fifo pointers */
144#define D64_FA_SEL_RFD 0xc0000 /* receive fifo data */
145#define D64_FA_SEL_RFP 0xd0000 /* receive fifo pointers */
146#define D64_FA_SEL_RSD 0xe0000 /* receive frame status data */
147#define D64_FA_SEL_RSP 0xf0000 /* receive frame status pointers */
148
149/* descriptor control flags 1 */
150#define D64_CTRL_COREFLAGS 0x0ff00000 /* core specific flags */
151#define D64_CTRL1_EOT ((u32)1 << 28) /* end of descriptor table */
152#define D64_CTRL1_IOC ((u32)1 << 29) /* interrupt on completion */
153#define D64_CTRL1_EOF ((u32)1 << 30) /* end of frame */
154#define D64_CTRL1_SOF ((u32)1 << 31) /* start of frame */
155
156/* descriptor control flags 2 */
157/* buffer byte count. real data len must <= 16KB */
158#define D64_CTRL2_BC_MASK 0x00007fff
159/* address extension bits */
160#define D64_CTRL2_AE 0x00030000
161#define D64_CTRL2_AE_SHIFT 16
162/* parity bit */
163#define D64_CTRL2_PARITY 0x00040000
164
165/* control flags in the range [27:20] are core-specific and not defined here */
166#define D64_CTRL_CORE_MASK 0x0ff00000
167
168#define D64_RX_FRM_STS_LEN 0x0000ffff /* frame length mask */
169#define D64_RX_FRM_STS_OVFL 0x00800000 /* RxOverFlow */
170#define D64_RX_FRM_STS_DSCRCNT 0x0f000000 /* no. of descriptors used - 1 */
171#define D64_RX_FRM_STS_DATATYPE 0xf0000000 /* core-dependent data type */
172
173/*
174 * packet headroom necessary to accommodate the largest header
175 * in the system, (i.e TXOFF). By doing, we avoid the need to
176 * allocate an extra buffer for the header when bridging to WL.
177 * There is a compile time check in wlc.c which ensure that this
178 * value is at least as big as TXOFF. This value is used in
179 * dma_rxfill().
180 */
181
182#define BCMEXTRAHDROOM 172
183
184/* debug/trace */
Joe Perches8ae74652012-01-15 00:38:38 -0800185#ifdef DEBUG
Joe Perches8505a7e2011-11-13 11:41:04 -0800186#define DMA_ERROR(fmt, ...) \
187do { \
188 if (*di->msg_level & 1) \
189 pr_debug("%s: " fmt, __func__, ##__VA_ARGS__); \
190} while (0)
191#define DMA_TRACE(fmt, ...) \
192do { \
193 if (*di->msg_level & 2) \
194 pr_debug("%s: " fmt, __func__, ##__VA_ARGS__); \
195} while (0)
Arend van Spriel5b435de2011-10-05 13:19:03 +0200196#else
Joe Perches8505a7e2011-11-13 11:41:04 -0800197#define DMA_ERROR(fmt, ...) \
198 no_printk(fmt, ##__VA_ARGS__)
199#define DMA_TRACE(fmt, ...) \
200 no_printk(fmt, ##__VA_ARGS__)
Joe Perches8ae74652012-01-15 00:38:38 -0800201#endif /* DEBUG */
Arend van Spriel5b435de2011-10-05 13:19:03 +0200202
Joe Perches8505a7e2011-11-13 11:41:04 -0800203#define DMA_NONE(fmt, ...) \
204 no_printk(fmt, ##__VA_ARGS__)
Arend van Spriel5b435de2011-10-05 13:19:03 +0200205
206#define MAXNAMEL 8 /* 8 char names */
207
208/* macros to convert between byte offsets and indexes */
209#define B2I(bytes, type) ((bytes) / sizeof(type))
210#define I2B(index, type) ((index) * sizeof(type))
211
212#define PCI32ADDR_HIGH 0xc0000000 /* address[31:30] */
213#define PCI32ADDR_HIGH_SHIFT 30 /* address[31:30] */
214
215#define PCI64ADDR_HIGH 0x80000000 /* address[63] */
216#define PCI64ADDR_HIGH_SHIFT 31 /* address[63] */
217
218/*
219 * DMA Descriptor
220 * Descriptors are only read by the hardware, never written back.
221 */
222struct dma64desc {
223 __le32 ctrl1; /* misc control bits & bufcount */
224 __le32 ctrl2; /* buffer count and address extension */
225 __le32 addrlow; /* memory address of the date buffer, bits 31:0 */
226 __le32 addrhigh; /* memory address of the date buffer, bits 63:32 */
227};
228
229/* dma engine software state */
230struct dma_info {
231 struct dma_pub dma; /* exported structure */
232 uint *msg_level; /* message level pointer */
233 char name[MAXNAMEL]; /* callers name for diag msgs */
234
Arend van Spriel3b758a62011-12-12 15:15:09 -0800235 struct bcma_device *core;
Arend van Spriel2e81b9b2011-12-08 15:06:52 -0800236 struct device *dmadev;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200237
Seth Forsheee041f652012-11-15 08:07:56 -0600238 /* session information for AMPDU */
239 struct brcms_ampdu_session ampdu_session;
240
Arend van Spriel5b435de2011-10-05 13:19:03 +0200241 bool dma64; /* this dma engine is operating in 64-bit mode */
242 bool addrext; /* this dma engine supports DmaExtendedAddrChanges */
243
244 /* 64-bit dma tx engine registers */
Arend van Spriele81da652011-12-08 15:06:53 -0800245 uint d64txregbase;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200246 /* 64-bit dma rx engine registers */
Arend van Spriele81da652011-12-08 15:06:53 -0800247 uint d64rxregbase;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200248 /* pointer to dma64 tx descriptor ring */
249 struct dma64desc *txd64;
250 /* pointer to dma64 rx descriptor ring */
251 struct dma64desc *rxd64;
252
253 u16 dmadesc_align; /* alignment requirement for dma descriptors */
254
255 u16 ntxd; /* # tx descriptors tunable */
256 u16 txin; /* index of next descriptor to reclaim */
257 u16 txout; /* index of next descriptor to post */
258 /* pointer to parallel array of pointers to packets */
259 struct sk_buff **txp;
260 /* Aligned physical address of descriptor ring */
261 dma_addr_t txdpa;
262 /* Original physical address of descriptor ring */
263 dma_addr_t txdpaorig;
264 u16 txdalign; /* #bytes added to alloc'd mem to align txd */
265 u32 txdalloc; /* #bytes allocated for the ring */
266 u32 xmtptrbase; /* When using unaligned descriptors, the ptr register
267 * is not just an index, it needs all 13 bits to be
268 * an offset from the addr register.
269 */
270
271 u16 nrxd; /* # rx descriptors tunable */
272 u16 rxin; /* index of next descriptor to reclaim */
273 u16 rxout; /* index of next descriptor to post */
274 /* pointer to parallel array of pointers to packets */
275 struct sk_buff **rxp;
276 /* Aligned physical address of descriptor ring */
277 dma_addr_t rxdpa;
278 /* Original physical address of descriptor ring */
279 dma_addr_t rxdpaorig;
280 u16 rxdalign; /* #bytes added to alloc'd mem to align rxd */
281 u32 rxdalloc; /* #bytes allocated for the ring */
282 u32 rcvptrbase; /* Base for ptr reg when using unaligned descriptors */
283
284 /* tunables */
285 unsigned int rxbufsize; /* rx buffer size in bytes, not including
286 * the extra headroom
287 */
288 uint rxextrahdrroom; /* extra rx headroom, reverseved to assist upper
289 * stack, e.g. some rx pkt buffers will be
290 * bridged to tx side without byte copying.
291 * The extra headroom needs to be large enough
292 * to fit txheader needs. Some dongle driver may
293 * not need it.
294 */
295 uint nrxpost; /* # rx buffers to keep posted */
296 unsigned int rxoffset; /* rxcontrol offset */
297 /* add to get dma address of descriptor ring, low 32 bits */
298 uint ddoffsetlow;
299 /* high 32 bits */
300 uint ddoffsethigh;
301 /* add to get dma address of data buffer, low 32 bits */
302 uint dataoffsetlow;
303 /* high 32 bits */
304 uint dataoffsethigh;
305 /* descriptor base need to be aligned or not */
306 bool aligndesc_4k;
307};
308
309/*
310 * default dma message level (if input msg_level
311 * pointer is null in dma_attach())
312 */
313static uint dma_msg_level;
314
315/* Check for odd number of 1's */
316static u32 parity32(__le32 data)
317{
318 /* no swap needed for counting 1's */
319 u32 par_data = *(u32 *)&data;
320
321 par_data ^= par_data >> 16;
322 par_data ^= par_data >> 8;
323 par_data ^= par_data >> 4;
324 par_data ^= par_data >> 2;
325 par_data ^= par_data >> 1;
326
327 return par_data & 1;
328}
329
330static bool dma64_dd_parity(struct dma64desc *dd)
331{
332 return parity32(dd->addrlow ^ dd->addrhigh ^ dd->ctrl1 ^ dd->ctrl2);
333}
334
335/* descriptor bumping functions */
336
337static uint xxd(uint x, uint n)
338{
339 return x & (n - 1); /* faster than %, but n must be power of 2 */
340}
341
342static uint txd(struct dma_info *di, uint x)
343{
344 return xxd(x, di->ntxd);
345}
346
347static uint rxd(struct dma_info *di, uint x)
348{
349 return xxd(x, di->nrxd);
350}
351
352static uint nexttxd(struct dma_info *di, uint i)
353{
354 return txd(di, i + 1);
355}
356
357static uint prevtxd(struct dma_info *di, uint i)
358{
359 return txd(di, i - 1);
360}
361
362static uint nextrxd(struct dma_info *di, uint i)
363{
Seth Forsheeb05618d2012-11-15 08:07:57 -0600364 return rxd(di, i + 1);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200365}
366
367static uint ntxdactive(struct dma_info *di, uint h, uint t)
368{
369 return txd(di, t-h);
370}
371
372static uint nrxdactive(struct dma_info *di, uint h, uint t)
373{
374 return rxd(di, t-h);
375}
376
377static uint _dma_ctrlflags(struct dma_info *di, uint mask, uint flags)
378{
Arend van Sprielae8e4672011-10-29 11:30:15 +0200379 uint dmactrlflags;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200380
381 if (di == NULL) {
Joe Perches8505a7e2011-11-13 11:41:04 -0800382 DMA_ERROR("NULL dma handle\n");
Arend van Spriel5b435de2011-10-05 13:19:03 +0200383 return 0;
384 }
385
Arend van Sprielae8e4672011-10-29 11:30:15 +0200386 dmactrlflags = di->dma.dmactrlflags;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200387 dmactrlflags &= ~mask;
388 dmactrlflags |= flags;
389
390 /* If trying to enable parity, check if parity is actually supported */
391 if (dmactrlflags & DMA_CTRL_PEN) {
392 u32 control;
393
Arend van Spriel3b758a62011-12-12 15:15:09 -0800394 control = bcma_read32(di->core, DMA64TXREGOFFS(di, control));
395 bcma_write32(di->core, DMA64TXREGOFFS(di, control),
Arend van Spriel5b435de2011-10-05 13:19:03 +0200396 control | D64_XC_PD);
Arend van Spriel3b758a62011-12-12 15:15:09 -0800397 if (bcma_read32(di->core, DMA64TXREGOFFS(di, control)) &
Arend van Spriele81da652011-12-08 15:06:53 -0800398 D64_XC_PD)
Arend van Spriel5b435de2011-10-05 13:19:03 +0200399 /* We *can* disable it so it is supported,
400 * restore control register
401 */
Arend van Spriel3b758a62011-12-12 15:15:09 -0800402 bcma_write32(di->core, DMA64TXREGOFFS(di, control),
Arend van Spriele81da652011-12-08 15:06:53 -0800403 control);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200404 else
405 /* Not supported, don't allow it to be enabled */
406 dmactrlflags &= ~DMA_CTRL_PEN;
407 }
408
409 di->dma.dmactrlflags = dmactrlflags;
410
411 return dmactrlflags;
412}
413
Arend van Spriele81da652011-12-08 15:06:53 -0800414static bool _dma64_addrext(struct dma_info *di, uint ctrl_offset)
Arend van Spriel5b435de2011-10-05 13:19:03 +0200415{
416 u32 w;
Arend van Spriel3b758a62011-12-12 15:15:09 -0800417 bcma_set32(di->core, ctrl_offset, D64_XC_AE);
418 w = bcma_read32(di->core, ctrl_offset);
419 bcma_mask32(di->core, ctrl_offset, ~D64_XC_AE);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200420 return (w & D64_XC_AE) == D64_XC_AE;
421}
422
423/*
424 * return true if this dma engine supports DmaExtendedAddrChanges,
425 * otherwise false
426 */
427static bool _dma_isaddrext(struct dma_info *di)
428{
429 /* DMA64 supports full 32- or 64-bit operation. AE is always valid */
430
431 /* not all tx or rx channel are available */
Arend van Spriele81da652011-12-08 15:06:53 -0800432 if (di->d64txregbase != 0) {
433 if (!_dma64_addrext(di, DMA64TXREGOFFS(di, control)))
Joe Perches8505a7e2011-11-13 11:41:04 -0800434 DMA_ERROR("%s: DMA64 tx doesn't have AE set\n",
435 di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200436 return true;
Arend van Spriele81da652011-12-08 15:06:53 -0800437 } else if (di->d64rxregbase != 0) {
438 if (!_dma64_addrext(di, DMA64RXREGOFFS(di, control)))
Joe Perches8505a7e2011-11-13 11:41:04 -0800439 DMA_ERROR("%s: DMA64 rx doesn't have AE set\n",
440 di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200441 return true;
442 }
443
444 return false;
445}
446
447static bool _dma_descriptor_align(struct dma_info *di)
448{
449 u32 addrl;
450
451 /* Check to see if the descriptors need to be aligned on 4K/8K or not */
Arend van Spriele81da652011-12-08 15:06:53 -0800452 if (di->d64txregbase != 0) {
Arend van Spriel3b758a62011-12-12 15:15:09 -0800453 bcma_write32(di->core, DMA64TXREGOFFS(di, addrlow), 0xff0);
454 addrl = bcma_read32(di->core, DMA64TXREGOFFS(di, addrlow));
Arend van Spriel5b435de2011-10-05 13:19:03 +0200455 if (addrl != 0)
456 return false;
Arend van Spriele81da652011-12-08 15:06:53 -0800457 } else if (di->d64rxregbase != 0) {
Arend van Spriel3b758a62011-12-12 15:15:09 -0800458 bcma_write32(di->core, DMA64RXREGOFFS(di, addrlow), 0xff0);
459 addrl = bcma_read32(di->core, DMA64RXREGOFFS(di, addrlow));
Arend van Spriel5b435de2011-10-05 13:19:03 +0200460 if (addrl != 0)
461 return false;
462 }
463 return true;
464}
465
466/*
467 * Descriptor table must start at the DMA hardware dictated alignment, so
468 * allocated memory must be large enough to support this requirement.
469 */
Arend van Spriel2e81b9b2011-12-08 15:06:52 -0800470static void *dma_alloc_consistent(struct dma_info *di, uint size,
Arend van Spriel5b435de2011-10-05 13:19:03 +0200471 u16 align_bits, uint *alloced,
472 dma_addr_t *pap)
473{
474 if (align_bits) {
475 u16 align = (1 << align_bits);
476 if (!IS_ALIGNED(PAGE_SIZE, align))
477 size += align;
478 *alloced = size;
479 }
Arend van Spriel2e81b9b2011-12-08 15:06:52 -0800480 return dma_alloc_coherent(di->dmadev, size, pap, GFP_ATOMIC);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200481}
482
483static
484u8 dma_align_sizetobits(uint size)
485{
486 u8 bitpos = 0;
487 while (size >>= 1)
488 bitpos++;
489 return bitpos;
490}
491
492/* This function ensures that the DMA descriptor ring will not get allocated
493 * across Page boundary. If the allocation is done across the page boundary
494 * at the first time, then it is freed and the allocation is done at
495 * descriptor ring size aligned location. This will ensure that the ring will
496 * not cross page boundary
497 */
498static void *dma_ringalloc(struct dma_info *di, u32 boundary, uint size,
499 u16 *alignbits, uint *alloced,
500 dma_addr_t *descpa)
501{
502 void *va;
503 u32 desc_strtaddr;
504 u32 alignbytes = 1 << *alignbits;
505
Arend van Spriel2e81b9b2011-12-08 15:06:52 -0800506 va = dma_alloc_consistent(di, size, *alignbits, alloced, descpa);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200507
508 if (NULL == va)
509 return NULL;
510
511 desc_strtaddr = (u32) roundup((unsigned long)va, alignbytes);
512 if (((desc_strtaddr + size - 1) & boundary) != (desc_strtaddr
513 & boundary)) {
514 *alignbits = dma_align_sizetobits(size);
Arend van Spriel2e81b9b2011-12-08 15:06:52 -0800515 dma_free_coherent(di->dmadev, size, va, *descpa);
516 va = dma_alloc_consistent(di, size, *alignbits,
Arend van Spriel5b435de2011-10-05 13:19:03 +0200517 alloced, descpa);
518 }
519 return va;
520}
521
522static bool dma64_alloc(struct dma_info *di, uint direction)
523{
524 u16 size;
525 uint ddlen;
526 void *va;
527 uint alloced = 0;
528 u16 align;
529 u16 align_bits;
530
531 ddlen = sizeof(struct dma64desc);
532
533 size = (direction == DMA_TX) ? (di->ntxd * ddlen) : (di->nrxd * ddlen);
534 align_bits = di->dmadesc_align;
535 align = (1 << align_bits);
536
537 if (direction == DMA_TX) {
538 va = dma_ringalloc(di, D64RINGALIGN, size, &align_bits,
539 &alloced, &di->txdpaorig);
540 if (va == NULL) {
Joe Perches8505a7e2011-11-13 11:41:04 -0800541 DMA_ERROR("%s: DMA_ALLOC_CONSISTENT(ntxd) failed\n",
542 di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200543 return false;
544 }
545 align = (1 << align_bits);
546 di->txd64 = (struct dma64desc *)
547 roundup((unsigned long)va, align);
548 di->txdalign = (uint) ((s8 *)di->txd64 - (s8 *) va);
549 di->txdpa = di->txdpaorig + di->txdalign;
550 di->txdalloc = alloced;
551 } else {
552 va = dma_ringalloc(di, D64RINGALIGN, size, &align_bits,
553 &alloced, &di->rxdpaorig);
554 if (va == NULL) {
Joe Perches8505a7e2011-11-13 11:41:04 -0800555 DMA_ERROR("%s: DMA_ALLOC_CONSISTENT(nrxd) failed\n",
556 di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200557 return false;
558 }
559 align = (1 << align_bits);
560 di->rxd64 = (struct dma64desc *)
561 roundup((unsigned long)va, align);
562 di->rxdalign = (uint) ((s8 *)di->rxd64 - (s8 *) va);
563 di->rxdpa = di->rxdpaorig + di->rxdalign;
564 di->rxdalloc = alloced;
565 }
566
567 return true;
568}
569
570static bool _dma_alloc(struct dma_info *di, uint direction)
571{
572 return dma64_alloc(di, direction);
573}
574
Seth Forsheee041f652012-11-15 08:07:56 -0600575struct dma_pub *dma_attach(char *name, struct brcms_c_info *wlc,
Arend van Spriele81da652011-12-08 15:06:53 -0800576 uint txregbase, uint rxregbase, uint ntxd, uint nrxd,
Arend van Spriel2e81b9b2011-12-08 15:06:52 -0800577 uint rxbufsize, int rxextheadroom,
578 uint nrxpost, uint rxoffset, uint *msg_level)
Arend van Spriel5b435de2011-10-05 13:19:03 +0200579{
Seth Forsheee041f652012-11-15 08:07:56 -0600580 struct si_pub *sih = wlc->hw->sih;
581 struct bcma_device *core = wlc->hw->d11core;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200582 struct dma_info *di;
Arend van Spriel3b758a62011-12-12 15:15:09 -0800583 u8 rev = core->id.rev;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200584 uint size;
Hauke Mehrtensec5ab1d2012-06-30 15:16:18 +0200585 struct si_info *sii = container_of(sih, struct si_info, pub);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200586
587 /* allocate private info structure */
588 di = kzalloc(sizeof(struct dma_info), GFP_ATOMIC);
589 if (di == NULL)
590 return NULL;
591
592 di->msg_level = msg_level ? msg_level : &dma_msg_level;
593
594
Arend van Spriela8779e42011-12-08 15:06:58 -0800595 di->dma64 =
Arend van Spriel3b758a62011-12-12 15:15:09 -0800596 ((bcma_aread32(core, BCMA_IOST) & SISF_DMA64) == SISF_DMA64);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200597
Arend van Spriele81da652011-12-08 15:06:53 -0800598 /* init dma reg info */
Arend van Spriel3b758a62011-12-12 15:15:09 -0800599 di->core = core;
Arend van Spriele81da652011-12-08 15:06:53 -0800600 di->d64txregbase = txregbase;
601 di->d64rxregbase = rxregbase;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200602
603 /*
604 * Default flags (which can be changed by the driver calling
605 * dma_ctrlflags before enable): For backwards compatibility
606 * both Rx Overflow Continue and Parity are DISABLED.
607 */
608 _dma_ctrlflags(di, DMA_CTRL_ROC | DMA_CTRL_PEN, 0);
609
Arend van Spriele81da652011-12-08 15:06:53 -0800610 DMA_TRACE("%s: %s flags 0x%x ntxd %d nrxd %d "
611 "rxbufsize %d rxextheadroom %d nrxpost %d rxoffset %d "
612 "txregbase %u rxregbase %u\n", name, "DMA64",
Joe Perches8505a7e2011-11-13 11:41:04 -0800613 di->dma.dmactrlflags, ntxd, nrxd, rxbufsize,
Arend van Spriele81da652011-12-08 15:06:53 -0800614 rxextheadroom, nrxpost, rxoffset, txregbase, rxregbase);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200615
616 /* make a private copy of our callers name */
617 strncpy(di->name, name, MAXNAMEL);
618 di->name[MAXNAMEL - 1] = '\0';
619
Arend van Spriel3b758a62011-12-12 15:15:09 -0800620 di->dmadev = core->dma_dev;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200621
622 /* save tunables */
623 di->ntxd = (u16) ntxd;
624 di->nrxd = (u16) nrxd;
625
626 /* the actual dma size doesn't include the extra headroom */
627 di->rxextrahdrroom =
628 (rxextheadroom == -1) ? BCMEXTRAHDROOM : rxextheadroom;
629 if (rxbufsize > BCMEXTRAHDROOM)
630 di->rxbufsize = (u16) (rxbufsize - di->rxextrahdrroom);
631 else
632 di->rxbufsize = (u16) rxbufsize;
633
634 di->nrxpost = (u16) nrxpost;
635 di->rxoffset = (u8) rxoffset;
636
637 /*
638 * figure out the DMA physical address offset for dd and data
639 * PCI/PCIE: they map silicon backplace address to zero
640 * based memory, need offset
641 * Other bus: use zero SI_BUS BIGENDIAN kludge: use sdram
642 * swapped region for data buffer, not descriptor
643 */
644 di->ddoffsetlow = 0;
645 di->dataoffsetlow = 0;
Hauke Mehrtensec5ab1d2012-06-30 15:16:18 +0200646 /* for pci bus, add offset */
647 if (sii->icbus->hosttype == BCMA_HOSTTYPE_PCI) {
648 /* add offset for pcie with DMA64 bus */
649 di->ddoffsetlow = 0;
650 di->ddoffsethigh = SI_PCIE_DMA_H32;
651 }
Arend van Spriel5b435de2011-10-05 13:19:03 +0200652 di->dataoffsetlow = di->ddoffsetlow;
653 di->dataoffsethigh = di->ddoffsethigh;
Hauke Mehrtensec5ab1d2012-06-30 15:16:18 +0200654
Arend van Spriel5b435de2011-10-05 13:19:03 +0200655 /* WAR64450 : DMACtl.Addr ext fields are not supported in SDIOD core. */
Hauke Mehrtens780b9c42012-06-30 15:16:12 +0200656 if ((core->id.id == BCMA_CORE_SDIO_DEV)
Arend van Spriel3b758a62011-12-12 15:15:09 -0800657 && ((rev > 0) && (rev <= 2)))
Rusty Russell3db1cd52011-12-19 13:56:45 +0000658 di->addrext = false;
Hauke Mehrtens780b9c42012-06-30 15:16:12 +0200659 else if ((core->id.id == BCMA_CORE_I2S) &&
Arend van Spriel3b758a62011-12-12 15:15:09 -0800660 ((rev == 0) || (rev == 1)))
Rusty Russell3db1cd52011-12-19 13:56:45 +0000661 di->addrext = false;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200662 else
663 di->addrext = _dma_isaddrext(di);
664
665 /* does the descriptor need to be aligned and if yes, on 4K/8K or not */
666 di->aligndesc_4k = _dma_descriptor_align(di);
667 if (di->aligndesc_4k) {
668 di->dmadesc_align = D64RINGALIGN_BITS;
669 if ((ntxd < D64MAXDD / 2) && (nrxd < D64MAXDD / 2))
670 /* for smaller dd table, HW relax alignment reqmnt */
671 di->dmadesc_align = D64RINGALIGN_BITS - 1;
672 } else {
673 di->dmadesc_align = 4; /* 16 byte alignment */
674 }
675
Joe Perches8505a7e2011-11-13 11:41:04 -0800676 DMA_NONE("DMA descriptor align_needed %d, align %d\n",
677 di->aligndesc_4k, di->dmadesc_align);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200678
679 /* allocate tx packet pointer vector */
680 if (ntxd) {
681 size = ntxd * sizeof(void *);
682 di->txp = kzalloc(size, GFP_ATOMIC);
683 if (di->txp == NULL)
684 goto fail;
685 }
686
687 /* allocate rx packet pointer vector */
688 if (nrxd) {
689 size = nrxd * sizeof(void *);
690 di->rxp = kzalloc(size, GFP_ATOMIC);
691 if (di->rxp == NULL)
692 goto fail;
693 }
694
695 /*
696 * allocate transmit descriptor ring, only need ntxd descriptors
697 * but it must be aligned
698 */
699 if (ntxd) {
700 if (!_dma_alloc(di, DMA_TX))
701 goto fail;
702 }
703
704 /*
705 * allocate receive descriptor ring, only need nrxd descriptors
706 * but it must be aligned
707 */
708 if (nrxd) {
709 if (!_dma_alloc(di, DMA_RX))
710 goto fail;
711 }
712
713 if ((di->ddoffsetlow != 0) && !di->addrext) {
714 if (di->txdpa > SI_PCI_DMA_SZ) {
Joe Perches8505a7e2011-11-13 11:41:04 -0800715 DMA_ERROR("%s: txdpa 0x%x: addrext not supported\n",
716 di->name, (u32)di->txdpa);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200717 goto fail;
718 }
719 if (di->rxdpa > SI_PCI_DMA_SZ) {
Joe Perches8505a7e2011-11-13 11:41:04 -0800720 DMA_ERROR("%s: rxdpa 0x%x: addrext not supported\n",
721 di->name, (u32)di->rxdpa);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200722 goto fail;
723 }
724 }
725
Seth Forsheee041f652012-11-15 08:07:56 -0600726 /* Initialize AMPDU session */
727 brcms_c_ampdu_reset_session(&di->ampdu_session, wlc);
728
Joe Perches8505a7e2011-11-13 11:41:04 -0800729 DMA_TRACE("ddoffsetlow 0x%x ddoffsethigh 0x%x dataoffsetlow 0x%x dataoffsethigh 0x%x addrext %d\n",
730 di->ddoffsetlow, di->ddoffsethigh,
731 di->dataoffsetlow, di->dataoffsethigh,
732 di->addrext);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200733
734 return (struct dma_pub *) di;
735
736 fail:
737 dma_detach((struct dma_pub *)di);
738 return NULL;
739}
740
741static inline void
742dma64_dd_upd(struct dma_info *di, struct dma64desc *ddring,
743 dma_addr_t pa, uint outidx, u32 *flags, u32 bufcount)
744{
745 u32 ctrl2 = bufcount & D64_CTRL2_BC_MASK;
746
747 /* PCI bus with big(>1G) physical address, use address extension */
748 if ((di->dataoffsetlow == 0) || !(pa & PCI32ADDR_HIGH)) {
749 ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow);
750 ddring[outidx].addrhigh = cpu_to_le32(di->dataoffsethigh);
751 ddring[outidx].ctrl1 = cpu_to_le32(*flags);
752 ddring[outidx].ctrl2 = cpu_to_le32(ctrl2);
753 } else {
754 /* address extension for 32-bit PCI */
755 u32 ae;
756
757 ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT;
758 pa &= ~PCI32ADDR_HIGH;
759
760 ctrl2 |= (ae << D64_CTRL2_AE_SHIFT) & D64_CTRL2_AE;
761 ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow);
762 ddring[outidx].addrhigh = cpu_to_le32(di->dataoffsethigh);
763 ddring[outidx].ctrl1 = cpu_to_le32(*flags);
764 ddring[outidx].ctrl2 = cpu_to_le32(ctrl2);
765 }
766 if (di->dma.dmactrlflags & DMA_CTRL_PEN) {
767 if (dma64_dd_parity(&ddring[outidx]))
768 ddring[outidx].ctrl2 =
769 cpu_to_le32(ctrl2 | D64_CTRL2_PARITY);
770 }
771}
772
773/* !! may be called with core in reset */
774void dma_detach(struct dma_pub *pub)
775{
776 struct dma_info *di = (struct dma_info *)pub;
777
Joe Perches8505a7e2011-11-13 11:41:04 -0800778 DMA_TRACE("%s:\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200779
780 /* free dma descriptor rings */
781 if (di->txd64)
Arend van Spriel2e81b9b2011-12-08 15:06:52 -0800782 dma_free_coherent(di->dmadev, di->txdalloc,
783 ((s8 *)di->txd64 - di->txdalign),
784 (di->txdpaorig));
Arend van Spriel5b435de2011-10-05 13:19:03 +0200785 if (di->rxd64)
Arend van Spriel2e81b9b2011-12-08 15:06:52 -0800786 dma_free_coherent(di->dmadev, di->rxdalloc,
787 ((s8 *)di->rxd64 - di->rxdalign),
788 (di->rxdpaorig));
Arend van Spriel5b435de2011-10-05 13:19:03 +0200789
790 /* free packet pointer vectors */
791 kfree(di->txp);
792 kfree(di->rxp);
793
794 /* free our private info structure */
795 kfree(di);
796
797}
798
799/* initialize descriptor table base address */
800static void
801_dma_ddtable_init(struct dma_info *di, uint direction, dma_addr_t pa)
802{
803 if (!di->aligndesc_4k) {
804 if (direction == DMA_TX)
805 di->xmtptrbase = pa;
806 else
807 di->rcvptrbase = pa;
808 }
809
810 if ((di->ddoffsetlow == 0)
811 || !(pa & PCI32ADDR_HIGH)) {
812 if (direction == DMA_TX) {
Arend van Spriel3b758a62011-12-12 15:15:09 -0800813 bcma_write32(di->core, DMA64TXREGOFFS(di, addrlow),
Arend van Spriele81da652011-12-08 15:06:53 -0800814 pa + di->ddoffsetlow);
Arend van Spriel3b758a62011-12-12 15:15:09 -0800815 bcma_write32(di->core, DMA64TXREGOFFS(di, addrhigh),
Arend van Spriele81da652011-12-08 15:06:53 -0800816 di->ddoffsethigh);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200817 } else {
Arend van Spriel3b758a62011-12-12 15:15:09 -0800818 bcma_write32(di->core, DMA64RXREGOFFS(di, addrlow),
Arend van Spriele81da652011-12-08 15:06:53 -0800819 pa + di->ddoffsetlow);
Arend van Spriel3b758a62011-12-12 15:15:09 -0800820 bcma_write32(di->core, DMA64RXREGOFFS(di, addrhigh),
Arend van Spriele81da652011-12-08 15:06:53 -0800821 di->ddoffsethigh);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200822 }
823 } else {
824 /* DMA64 32bits address extension */
825 u32 ae;
826
827 /* shift the high bit(s) from pa to ae */
828 ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT;
829 pa &= ~PCI32ADDR_HIGH;
830
831 if (direction == DMA_TX) {
Arend van Spriel3b758a62011-12-12 15:15:09 -0800832 bcma_write32(di->core, DMA64TXREGOFFS(di, addrlow),
Arend van Spriele81da652011-12-08 15:06:53 -0800833 pa + di->ddoffsetlow);
Arend van Spriel3b758a62011-12-12 15:15:09 -0800834 bcma_write32(di->core, DMA64TXREGOFFS(di, addrhigh),
Arend van Spriele81da652011-12-08 15:06:53 -0800835 di->ddoffsethigh);
Arend van Spriel3b758a62011-12-12 15:15:09 -0800836 bcma_maskset32(di->core, DMA64TXREGOFFS(di, control),
Arend van Spriele81da652011-12-08 15:06:53 -0800837 D64_XC_AE, (ae << D64_XC_AE_SHIFT));
Arend van Spriel5b435de2011-10-05 13:19:03 +0200838 } else {
Arend van Spriel3b758a62011-12-12 15:15:09 -0800839 bcma_write32(di->core, DMA64RXREGOFFS(di, addrlow),
Arend van Spriele81da652011-12-08 15:06:53 -0800840 pa + di->ddoffsetlow);
Arend van Spriel3b758a62011-12-12 15:15:09 -0800841 bcma_write32(di->core, DMA64RXREGOFFS(di, addrhigh),
Arend van Spriele81da652011-12-08 15:06:53 -0800842 di->ddoffsethigh);
Arend van Spriel3b758a62011-12-12 15:15:09 -0800843 bcma_maskset32(di->core, DMA64RXREGOFFS(di, control),
Arend van Spriele81da652011-12-08 15:06:53 -0800844 D64_RC_AE, (ae << D64_RC_AE_SHIFT));
Arend van Spriel5b435de2011-10-05 13:19:03 +0200845 }
846 }
847}
848
849static void _dma_rxenable(struct dma_info *di)
850{
851 uint dmactrlflags = di->dma.dmactrlflags;
852 u32 control;
853
Joe Perches8505a7e2011-11-13 11:41:04 -0800854 DMA_TRACE("%s:\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200855
Arend van Spriel3b758a62011-12-12 15:15:09 -0800856 control = D64_RC_RE | (bcma_read32(di->core,
Arend van Spriele81da652011-12-08 15:06:53 -0800857 DMA64RXREGOFFS(di, control)) &
858 D64_RC_AE);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200859
860 if ((dmactrlflags & DMA_CTRL_PEN) == 0)
861 control |= D64_RC_PD;
862
863 if (dmactrlflags & DMA_CTRL_ROC)
864 control |= D64_RC_OC;
865
Arend van Spriel3b758a62011-12-12 15:15:09 -0800866 bcma_write32(di->core, DMA64RXREGOFFS(di, control),
Arend van Spriel5b435de2011-10-05 13:19:03 +0200867 ((di->rxoffset << D64_RC_RO_SHIFT) | control));
868}
869
870void dma_rxinit(struct dma_pub *pub)
871{
872 struct dma_info *di = (struct dma_info *)pub;
873
Joe Perches8505a7e2011-11-13 11:41:04 -0800874 DMA_TRACE("%s:\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200875
876 if (di->nrxd == 0)
877 return;
878
879 di->rxin = di->rxout = 0;
880
881 /* clear rx descriptor ring */
882 memset(di->rxd64, '\0', di->nrxd * sizeof(struct dma64desc));
883
884 /* DMA engine with out alignment requirement requires table to be inited
885 * before enabling the engine
886 */
887 if (!di->aligndesc_4k)
888 _dma_ddtable_init(di, DMA_RX, di->rxdpa);
889
890 _dma_rxenable(di);
891
892 if (di->aligndesc_4k)
893 _dma_ddtable_init(di, DMA_RX, di->rxdpa);
894}
895
896static struct sk_buff *dma64_getnextrxp(struct dma_info *di, bool forceall)
897{
898 uint i, curr;
899 struct sk_buff *rxp;
900 dma_addr_t pa;
901
902 i = di->rxin;
903
904 /* return if no packets posted */
905 if (i == di->rxout)
906 return NULL;
907
908 curr =
Arend van Spriel3b758a62011-12-12 15:15:09 -0800909 B2I(((bcma_read32(di->core,
Arend van Spriele81da652011-12-08 15:06:53 -0800910 DMA64RXREGOFFS(di, status0)) & D64_RS0_CD_MASK) -
Arend van Spriel5b435de2011-10-05 13:19:03 +0200911 di->rcvptrbase) & D64_RS0_CD_MASK, struct dma64desc);
912
913 /* ignore curr if forceall */
914 if (!forceall && (i == curr))
915 return NULL;
916
917 /* get the packet pointer that corresponds to the rx descriptor */
918 rxp = di->rxp[i];
919 di->rxp[i] = NULL;
920
921 pa = le32_to_cpu(di->rxd64[i].addrlow) - di->dataoffsetlow;
922
923 /* clear this packet from the descriptor ring */
Arend van Spriel2e81b9b2011-12-08 15:06:52 -0800924 dma_unmap_single(di->dmadev, pa, di->rxbufsize, DMA_FROM_DEVICE);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200925
926 di->rxd64[i].addrlow = cpu_to_le32(0xdeadbeef);
927 di->rxd64[i].addrhigh = cpu_to_le32(0xdeadbeef);
928
929 di->rxin = nextrxd(di, i);
930
931 return rxp;
932}
933
934static struct sk_buff *_dma_getnextrxp(struct dma_info *di, bool forceall)
935{
936 if (di->nrxd == 0)
937 return NULL;
938
939 return dma64_getnextrxp(di, forceall);
940}
941
942/*
943 * !! rx entry routine
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200944 * returns the number packages in the next frame, or 0 if there are no more
Arend van Spriel5b435de2011-10-05 13:19:03 +0200945 * if DMA_CTRL_RXMULTI is defined, DMA scattering(multiple buffers) is
946 * supported with pkts chain
947 * otherwise, it's treated as giant pkt and will be tossed.
948 * The DMA scattering starts with normal DMA header, followed by first
949 * buffer data. After it reaches the max size of buffer, the data continues
950 * in next DMA descriptor buffer WITHOUT DMA header
951 */
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200952int dma_rx(struct dma_pub *pub, struct sk_buff_head *skb_list)
Arend van Spriel5b435de2011-10-05 13:19:03 +0200953{
954 struct dma_info *di = (struct dma_info *)pub;
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200955 struct sk_buff_head dma_frames;
956 struct sk_buff *p, *next;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200957 uint len;
958 uint pkt_len;
959 int resid = 0;
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200960 int pktcnt = 1;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200961
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200962 skb_queue_head_init(&dma_frames);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200963 next_frame:
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200964 p = _dma_getnextrxp(di, false);
965 if (p == NULL)
966 return 0;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200967
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200968 len = le16_to_cpu(*(__le16 *) (p->data));
Joe Perches8505a7e2011-11-13 11:41:04 -0800969 DMA_TRACE("%s: dma_rx len %d\n", di->name, len);
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200970 dma_spin_for_len(len, p);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200971
972 /* set actual length */
973 pkt_len = min((di->rxoffset + len), di->rxbufsize);
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200974 __skb_trim(p, pkt_len);
975 skb_queue_tail(&dma_frames, p);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200976 resid = len - (di->rxbufsize - di->rxoffset);
977
978 /* check for single or multi-buffer rx */
979 if (resid > 0) {
Arend van Spriel5b435de2011-10-05 13:19:03 +0200980 while ((resid > 0) && (p = _dma_getnextrxp(di, false))) {
Arend van Spriel5b435de2011-10-05 13:19:03 +0200981 pkt_len = min_t(uint, resid, di->rxbufsize);
982 __skb_trim(p, pkt_len);
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200983 skb_queue_tail(&dma_frames, p);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200984 resid -= di->rxbufsize;
Arend van Spriel3fd172d2011-10-21 16:16:31 +0200985 pktcnt++;
Arend van Spriel5b435de2011-10-05 13:19:03 +0200986 }
987
Joe Perches8ae74652012-01-15 00:38:38 -0800988#ifdef DEBUG
Arend van Spriel5b435de2011-10-05 13:19:03 +0200989 if (resid > 0) {
990 uint cur;
991 cur =
Arend van Spriel3b758a62011-12-12 15:15:09 -0800992 B2I(((bcma_read32(di->core,
Arend van Spriele81da652011-12-08 15:06:53 -0800993 DMA64RXREGOFFS(di, status0)) &
994 D64_RS0_CD_MASK) - di->rcvptrbase) &
995 D64_RS0_CD_MASK, struct dma64desc);
Joe Perches8505a7e2011-11-13 11:41:04 -0800996 DMA_ERROR("rxin %d rxout %d, hw_curr %d\n",
Arend van Spriele81da652011-12-08 15:06:53 -0800997 di->rxin, di->rxout, cur);
Arend van Spriel5b435de2011-10-05 13:19:03 +0200998 }
Joe Perches8ae74652012-01-15 00:38:38 -0800999#endif /* DEBUG */
Arend van Spriel5b435de2011-10-05 13:19:03 +02001000
1001 if ((di->dma.dmactrlflags & DMA_CTRL_RXMULTI) == 0) {
Joe Perches8505a7e2011-11-13 11:41:04 -08001002 DMA_ERROR("%s: bad frame length (%d)\n",
1003 di->name, len);
Arend van Spriel3fd172d2011-10-21 16:16:31 +02001004 skb_queue_walk_safe(&dma_frames, p, next) {
1005 skb_unlink(p, &dma_frames);
1006 brcmu_pkt_buf_free_skb(p);
1007 }
Arend van Spriel5b435de2011-10-05 13:19:03 +02001008 di->dma.rxgiants++;
Arend van Spriel3fd172d2011-10-21 16:16:31 +02001009 pktcnt = 1;
Arend van Spriel5b435de2011-10-05 13:19:03 +02001010 goto next_frame;
1011 }
1012 }
1013
Arend van Spriel3fd172d2011-10-21 16:16:31 +02001014 skb_queue_splice_tail(&dma_frames, skb_list);
1015 return pktcnt;
Arend van Spriel5b435de2011-10-05 13:19:03 +02001016}
1017
1018static bool dma64_rxidle(struct dma_info *di)
1019{
Joe Perches8505a7e2011-11-13 11:41:04 -08001020 DMA_TRACE("%s:\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001021
1022 if (di->nrxd == 0)
1023 return true;
1024
Arend van Spriel3b758a62011-12-12 15:15:09 -08001025 return ((bcma_read32(di->core,
Arend van Spriele81da652011-12-08 15:06:53 -08001026 DMA64RXREGOFFS(di, status0)) & D64_RS0_CD_MASK) ==
Arend van Spriel3b758a62011-12-12 15:15:09 -08001027 (bcma_read32(di->core, DMA64RXREGOFFS(di, ptr)) &
Arend van Spriele81da652011-12-08 15:06:53 -08001028 D64_RS0_CD_MASK));
Arend van Spriel5b435de2011-10-05 13:19:03 +02001029}
1030
Seth Forsheee041f652012-11-15 08:07:56 -06001031static bool dma64_txidle(struct dma_info *di)
1032{
1033 if (di->ntxd == 0)
1034 return true;
1035
1036 return ((bcma_read32(di->core,
1037 DMA64TXREGOFFS(di, status0)) & D64_XS0_CD_MASK) ==
1038 (bcma_read32(di->core, DMA64TXREGOFFS(di, ptr)) &
1039 D64_XS0_CD_MASK));
1040}
1041
Arend van Spriel5b435de2011-10-05 13:19:03 +02001042/*
1043 * post receive buffers
1044 * return false is refill failed completely and ring is empty this will stall
1045 * the rx dma and user might want to call rxfill again asap. This unlikely
1046 * happens on memory-rich NIC, but often on memory-constrained dongle
1047 */
1048bool dma_rxfill(struct dma_pub *pub)
1049{
1050 struct dma_info *di = (struct dma_info *)pub;
1051 struct sk_buff *p;
1052 u16 rxin, rxout;
1053 u32 flags = 0;
1054 uint n;
1055 uint i;
1056 dma_addr_t pa;
1057 uint extra_offset = 0;
1058 bool ring_empty;
1059
1060 ring_empty = false;
1061
1062 /*
1063 * Determine how many receive buffers we're lacking
1064 * from the full complement, allocate, initialize,
1065 * and post them, then update the chip rx lastdscr.
1066 */
1067
1068 rxin = di->rxin;
1069 rxout = di->rxout;
1070
1071 n = di->nrxpost - nrxdactive(di, rxin, rxout);
1072
Joe Perches8505a7e2011-11-13 11:41:04 -08001073 DMA_TRACE("%s: post %d\n", di->name, n);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001074
1075 if (di->rxbufsize > BCMEXTRAHDROOM)
1076 extra_offset = di->rxextrahdrroom;
1077
1078 for (i = 0; i < n; i++) {
1079 /*
1080 * the di->rxbufsize doesn't include the extra headroom,
1081 * we need to add it to the size to be allocated
1082 */
1083 p = brcmu_pkt_buf_get_skb(di->rxbufsize + extra_offset);
1084
1085 if (p == NULL) {
Joe Perches8505a7e2011-11-13 11:41:04 -08001086 DMA_ERROR("%s: out of rxbufs\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001087 if (i == 0 && dma64_rxidle(di)) {
Joe Perches8505a7e2011-11-13 11:41:04 -08001088 DMA_ERROR("%s: ring is empty !\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001089 ring_empty = true;
1090 }
1091 di->dma.rxnobuf++;
1092 break;
1093 }
1094 /* reserve an extra headroom, if applicable */
1095 if (extra_offset)
1096 skb_pull(p, extra_offset);
1097
1098 /* Do a cached write instead of uncached write since DMA_MAP
1099 * will flush the cache.
1100 */
1101 *(u32 *) (p->data) = 0;
1102
Arend van Spriel2e81b9b2011-12-08 15:06:52 -08001103 pa = dma_map_single(di->dmadev, p->data, di->rxbufsize,
1104 DMA_FROM_DEVICE);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001105
1106 /* save the free packet pointer */
1107 di->rxp[rxout] = p;
1108
1109 /* reset flags for each descriptor */
1110 flags = 0;
1111 if (rxout == (di->nrxd - 1))
1112 flags = D64_CTRL1_EOT;
1113
1114 dma64_dd_upd(di, di->rxd64, pa, rxout, &flags,
1115 di->rxbufsize);
1116 rxout = nextrxd(di, rxout);
1117 }
1118
1119 di->rxout = rxout;
1120
1121 /* update the chip lastdscr pointer */
Arend van Spriel3b758a62011-12-12 15:15:09 -08001122 bcma_write32(di->core, DMA64RXREGOFFS(di, ptr),
Arend van Spriel5b435de2011-10-05 13:19:03 +02001123 di->rcvptrbase + I2B(rxout, struct dma64desc));
1124
1125 return ring_empty;
1126}
1127
1128void dma_rxreclaim(struct dma_pub *pub)
1129{
1130 struct dma_info *di = (struct dma_info *)pub;
1131 struct sk_buff *p;
1132
Joe Perches8505a7e2011-11-13 11:41:04 -08001133 DMA_TRACE("%s:\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001134
1135 while ((p = _dma_getnextrxp(di, true)))
1136 brcmu_pkt_buf_free_skb(p);
1137}
1138
1139void dma_counterreset(struct dma_pub *pub)
1140{
1141 /* reset all software counters */
1142 pub->rxgiants = 0;
1143 pub->rxnobuf = 0;
1144 pub->txnobuf = 0;
1145}
1146
1147/* get the address of the var in order to change later */
1148unsigned long dma_getvar(struct dma_pub *pub, const char *name)
1149{
1150 struct dma_info *di = (struct dma_info *)pub;
1151
1152 if (!strcmp(name, "&txavail"))
1153 return (unsigned long)&(di->dma.txavail);
1154 return 0;
1155}
1156
1157/* 64-bit DMA functions */
1158
1159void dma_txinit(struct dma_pub *pub)
1160{
1161 struct dma_info *di = (struct dma_info *)pub;
1162 u32 control = D64_XC_XE;
1163
Joe Perches8505a7e2011-11-13 11:41:04 -08001164 DMA_TRACE("%s:\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001165
1166 if (di->ntxd == 0)
1167 return;
1168
1169 di->txin = di->txout = 0;
1170 di->dma.txavail = di->ntxd - 1;
1171
1172 /* clear tx descriptor ring */
1173 memset(di->txd64, '\0', (di->ntxd * sizeof(struct dma64desc)));
1174
1175 /* DMA engine with out alignment requirement requires table to be inited
1176 * before enabling the engine
1177 */
1178 if (!di->aligndesc_4k)
1179 _dma_ddtable_init(di, DMA_TX, di->txdpa);
1180
1181 if ((di->dma.dmactrlflags & DMA_CTRL_PEN) == 0)
1182 control |= D64_XC_PD;
Arend van Spriel3b758a62011-12-12 15:15:09 -08001183 bcma_set32(di->core, DMA64TXREGOFFS(di, control), control);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001184
1185 /* DMA engine with alignment requirement requires table to be inited
1186 * before enabling the engine
1187 */
1188 if (di->aligndesc_4k)
1189 _dma_ddtable_init(di, DMA_TX, di->txdpa);
1190}
1191
1192void dma_txsuspend(struct dma_pub *pub)
1193{
1194 struct dma_info *di = (struct dma_info *)pub;
1195
Joe Perches8505a7e2011-11-13 11:41:04 -08001196 DMA_TRACE("%s:\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001197
1198 if (di->ntxd == 0)
1199 return;
1200
Arend van Spriel3b758a62011-12-12 15:15:09 -08001201 bcma_set32(di->core, DMA64TXREGOFFS(di, control), D64_XC_SE);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001202}
1203
1204void dma_txresume(struct dma_pub *pub)
1205{
1206 struct dma_info *di = (struct dma_info *)pub;
1207
Joe Perches8505a7e2011-11-13 11:41:04 -08001208 DMA_TRACE("%s:\n", di->name);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001209
1210 if (di->ntxd == 0)
1211 return;
1212
Arend van Spriel3b758a62011-12-12 15:15:09 -08001213 bcma_mask32(di->core, DMA64TXREGOFFS(di, control), ~D64_XC_SE);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001214}
1215
1216bool dma_txsuspended(struct dma_pub *pub)
1217{
1218 struct dma_info *di = (struct dma_info *)pub;
1219
1220 return (di->ntxd == 0) ||
Arend van Spriel3b758a62011-12-12 15:15:09 -08001221 ((bcma_read32(di->core,
Arend van Spriele81da652011-12-08 15:06:53 -08001222 DMA64TXREGOFFS(di, control)) & D64_XC_SE) ==
1223 D64_XC_SE);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001224}
1225
1226void dma_txreclaim(struct dma_pub *pub, enum txd_range range)
1227{
1228 struct dma_info *di = (struct dma_info *)pub;
1229 struct sk_buff *p;
1230
Joe Perches8505a7e2011-11-13 11:41:04 -08001231 DMA_TRACE("%s: %s\n",
1232 di->name,
1233 range == DMA_RANGE_ALL ? "all" :
1234 range == DMA_RANGE_TRANSMITTED ? "transmitted" :
1235 "transferred");
Arend van Spriel5b435de2011-10-05 13:19:03 +02001236
1237 if (di->txin == di->txout)
1238 return;
1239
1240 while ((p = dma_getnexttxp(pub, range))) {
1241 /* For unframed data, we don't have any packets to free */
1242 if (!(di->dma.dmactrlflags & DMA_CTRL_UNFRAMED))
1243 brcmu_pkt_buf_free_skb(p);
1244 }
1245}
1246
1247bool dma_txreset(struct dma_pub *pub)
1248{
1249 struct dma_info *di = (struct dma_info *)pub;
1250 u32 status;
1251
1252 if (di->ntxd == 0)
1253 return true;
1254
1255 /* suspend tx DMA first */
Arend van Spriel3b758a62011-12-12 15:15:09 -08001256 bcma_write32(di->core, DMA64TXREGOFFS(di, control), D64_XC_SE);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001257 SPINWAIT(((status =
Arend van Spriel3b758a62011-12-12 15:15:09 -08001258 (bcma_read32(di->core, DMA64TXREGOFFS(di, status0)) &
Arend van Spriele81da652011-12-08 15:06:53 -08001259 D64_XS0_XS_MASK)) != D64_XS0_XS_DISABLED) &&
1260 (status != D64_XS0_XS_IDLE) && (status != D64_XS0_XS_STOPPED),
1261 10000);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001262
Arend van Spriel3b758a62011-12-12 15:15:09 -08001263 bcma_write32(di->core, DMA64TXREGOFFS(di, control), 0);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001264 SPINWAIT(((status =
Arend van Spriel3b758a62011-12-12 15:15:09 -08001265 (bcma_read32(di->core, DMA64TXREGOFFS(di, status0)) &
Arend van Spriele81da652011-12-08 15:06:53 -08001266 D64_XS0_XS_MASK)) != D64_XS0_XS_DISABLED), 10000);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001267
1268 /* wait for the last transaction to complete */
1269 udelay(300);
1270
1271 return status == D64_XS0_XS_DISABLED;
1272}
1273
1274bool dma_rxreset(struct dma_pub *pub)
1275{
1276 struct dma_info *di = (struct dma_info *)pub;
1277 u32 status;
1278
1279 if (di->nrxd == 0)
1280 return true;
1281
Arend van Spriel3b758a62011-12-12 15:15:09 -08001282 bcma_write32(di->core, DMA64RXREGOFFS(di, control), 0);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001283 SPINWAIT(((status =
Arend van Spriel3b758a62011-12-12 15:15:09 -08001284 (bcma_read32(di->core, DMA64RXREGOFFS(di, status0)) &
Arend van Spriele81da652011-12-08 15:06:53 -08001285 D64_RS0_RS_MASK)) != D64_RS0_RS_DISABLED), 10000);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001286
1287 return status == D64_RS0_RS_DISABLED;
1288}
1289
Seth Forsheee041f652012-11-15 08:07:56 -06001290static void dma_txenq(struct dma_info *di, struct sk_buff *p)
Seth Forshee05f8a612012-11-15 08:07:53 -06001291{
Arend van Spriel5b435de2011-10-05 13:19:03 +02001292 unsigned char *data;
1293 uint len;
1294 u16 txout;
1295 u32 flags = 0;
1296 dma_addr_t pa;
1297
Arend van Spriel5b435de2011-10-05 13:19:03 +02001298 txout = di->txout;
1299
Seth Forsheee041f652012-11-15 08:07:56 -06001300 if (WARN_ON(nexttxd(di, txout) == di->txin))
1301 return;
1302
Arend van Spriel5b435de2011-10-05 13:19:03 +02001303 /*
Arend van Spriel30307942011-11-22 17:21:37 -08001304 * obtain and initialize transmit descriptor entry.
Arend van Spriel5b435de2011-10-05 13:19:03 +02001305 */
Arend van Spriel30307942011-11-22 17:21:37 -08001306 data = p->data;
1307 len = p->len;
Arend van Spriel5b435de2011-10-05 13:19:03 +02001308
Arend van Spriel30307942011-11-22 17:21:37 -08001309 /* get physical address of buffer start */
Arend van Spriel2e81b9b2011-12-08 15:06:52 -08001310 pa = dma_map_single(di->dmadev, data, len, DMA_TO_DEVICE);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001311
Arend van Spriel30307942011-11-22 17:21:37 -08001312 /* With a DMA segment list, Descriptor table is filled
1313 * using the segment list instead of looping over
1314 * buffers in multi-chain DMA. Therefore, EOF for SGLIST
1315 * is when end of segment list is reached.
1316 */
1317 flags = D64_CTRL1_SOF | D64_CTRL1_IOC | D64_CTRL1_EOF;
1318 if (txout == (di->ntxd - 1))
1319 flags |= D64_CTRL1_EOT;
Arend van Spriel5b435de2011-10-05 13:19:03 +02001320
Arend van Spriel30307942011-11-22 17:21:37 -08001321 dma64_dd_upd(di, di->txd64, pa, txout, &flags, len);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001322
Arend van Spriel30307942011-11-22 17:21:37 -08001323 txout = nexttxd(di, txout);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001324
1325 /* save the packet */
Arend van Spriel30307942011-11-22 17:21:37 -08001326 di->txp[prevtxd(di, txout)] = p;
Arend van Spriel5b435de2011-10-05 13:19:03 +02001327
1328 /* bump the tx descriptor index */
1329 di->txout = txout;
Seth Forsheee041f652012-11-15 08:07:56 -06001330}
Arend van Spriel5b435de2011-10-05 13:19:03 +02001331
Seth Forsheee041f652012-11-15 08:07:56 -06001332static void ampdu_finalize(struct dma_info *di)
1333{
1334 struct brcms_ampdu_session *session = &di->ampdu_session;
1335 struct sk_buff *p;
1336
1337 if (WARN_ON(skb_queue_empty(&session->skb_list)))
1338 return;
1339
1340 brcms_c_ampdu_finalize(session);
1341
1342 while (!skb_queue_empty(&session->skb_list)) {
1343 p = skb_dequeue(&session->skb_list);
1344 dma_txenq(di, p);
1345 }
1346
1347 bcma_write32(di->core, DMA64TXREGOFFS(di, ptr),
1348 di->xmtptrbase + I2B(di->txout, struct dma64desc));
1349 brcms_c_ampdu_reset_session(session, session->wlc);
1350}
1351
1352static void prep_ampdu_frame(struct dma_info *di, struct sk_buff *p)
1353{
1354 struct brcms_ampdu_session *session = &di->ampdu_session;
1355 int ret;
1356
1357 ret = brcms_c_ampdu_add_frame(session, p);
1358 if (ret == -ENOSPC) {
1359 /*
1360 * AMPDU cannot accomodate this frame. Close out the in-
1361 * progress AMPDU session and start a new one.
1362 */
1363 ampdu_finalize(di);
1364 ret = brcms_c_ampdu_add_frame(session, p);
1365 }
1366
1367 WARN_ON(ret);
1368}
1369
1370/* Update count of available tx descriptors based on current DMA state */
1371static void dma_update_txavail(struct dma_info *di)
1372{
1373 /*
1374 * Available space is number of descriptors less the number of
1375 * active descriptors and the number of queued AMPDU frames.
1376 */
1377 di->dma.txavail = di->ntxd - ntxdactive(di, di->txin, di->txout) -
1378 skb_queue_len(&di->ampdu_session.skb_list) - 1;
1379}
1380
1381/*
1382 * !! tx entry routine
1383 * WARNING: call must check the return value for error.
1384 * the error(toss frames) could be fatal and cause many subsequent hard
1385 * to debug problems
1386 */
1387int dma_txfast(struct brcms_c_info *wlc, struct dma_pub *pub,
1388 struct sk_buff *p)
1389{
1390 struct dma_info *di = (struct dma_info *)pub;
1391 struct brcms_ampdu_session *session = &di->ampdu_session;
1392 struct ieee80211_tx_info *tx_info;
1393 bool is_ampdu;
1394
1395 DMA_TRACE("%s:\n", di->name);
1396
1397 /* no use to transmit a zero length packet */
1398 if (p->len == 0)
1399 return 0;
1400
1401 /* return nonzero if out of tx descriptors */
1402 if (di->dma.txavail == 0 || nexttxd(di, di->txout) == di->txin)
1403 goto outoftxd;
1404
1405 tx_info = IEEE80211_SKB_CB(p);
1406 is_ampdu = tx_info->flags & IEEE80211_TX_CTL_AMPDU;
1407 if (is_ampdu)
1408 prep_ampdu_frame(di, p);
1409 else
1410 dma_txenq(di, p);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001411
1412 /* tx flow control */
Seth Forshee05f8a612012-11-15 08:07:53 -06001413 dma_update_txavail(di);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001414
Seth Forsheee041f652012-11-15 08:07:56 -06001415 /* kick the chip */
1416 if (is_ampdu) {
1417 /*
1418 * Start sending data if we've got a full AMPDU, there's
1419 * no more space in the DMA ring, or the ring isn't
1420 * currently transmitting.
1421 */
1422 if (skb_queue_len(&session->skb_list) == session->max_ampdu_frames ||
1423 di->dma.txavail == 0 || dma64_txidle(di))
1424 ampdu_finalize(di);
1425 } else {
1426 bcma_write32(di->core, DMA64TXREGOFFS(di, ptr),
1427 di->xmtptrbase + I2B(di->txout, struct dma64desc));
1428 }
1429
Arend van Spriel5b435de2011-10-05 13:19:03 +02001430 return 0;
1431
1432 outoftxd:
Joe Perches8505a7e2011-11-13 11:41:04 -08001433 DMA_ERROR("%s: out of txds !!!\n", di->name);
Arend van Spriel30307942011-11-22 17:21:37 -08001434 brcmu_pkt_buf_free_skb(p);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001435 di->dma.txavail = 0;
1436 di->dma.txnobuf++;
Seth Forsheee041f652012-11-15 08:07:56 -06001437 return -ENOSPC;
1438}
1439
1440void dma_txflush(struct dma_pub *pub)
1441{
1442 struct dma_info *di = (struct dma_info *)pub;
1443 struct brcms_ampdu_session *session = &di->ampdu_session;
1444
1445 if (!skb_queue_empty(&session->skb_list))
1446 ampdu_finalize(di);
1447}
1448
1449int dma_txpending(struct dma_pub *pub)
1450{
1451 struct dma_info *di = (struct dma_info *)pub;
1452 return ntxdactive(di, di->txin, di->txout);
1453}
1454
1455/*
1456 * If we have an active AMPDU session and are not transmitting,
1457 * this function will force tx to start.
1458 */
1459void dma_kick_tx(struct dma_pub *pub)
1460{
1461 struct dma_info *di = (struct dma_info *)pub;
1462 struct brcms_ampdu_session *session = &di->ampdu_session;
1463
1464 if (!skb_queue_empty(&session->skb_list) && dma64_txidle(di))
1465 ampdu_finalize(di);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001466}
1467
1468/*
1469 * Reclaim next completed txd (txds if using chained buffers) in the range
1470 * specified and return associated packet.
1471 * If range is DMA_RANGE_TRANSMITTED, reclaim descriptors that have be
1472 * transmitted as noted by the hardware "CurrDescr" pointer.
1473 * If range is DMA_RANGE_TRANSFERED, reclaim descriptors that have be
1474 * transferred by the DMA as noted by the hardware "ActiveDescr" pointer.
1475 * If range is DMA_RANGE_ALL, reclaim all txd(s) posted to the ring and
1476 * return associated packet regardless of the value of hardware pointers.
1477 */
1478struct sk_buff *dma_getnexttxp(struct dma_pub *pub, enum txd_range range)
1479{
1480 struct dma_info *di = (struct dma_info *)pub;
1481 u16 start, end, i;
1482 u16 active_desc;
1483 struct sk_buff *txp;
1484
Joe Perches8505a7e2011-11-13 11:41:04 -08001485 DMA_TRACE("%s: %s\n",
1486 di->name,
1487 range == DMA_RANGE_ALL ? "all" :
1488 range == DMA_RANGE_TRANSMITTED ? "transmitted" :
1489 "transferred");
Arend van Spriel5b435de2011-10-05 13:19:03 +02001490
1491 if (di->ntxd == 0)
1492 return NULL;
1493
1494 txp = NULL;
1495
1496 start = di->txin;
1497 if (range == DMA_RANGE_ALL)
1498 end = di->txout;
1499 else {
Arend van Spriel3b758a62011-12-12 15:15:09 -08001500 end = (u16) (B2I(((bcma_read32(di->core,
Arend van Spriele81da652011-12-08 15:06:53 -08001501 DMA64TXREGOFFS(di, status0)) &
1502 D64_XS0_CD_MASK) - di->xmtptrbase) &
1503 D64_XS0_CD_MASK, struct dma64desc));
Arend van Spriel5b435de2011-10-05 13:19:03 +02001504
1505 if (range == DMA_RANGE_TRANSFERED) {
1506 active_desc =
Arend van Spriel3b758a62011-12-12 15:15:09 -08001507 (u16)(bcma_read32(di->core,
Arend van Spriele81da652011-12-08 15:06:53 -08001508 DMA64TXREGOFFS(di, status1)) &
Arend van Spriel5b435de2011-10-05 13:19:03 +02001509 D64_XS1_AD_MASK);
1510 active_desc =
1511 (active_desc - di->xmtptrbase) & D64_XS0_CD_MASK;
1512 active_desc = B2I(active_desc, struct dma64desc);
1513 if (end != active_desc)
1514 end = prevtxd(di, active_desc);
1515 }
1516 }
1517
1518 if ((start == 0) && (end > di->txout))
1519 goto bogus;
1520
1521 for (i = start; i != end && !txp; i = nexttxd(di, i)) {
1522 dma_addr_t pa;
1523 uint size;
1524
1525 pa = le32_to_cpu(di->txd64[i].addrlow) - di->dataoffsetlow;
1526
1527 size =
1528 (le32_to_cpu(di->txd64[i].ctrl2) &
1529 D64_CTRL2_BC_MASK);
1530
1531 di->txd64[i].addrlow = cpu_to_le32(0xdeadbeef);
1532 di->txd64[i].addrhigh = cpu_to_le32(0xdeadbeef);
1533
1534 txp = di->txp[i];
1535 di->txp[i] = NULL;
1536
Arend van Spriel2e81b9b2011-12-08 15:06:52 -08001537 dma_unmap_single(di->dmadev, pa, size, DMA_TO_DEVICE);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001538 }
1539
1540 di->txin = i;
1541
1542 /* tx flow control */
Seth Forshee05f8a612012-11-15 08:07:53 -06001543 dma_update_txavail(di);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001544
1545 return txp;
1546
1547 bogus:
Joe Perches8505a7e2011-11-13 11:41:04 -08001548 DMA_NONE("bogus curr: start %d end %d txout %d\n",
1549 start, end, di->txout);
Arend van Spriel5b435de2011-10-05 13:19:03 +02001550 return NULL;
1551}
1552
1553/*
1554 * Mac80211 initiated actions sometimes require packets in the DMA queue to be
1555 * modified. The modified portion of the packet is not under control of the DMA
1556 * engine. This function calls a caller-supplied function for each packet in
1557 * the caller specified dma chain.
1558 */
1559void dma_walk_packets(struct dma_pub *dmah, void (*callback_fnc)
1560 (void *pkt, void *arg_a), void *arg_a)
1561{
1562 struct dma_info *di = (struct dma_info *) dmah;
1563 uint i = di->txin;
1564 uint end = di->txout;
1565 struct sk_buff *skb;
1566 struct ieee80211_tx_info *tx_info;
1567
1568 while (i != end) {
Joe Perches2c208892012-06-04 12:44:17 +00001569 skb = di->txp[i];
Arend van Spriel5b435de2011-10-05 13:19:03 +02001570 if (skb != NULL) {
1571 tx_info = (struct ieee80211_tx_info *)skb->cb;
1572 (callback_fnc)(tx_info, arg_a);
1573 }
1574 i = nexttxd(di, i);
1575 }
1576}