2024-09-09 08:52:07 +00:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2010 Broadcom Corporation
|
|
|
|
*
|
|
|
|
* Permission to use, copy, modify, and/or distribute this software for any
|
|
|
|
* purpose with or without fee is hereby granted, provided that the above
|
|
|
|
* copyright notice and this permission notice appear in all copies.
|
|
|
|
*
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
|
|
|
|
* WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
|
|
|
|
* MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
|
|
|
|
* SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
|
|
|
|
* WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
|
|
|
|
* OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
|
|
|
|
* CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef _BRCM_DMA_H_
|
|
|
|
#define _BRCM_DMA_H_
|
|
|
|
|
|
|
|
#include <linux/delay.h>
|
|
|
|
#include <linux/skbuff.h>
|
|
|
|
#include "types.h" /* forward structure declarations */
|
|
|
|
|
|
|
|
/* map/unmap direction */
|
|
|
|
#define DMA_TX 1 /* TX direction for DMA */
|
|
|
|
#define DMA_RX 2 /* RX direction for DMA */
|
|
|
|
|
|
|
|
/* DMA structure:
|
|
|
|
* support two DMA engines: 32 bits address or 64 bit addressing
|
|
|
|
* basic DMA register set is per channel(transmit or receive)
|
|
|
|
* a pair of channels is defined for convenience
|
|
|
|
*/
|
|
|
|
|
|
|
|
/* 32 bits addressing */
|
|
|
|
|
|
|
|
struct dma32diag { /* diag access */
|
|
|
|
u32 fifoaddr; /* diag address */
|
|
|
|
u32 fifodatalow; /* low 32bits of data */
|
|
|
|
u32 fifodatahigh; /* high 32bits of data */
|
|
|
|
u32 pad; /* reserved */
|
|
|
|
};
|
|
|
|
|
|
|
|
/* 64 bits addressing */
|
|
|
|
|
|
|
|
/* dma registers per channel(xmt or rcv) */
|
|
|
|
struct dma64regs {
|
|
|
|
u32 control; /* enable, et al */
|
|
|
|
u32 ptr; /* last descriptor posted to chip */
|
|
|
|
u32 addrlow; /* desc ring base address low 32-bits (8K aligned) */
|
|
|
|
u32 addrhigh; /* desc ring base address bits 63:32 (8K aligned) */
|
|
|
|
u32 status0; /* current descriptor, xmt state */
|
|
|
|
u32 status1; /* active descriptor, xmt error */
|
|
|
|
};
|
|
|
|
|
|
|
|
/* range param for dma_getnexttxp() and dma_txreclaim */
|
|
|
|
enum txd_range {
|
|
|
|
DMA_RANGE_ALL = 1,
|
|
|
|
DMA_RANGE_TRANSMITTED,
|
|
|
|
DMA_RANGE_TRANSFERED
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Exported data structure (read-only)
|
|
|
|
*/
|
|
|
|
/* export structure */
|
|
|
|
struct dma_pub {
|
|
|
|
uint txavail; /* # free tx descriptors */
|
|
|
|
uint dmactrlflags; /* dma control flags */
|
|
|
|
|
|
|
|
/* rx error counters */
|
|
|
|
uint rxgiants; /* rx giant frames */
|
|
|
|
uint rxnobuf; /* rx out of dma descriptors */
|
|
|
|
/* tx error counters */
|
|
|
|
uint txnobuf; /* tx out of dma descriptors */
|
|
|
|
};
|
|
|
|
|
|
|
|
extern struct dma_pub *dma_attach(char *name, struct si_pub *sih,
|
2024-09-09 08:57:42 +00:00
|
|
|
void __iomem *dmaregstx, void __iomem *dmaregsrx,
|
|
|
|
uint ntxd, uint nrxd,
|
|
|
|
uint rxbufsize, int rxextheadroom,
|
|
|
|
uint nrxpost, uint rxoffset, uint *msg_level);
|
2024-09-09 08:52:07 +00:00
|
|
|
|
|
|
|
void dma_rxinit(struct dma_pub *pub);
|
|
|
|
int dma_rx(struct dma_pub *pub, struct sk_buff_head *skb_list);
|
|
|
|
bool dma_rxfill(struct dma_pub *pub);
|
|
|
|
bool dma_rxreset(struct dma_pub *pub);
|
|
|
|
bool dma_txreset(struct dma_pub *pub);
|
|
|
|
void dma_txinit(struct dma_pub *pub);
|
|
|
|
int dma_txfast(struct dma_pub *pub, struct sk_buff *p0, bool commit);
|
|
|
|
void dma_txsuspend(struct dma_pub *pub);
|
|
|
|
bool dma_txsuspended(struct dma_pub *pub);
|
|
|
|
void dma_txresume(struct dma_pub *pub);
|
|
|
|
void dma_txreclaim(struct dma_pub *pub, enum txd_range range);
|
|
|
|
void dma_rxreclaim(struct dma_pub *pub);
|
|
|
|
void dma_detach(struct dma_pub *pub);
|
|
|
|
unsigned long dma_getvar(struct dma_pub *pub, const char *name);
|
|
|
|
struct sk_buff *dma_getnexttxp(struct dma_pub *pub, enum txd_range range);
|
|
|
|
void dma_counterreset(struct dma_pub *pub);
|
|
|
|
|
|
|
|
void dma_walk_packets(struct dma_pub *dmah, void (*callback_fnc)
|
|
|
|
(void *pkt, void *arg_a), void *arg_a);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* DMA(Bug) on bcm47xx chips seems to declare that the packet is ready, but
|
|
|
|
* the packet length is not updated yet (by DMA) on the expected time.
|
|
|
|
* Workaround is to hold processor till DMA updates the length, and stay off
|
|
|
|
* the bus to allow DMA update the length in buffer
|
|
|
|
*/
|
|
|
|
static inline void dma_spin_for_len(uint len, struct sk_buff *head)
|
|
|
|
{
|
|
|
|
#if defined(CONFIG_BCM47XX)
|
|
|
|
if (!len) {
|
|
|
|
while (!(len = *(u16 *) KSEG1ADDR(head->data)))
|
|
|
|
udelay(1);
|
|
|
|
|
|
|
|
*(u16 *) (head->data) = cpu_to_le16((u16) len);
|
|
|
|
}
|
|
|
|
#endif /* defined(CONFIG_BCM47XX) */
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* _BRCM_DMA_H_ */
|