| Index: chromeos/drivers/ath6kl/htc2/AR6000/ar6k.c
|
| diff --git a/chromeos/drivers/ath6kl/htc2/AR6000/ar6k.c b/chromeos/drivers/ath6kl/htc2/AR6000/ar6k.c
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..48f7c09b06e98660fe1fd50782a43b89dc08a9cd
|
| --- /dev/null
|
| +++ b/chromeos/drivers/ath6kl/htc2/AR6000/ar6k.c
|
| @@ -0,0 +1,1393 @@
|
| +//------------------------------------------------------------------------------
|
| +// <copyright file="ar6k.c" company="Atheros">
|
| +// Copyright (c) 2007-2008 Atheros Corporation. All rights reserved.
|
| +//
|
| +// This program is free software; you can redistribute it and/or modify
|
| +// it under the terms of the GNU General Public License version 2 as
|
| +// published by the Free Software Foundation;
|
| +//
|
| +// Software distributed under the License is distributed on an "AS
|
| +// IS" basis, WITHOUT WARRANTY OF ANY KIND, either express or
|
| +// implied. See the License for the specific language governing
|
| +// rights and limitations under the License.
|
| +//
|
| +//
|
| +//------------------------------------------------------------------------------
|
| +//==============================================================================
|
| +// AR6K device layer that handles register level I/O
|
| +//
|
| +// Author(s): ="Atheros"
|
| +//==============================================================================
|
| +#include "a_config.h"
|
| +#include "athdefs.h"
|
| +#include "a_types.h"
|
| +#include "AR6002/hw2.0/hw/mbox_host_reg.h"
|
| +#include "a_osapi.h"
|
| +#include "../htc_debug.h"
|
| +#include "hif.h"
|
| +#include "htc_packet.h"
|
| +#include "ar6k.h"
|
| +
|
| +#define MAILBOX_FOR_BLOCK_SIZE 1
|
| +
|
| +A_STATUS DevEnableInterrupts(AR6K_DEVICE *pDev);
|
| +A_STATUS DevDisableInterrupts(AR6K_DEVICE *pDev);
|
| +
|
| +static void DevCleanupVirtualScatterSupport(AR6K_DEVICE *pDev);
|
| +
|
| +void AR6KFreeIOPacket(AR6K_DEVICE *pDev, HTC_PACKET *pPacket)
|
| +{
|
| + LOCK_AR6K(pDev);
|
| + HTC_PACKET_ENQUEUE(&pDev->RegisterIOList,pPacket);
|
| + UNLOCK_AR6K(pDev);
|
| +}
|
| +
|
| +HTC_PACKET *AR6KAllocIOPacket(AR6K_DEVICE *pDev)
|
| +{
|
| + HTC_PACKET *pPacket;
|
| +
|
| + LOCK_AR6K(pDev);
|
| + pPacket = HTC_PACKET_DEQUEUE(&pDev->RegisterIOList);
|
| + UNLOCK_AR6K(pDev);
|
| +
|
| + return pPacket;
|
| +}
|
| +
|
| +void DevCleanup(AR6K_DEVICE *pDev)
|
| +{
|
| + DevCleanupGMbox(pDev);
|
| +
|
| + if (pDev->HifAttached) {
|
| + HIFDetachHTC(pDev->HIFDevice);
|
| + pDev->HifAttached = FALSE;
|
| + }
|
| +
|
| + DevCleanupVirtualScatterSupport(pDev);
|
| +
|
| + if (A_IS_MUTEX_VALID(&pDev->Lock)) {
|
| + A_MUTEX_DELETE(&pDev->Lock);
|
| + }
|
| +}
|
| +
|
| +A_STATUS DevSetup(AR6K_DEVICE *pDev)
|
| +{
|
| + A_UINT32 blocksizes[AR6K_MAILBOXES];
|
| + A_STATUS status = A_OK;
|
| + int i;
|
| + HTC_CALLBACKS htcCallbacks;
|
| +
|
| + do {
|
| +
|
| + DL_LIST_INIT(&pDev->ScatterReqHead);
|
| + /* initialize our free list of IO packets */
|
| + INIT_HTC_PACKET_QUEUE(&pDev->RegisterIOList);
|
| + A_MUTEX_INIT(&pDev->Lock);
|
| +
|
| + A_MEMZERO(&htcCallbacks, sizeof(HTC_CALLBACKS));
|
| + /* the device layer handles these */
|
| + htcCallbacks.rwCompletionHandler = DevRWCompletionHandler;
|
| + htcCallbacks.dsrHandler = DevDsrHandler;
|
| + htcCallbacks.context = pDev;
|
| +
|
| + status = HIFAttachHTC(pDev->HIFDevice, &htcCallbacks);
|
| +
|
| + if (A_FAILED(status)) {
|
| + break;
|
| + }
|
| +
|
| + pDev->HifAttached = TRUE;
|
| +
|
| + /* get the addresses for all 4 mailboxes */
|
| + status = HIFConfigureDevice(pDev->HIFDevice, HIF_DEVICE_GET_MBOX_ADDR,
|
| + &pDev->MailBoxInfo, sizeof(pDev->MailBoxInfo));
|
| +
|
| + if (status != A_OK) {
|
| + A_ASSERT(FALSE);
|
| + break;
|
| + }
|
| +
|
| + /* carve up register I/O packets (these are for ASYNC register I/O ) */
|
| + for (i = 0; i < AR6K_MAX_REG_IO_BUFFERS; i++) {
|
| + HTC_PACKET *pIOPacket;
|
| + pIOPacket = &pDev->RegIOBuffers[i].HtcPacket;
|
| + SET_HTC_PACKET_INFO_RX_REFILL(pIOPacket,
|
| + pDev,
|
| + pDev->RegIOBuffers[i].Buffer,
|
| + AR6K_REG_IO_BUFFER_SIZE,
|
| + 0); /* don't care */
|
| + AR6KFreeIOPacket(pDev,pIOPacket);
|
| + }
|
| +
|
| + /* get the block sizes */
|
| + status = HIFConfigureDevice(pDev->HIFDevice, HIF_DEVICE_GET_MBOX_BLOCK_SIZE,
|
| + blocksizes, sizeof(blocksizes));
|
| +
|
| + if (status != A_OK) {
|
| + A_ASSERT(FALSE);
|
| + break;
|
| + }
|
| +
|
| + /* note: we actually get the block size of a mailbox other than 0, for SDIO the block
|
| + * size on mailbox 0 is artificially set to 1. So we use the block size that is set
|
| + * for the other 3 mailboxes */
|
| + pDev->BlockSize = blocksizes[MAILBOX_FOR_BLOCK_SIZE];
|
| + /* must be a power of 2 */
|
| + A_ASSERT((pDev->BlockSize & (pDev->BlockSize - 1)) == 0);
|
| +
|
| + /* assemble mask, used for padding to a block */
|
| + pDev->BlockMask = pDev->BlockSize - 1;
|
| +
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_TRC,("BlockSize: %d, MailboxAddress:0x%X \n",
|
| + pDev->BlockSize, pDev->MailBoxInfo.MboxAddresses[HTC_MAILBOX]));
|
| +
|
| + pDev->GetPendingEventsFunc = NULL;
|
| + /* see if the HIF layer implements the get pending events function */
|
| + HIFConfigureDevice(pDev->HIFDevice,
|
| + HIF_DEVICE_GET_PENDING_EVENTS_FUNC,
|
| + &pDev->GetPendingEventsFunc,
|
| + sizeof(pDev->GetPendingEventsFunc));
|
| +
|
| + /* assume we can process HIF interrupt events asynchronously */
|
| + pDev->HifIRQProcessingMode = HIF_DEVICE_IRQ_ASYNC_SYNC;
|
| +
|
| + /* see if the HIF layer overrides this assumption */
|
| + HIFConfigureDevice(pDev->HIFDevice,
|
| + HIF_DEVICE_GET_IRQ_PROC_MODE,
|
| + &pDev->HifIRQProcessingMode,
|
| + sizeof(pDev->HifIRQProcessingMode));
|
| +
|
| + switch (pDev->HifIRQProcessingMode) {
|
| + case HIF_DEVICE_IRQ_SYNC_ONLY:
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_WARN,("HIF Interrupt processing is SYNC ONLY\n"));
|
| + /* see if HIF layer wants HTC to yield */
|
| + HIFConfigureDevice(pDev->HIFDevice,
|
| + HIF_DEVICE_GET_IRQ_YIELD_PARAMS,
|
| + &pDev->HifIRQYieldParams,
|
| + sizeof(pDev->HifIRQYieldParams));
|
| +
|
| + if (pDev->HifIRQYieldParams.RecvPacketYieldCount > 0) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_WARN,
|
| + ("HIF requests that DSR yield per %d RECV packets \n",
|
| + pDev->HifIRQYieldParams.RecvPacketYieldCount));
|
| + pDev->DSRCanYield = TRUE;
|
| + }
|
| + break;
|
| + case HIF_DEVICE_IRQ_ASYNC_SYNC:
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_TRC,("HIF Interrupt processing is ASYNC and SYNC\n"));
|
| + break;
|
| + default:
|
| + A_ASSERT(FALSE);
|
| + }
|
| +
|
| + pDev->HifMaskUmaskRecvEvent = NULL;
|
| +
|
| + /* see if the HIF layer implements the mask/unmask recv events function */
|
| + HIFConfigureDevice(pDev->HIFDevice,
|
| + HIF_DEVICE_GET_RECV_EVENT_MASK_UNMASK_FUNC,
|
| + &pDev->HifMaskUmaskRecvEvent,
|
| + sizeof(pDev->HifMaskUmaskRecvEvent));
|
| +
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_TRC,("HIF special overrides : 0x%X , 0x%X\n",
|
| + (A_UINT32)pDev->GetPendingEventsFunc, (A_UINT32)pDev->HifMaskUmaskRecvEvent));
|
| +
|
| + status = DevDisableInterrupts(pDev);
|
| +
|
| + if (A_FAILED(status)) {
|
| + break;
|
| + }
|
| +
|
| + status = DevSetupGMbox(pDev);
|
| +
|
| + } while (FALSE);
|
| +
|
| + if (A_FAILED(status)) {
|
| + if (pDev->HifAttached) {
|
| + HIFDetachHTC(pDev->HIFDevice);
|
| + pDev->HifAttached = FALSE;
|
| + }
|
| + }
|
| +
|
| + return status;
|
| +
|
| +}
|
| +
|
| +A_STATUS DevEnableInterrupts(AR6K_DEVICE *pDev)
|
| +{
|
| + A_STATUS status;
|
| + AR6K_IRQ_ENABLE_REGISTERS regs;
|
| +
|
| + LOCK_AR6K(pDev);
|
| +
|
| + /* Enable all the interrupts except for the internal AR6000 CPU interrupt */
|
| + pDev->IrqEnableRegisters.int_status_enable = INT_STATUS_ENABLE_ERROR_SET(0x01) |
|
| + INT_STATUS_ENABLE_CPU_SET(0x01) |
|
| + INT_STATUS_ENABLE_COUNTER_SET(0x01);
|
| +
|
| + if (NULL == pDev->GetPendingEventsFunc) {
|
| + pDev->IrqEnableRegisters.int_status_enable |= INT_STATUS_ENABLE_MBOX_DATA_SET(0x01);
|
| + } else {
|
| + /* The HIF layer provided us with a pending events function which means that
|
| + * the detection of pending mbox messages is handled in the HIF layer.
|
| + * This is the case for the SPI2 interface.
|
| + * In the normal case we enable MBOX interrupts, for the case
|
| + * with HIFs that offer this mechanism, we keep these interrupts
|
| + * masked */
|
| + pDev->IrqEnableRegisters.int_status_enable &= ~INT_STATUS_ENABLE_MBOX_DATA_SET(0x01);
|
| + }
|
| +
|
| +
|
| + /* Set up the CPU Interrupt Status Register */
|
| + pDev->IrqEnableRegisters.cpu_int_status_enable = CPU_INT_STATUS_ENABLE_BIT_SET(0x00);
|
| +
|
| + /* Set up the Error Interrupt Status Register */
|
| + pDev->IrqEnableRegisters.error_status_enable =
|
| + ERROR_STATUS_ENABLE_RX_UNDERFLOW_SET(0x01) |
|
| + ERROR_STATUS_ENABLE_TX_OVERFLOW_SET(0x01);
|
| +
|
| + /* Set up the Counter Interrupt Status Register (only for debug interrupt to catch fatal errors) */
|
| + pDev->IrqEnableRegisters.counter_int_status_enable =
|
| + COUNTER_INT_STATUS_ENABLE_BIT_SET(AR6K_TARGET_DEBUG_INTR_MASK);
|
| +
|
| + /* copy into our temp area */
|
| + A_MEMCPY(®s,&pDev->IrqEnableRegisters,AR6K_IRQ_ENABLE_REGS_SIZE);
|
| +
|
| + UNLOCK_AR6K(pDev);
|
| +
|
| + /* always synchronous */
|
| + status = HIFReadWrite(pDev->HIFDevice,
|
| + INT_STATUS_ENABLE_ADDRESS,
|
| + ®s.int_status_enable,
|
| + AR6K_IRQ_ENABLE_REGS_SIZE,
|
| + HIF_WR_SYNC_BYTE_INC,
|
| + NULL);
|
| +
|
| + if (status != A_OK) {
|
| + /* Can't write it for some reason */
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR,
|
| + ("Failed to update interrupt control registers err: %d\n", status));
|
| +
|
| + }
|
| +
|
| + return status;
|
| +}
|
| +
|
| +A_STATUS DevDisableInterrupts(AR6K_DEVICE *pDev)
|
| +{
|
| + AR6K_IRQ_ENABLE_REGISTERS regs;
|
| +
|
| + LOCK_AR6K(pDev);
|
| + /* Disable all interrupts */
|
| + pDev->IrqEnableRegisters.int_status_enable = 0;
|
| + pDev->IrqEnableRegisters.cpu_int_status_enable = 0;
|
| + pDev->IrqEnableRegisters.error_status_enable = 0;
|
| + pDev->IrqEnableRegisters.counter_int_status_enable = 0;
|
| + /* copy into our temp area */
|
| + A_MEMCPY(®s,&pDev->IrqEnableRegisters,AR6K_IRQ_ENABLE_REGS_SIZE);
|
| +
|
| + UNLOCK_AR6K(pDev);
|
| +
|
| + /* always synchronous */
|
| + return HIFReadWrite(pDev->HIFDevice,
|
| + INT_STATUS_ENABLE_ADDRESS,
|
| + ®s.int_status_enable,
|
| + AR6K_IRQ_ENABLE_REGS_SIZE,
|
| + HIF_WR_SYNC_BYTE_INC,
|
| + NULL);
|
| +}
|
| +
|
| +/* enable device interrupts */
|
| +A_STATUS DevUnmaskInterrupts(AR6K_DEVICE *pDev)
|
| +{
|
| + /* for good measure, make sure interrupt are disabled before unmasking at the HIF
|
| + * layer.
|
| + * The rationale here is that between device insertion (where we clear the interrupts the first time)
|
| + * and when HTC is finally ready to handle interrupts, other software can perform target "soft" resets.
|
| + * The AR6K interrupt enables reset back to an "enabled" state when this happens.
|
| + * */
|
| + DevDisableInterrupts(pDev);
|
| +
|
| + /* Unmask the host controller interrupts */
|
| + HIFUnMaskInterrupt(pDev->HIFDevice);
|
| +
|
| + return DevEnableInterrupts(pDev);
|
| +}
|
| +
|
| +/* disable all device interrupts */
|
| +A_STATUS DevMaskInterrupts(AR6K_DEVICE *pDev)
|
| +{
|
| + /* mask the interrupt at the HIF layer, we don't want a stray interrupt taken while
|
| + * we zero out our shadow registers in DevDisableInterrupts()*/
|
| + HIFMaskInterrupt(pDev->HIFDevice);
|
| +
|
| + return DevDisableInterrupts(pDev);
|
| +}
|
| +
|
| +/* callback when our fetch to enable/disable completes */
|
| +static void DevDoEnableDisableRecvAsyncHandler(void *Context, HTC_PACKET *pPacket)
|
| +{
|
| + AR6K_DEVICE *pDev = (AR6K_DEVICE *)Context;
|
| +
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_IRQ,("+DevDoEnableDisableRecvAsyncHandler: (dev: 0x%X)\n", (A_UINT32)pDev));
|
| +
|
| + if (A_FAILED(pPacket->Status)) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR,
|
| + (" Failed to disable receiver, status:%d \n", pPacket->Status));
|
| + }
|
| + /* free this IO packet */
|
| + AR6KFreeIOPacket(pDev,pPacket);
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_IRQ,("-DevDoEnableDisableRecvAsyncHandler \n"));
|
| +}
|
| +
|
| +/* disable packet reception (used in case the host runs out of buffers)
|
| + * this is the "override" method when the HIF reports another methods to
|
| + * disable recv events */
|
| +static A_STATUS DevDoEnableDisableRecvOverride(AR6K_DEVICE *pDev, A_BOOL EnableRecv, A_BOOL AsyncMode)
|
| +{
|
| + A_STATUS status = A_OK;
|
| + HTC_PACKET *pIOPacket = NULL;
|
| +
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_TRC,("DevDoEnableDisableRecvOverride: Enable:%d Mode:%d\n",
|
| + EnableRecv,AsyncMode));
|
| +
|
| + do {
|
| +
|
| + if (AsyncMode) {
|
| +
|
| + pIOPacket = AR6KAllocIOPacket(pDev);
|
| +
|
| + if (NULL == pIOPacket) {
|
| + status = A_NO_MEMORY;
|
| + A_ASSERT(FALSE);
|
| + break;
|
| + }
|
| +
|
| + /* stick in our completion routine when the I/O operation completes */
|
| + pIOPacket->Completion = DevDoEnableDisableRecvAsyncHandler;
|
| + pIOPacket->pContext = pDev;
|
| +
|
| + /* call the HIF layer override and do this asynchronously */
|
| + status = pDev->HifMaskUmaskRecvEvent(pDev->HIFDevice,
|
| + EnableRecv ? HIF_UNMASK_RECV : HIF_MASK_RECV,
|
| + pIOPacket);
|
| + break;
|
| + }
|
| +
|
| + /* if we get here we are doing it synchronously */
|
| + status = pDev->HifMaskUmaskRecvEvent(pDev->HIFDevice,
|
| + EnableRecv ? HIF_UNMASK_RECV : HIF_MASK_RECV,
|
| + NULL);
|
| +
|
| + } while (FALSE);
|
| +
|
| + if (A_FAILED(status) && (pIOPacket != NULL)) {
|
| + AR6KFreeIOPacket(pDev,pIOPacket);
|
| + }
|
| +
|
| + return status;
|
| +}
|
| +
|
| +/* disable packet reception (used in case the host runs out of buffers)
|
| + * this is the "normal" method using the interrupt enable registers through
|
| + * the host I/F */
|
| +static A_STATUS DevDoEnableDisableRecvNormal(AR6K_DEVICE *pDev, A_BOOL EnableRecv, A_BOOL AsyncMode)
|
| +{
|
| + A_STATUS status = A_OK;
|
| + HTC_PACKET *pIOPacket = NULL;
|
| + AR6K_IRQ_ENABLE_REGISTERS regs;
|
| +
|
| + /* take the lock to protect interrupt enable shadows */
|
| + LOCK_AR6K(pDev);
|
| +
|
| + if (EnableRecv) {
|
| + pDev->IrqEnableRegisters.int_status_enable |= INT_STATUS_ENABLE_MBOX_DATA_SET(0x01);
|
| + } else {
|
| + pDev->IrqEnableRegisters.int_status_enable &= ~INT_STATUS_ENABLE_MBOX_DATA_SET(0x01);
|
| + }
|
| +
|
| + /* copy into our temp area */
|
| + A_MEMCPY(®s,&pDev->IrqEnableRegisters,AR6K_IRQ_ENABLE_REGS_SIZE);
|
| + UNLOCK_AR6K(pDev);
|
| +
|
| + do {
|
| +
|
| + if (AsyncMode) {
|
| +
|
| + pIOPacket = AR6KAllocIOPacket(pDev);
|
| +
|
| + if (NULL == pIOPacket) {
|
| + status = A_NO_MEMORY;
|
| + A_ASSERT(FALSE);
|
| + break;
|
| + }
|
| +
|
| + /* copy values to write to our async I/O buffer */
|
| + A_MEMCPY(pIOPacket->pBuffer,®s,AR6K_IRQ_ENABLE_REGS_SIZE);
|
| +
|
| + /* stick in our completion routine when the I/O operation completes */
|
| + pIOPacket->Completion = DevDoEnableDisableRecvAsyncHandler;
|
| + pIOPacket->pContext = pDev;
|
| +
|
| + /* write it out asynchronously */
|
| + HIFReadWrite(pDev->HIFDevice,
|
| + INT_STATUS_ENABLE_ADDRESS,
|
| + pIOPacket->pBuffer,
|
| + AR6K_IRQ_ENABLE_REGS_SIZE,
|
| + HIF_WR_ASYNC_BYTE_INC,
|
| + pIOPacket);
|
| + break;
|
| + }
|
| +
|
| + /* if we get here we are doing it synchronously */
|
| +
|
| + status = HIFReadWrite(pDev->HIFDevice,
|
| + INT_STATUS_ENABLE_ADDRESS,
|
| + ®s.int_status_enable,
|
| + AR6K_IRQ_ENABLE_REGS_SIZE,
|
| + HIF_WR_SYNC_BYTE_INC,
|
| + NULL);
|
| +
|
| + } while (FALSE);
|
| +
|
| + if (A_FAILED(status) && (pIOPacket != NULL)) {
|
| + AR6KFreeIOPacket(pDev,pIOPacket);
|
| + }
|
| +
|
| + return status;
|
| +}
|
| +
|
| +
|
| +A_STATUS DevStopRecv(AR6K_DEVICE *pDev, A_BOOL AsyncMode)
|
| +{
|
| + if (NULL == pDev->HifMaskUmaskRecvEvent) {
|
| + return DevDoEnableDisableRecvNormal(pDev,FALSE,AsyncMode);
|
| + } else {
|
| + return DevDoEnableDisableRecvOverride(pDev,FALSE,AsyncMode);
|
| + }
|
| +}
|
| +
|
| +A_STATUS DevEnableRecv(AR6K_DEVICE *pDev, A_BOOL AsyncMode)
|
| +{
|
| + if (NULL == pDev->HifMaskUmaskRecvEvent) {
|
| + return DevDoEnableDisableRecvNormal(pDev,TRUE,AsyncMode);
|
| + } else {
|
| + return DevDoEnableDisableRecvOverride(pDev,TRUE,AsyncMode);
|
| + }
|
| +}
|
| +
|
| +void DevDumpRegisters(AR6K_DEVICE *pDev,
|
| + AR6K_IRQ_PROC_REGISTERS *pIrqProcRegs,
|
| + AR6K_IRQ_ENABLE_REGISTERS *pIrqEnableRegs)
|
| +{
|
| +
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY, ("\n<------- Register Table -------->\n"));
|
| +
|
| + if (pIrqProcRegs != NULL) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("Host Int Status: 0x%x\n",pIrqProcRegs->host_int_status));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("CPU Int Status: 0x%x\n",pIrqProcRegs->cpu_int_status));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("Error Int Status: 0x%x\n",pIrqProcRegs->error_int_status));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("Counter Int Status: 0x%x\n",pIrqProcRegs->counter_int_status));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("Mbox Frame: 0x%x\n",pIrqProcRegs->mbox_frame));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("Rx Lookahead Valid: 0x%x\n",pIrqProcRegs->rx_lookahead_valid));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("Rx Lookahead 0: 0x%x\n",pIrqProcRegs->rx_lookahead[0]));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("Rx Lookahead 1: 0x%x\n",pIrqProcRegs->rx_lookahead[1]));
|
| +
|
| + if (pDev->MailBoxInfo.GMboxAddress != 0) {
|
| + /* if the target supports GMBOX hardware, dump some additional state */
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("GMBOX Host Int Status 2: 0x%x\n",pIrqProcRegs->host_int_status2));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("GMBOX RX Avail: 0x%x\n",pIrqProcRegs->gmbox_rx_avail));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("GMBOX lookahead alias 0: 0x%x\n",pIrqProcRegs->rx_gmbox_lookahead_alias[0]));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("GMBOX lookahead alias 1: 0x%x\n",pIrqProcRegs->rx_gmbox_lookahead_alias[1]));
|
| + }
|
| +
|
| + }
|
| +
|
| + if (pIrqEnableRegs != NULL) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("Int Status Enable: 0x%x\n",pIrqEnableRegs->int_status_enable));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("Counter Int Status Enable: 0x%x\n",pIrqEnableRegs->counter_int_status_enable));
|
| + }
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY, ("<------------------------------->\n"));
|
| +}
|
| +
|
| +
|
| +#define DEV_GET_VIRT_DMA_INFO(p) ((DEV_SCATTER_DMA_VIRTUAL_INFO *)((p)->HIFPrivate[0]))
|
| +
|
| +static HIF_SCATTER_REQ *DevAllocScatterReq(HIF_DEVICE *Context)
|
| +{
|
| + DL_LIST *pItem;
|
| + AR6K_DEVICE *pDev = (AR6K_DEVICE *)Context;
|
| + LOCK_AR6K(pDev);
|
| + pItem = DL_ListRemoveItemFromHead(&pDev->ScatterReqHead);
|
| + UNLOCK_AR6K(pDev);
|
| + if (pItem != NULL) {
|
| + return A_CONTAINING_STRUCT(pItem, HIF_SCATTER_REQ, ListLink);
|
| + }
|
| + return NULL;
|
| +}
|
| +
|
| +static void DevFreeScatterReq(HIF_DEVICE *Context, HIF_SCATTER_REQ *pReq)
|
| +{
|
| + AR6K_DEVICE *pDev = (AR6K_DEVICE *)Context;
|
| + LOCK_AR6K(pDev);
|
| + DL_ListInsertTail(&pDev->ScatterReqHead, &pReq->ListLink);
|
| + UNLOCK_AR6K(pDev);
|
| +}
|
| +
|
| +A_STATUS DevCopyScatterListToFromDMABuffer(HIF_SCATTER_REQ *pReq, A_BOOL FromDMA)
|
| +{
|
| + A_UINT8 *pDMABuffer = NULL;
|
| + int i, remaining;
|
| + A_UINT32 length;
|
| +
|
| + pDMABuffer = pReq->pScatterBounceBuffer;
|
| +
|
| + if (pDMABuffer == NULL) {
|
| + A_ASSERT(FALSE);
|
| + return A_EINVAL;
|
| + }
|
| +
|
| + remaining = (int)pReq->TotalLength;
|
| +
|
| + for (i = 0; i < pReq->ValidScatterEntries; i++) {
|
| +
|
| + length = min((int)pReq->ScatterList[i].Length, remaining);
|
| +
|
| + if (length != (int)pReq->ScatterList[i].Length) {
|
| + A_ASSERT(FALSE);
|
| + /* there is a problem with the scatter list */
|
| + return A_EINVAL;
|
| + }
|
| +
|
| + if (FromDMA) {
|
| + /* from DMA buffer */
|
| + A_MEMCPY(pReq->ScatterList[i].pBuffer, pDMABuffer , length);
|
| + } else {
|
| + /* to DMA buffer */
|
| + A_MEMCPY(pDMABuffer, pReq->ScatterList[i].pBuffer, length);
|
| + }
|
| +
|
| + pDMABuffer += length;
|
| + remaining -= length;
|
| + }
|
| +
|
| + return A_OK;
|
| +}
|
| +
|
| +static void DevReadWriteScatterAsyncHandler(void *Context, HTC_PACKET *pPacket)
|
| +{
|
| + AR6K_DEVICE *pDev = (AR6K_DEVICE *)Context;
|
| + HIF_SCATTER_REQ *pReq = (HIF_SCATTER_REQ *)pPacket->pPktContext;
|
| +
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_RECV,("+DevReadWriteScatterAsyncHandler: (dev: 0x%X)\n", (A_UINT32)pDev));
|
| +
|
| + pReq->CompletionStatus = pPacket->Status;
|
| +
|
| + AR6KFreeIOPacket(pDev,pPacket);
|
| +
|
| + pReq->CompletionRoutine(pReq);
|
| +
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_RECV,("-DevReadWriteScatterAsyncHandler \n"));
|
| +}
|
| +
|
| +static A_STATUS DevReadWriteScatter(HIF_DEVICE *Context, HIF_SCATTER_REQ *pReq)
|
| +{
|
| + AR6K_DEVICE *pDev = (AR6K_DEVICE *)Context;
|
| + A_STATUS status = A_OK;
|
| + HTC_PACKET *pIOPacket = NULL;
|
| + A_UINT32 request = pReq->Request;
|
| +
|
| + do {
|
| +
|
| + if (pReq->TotalLength > AR6K_MAX_TRANSFER_SIZE_PER_SCATTER) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR,
|
| + ("Invalid length: %d \n", pReq->TotalLength));
|
| + break;
|
| + }
|
| +
|
| + if (pReq->TotalLength == 0) {
|
| + A_ASSERT(FALSE);
|
| + break;
|
| + }
|
| +
|
| + if (request & HIF_ASYNCHRONOUS) {
|
| + /* use an I/O packet to carry this request */
|
| + pIOPacket = AR6KAllocIOPacket(pDev);
|
| + if (NULL == pIOPacket) {
|
| + status = A_NO_MEMORY;
|
| + break;
|
| + }
|
| +
|
| + /* save the request */
|
| + pIOPacket->pPktContext = pReq;
|
| + /* stick in our completion routine when the I/O operation completes */
|
| + pIOPacket->Completion = DevReadWriteScatterAsyncHandler;
|
| + pIOPacket->pContext = pDev;
|
| + }
|
| +
|
| + if (request & HIF_WRITE) {
|
| + /* in virtual DMA, we are issuing the requests through the legacy HIFReadWrite API
|
| + * this API will adjust the address automatically for the last byte to fall on the mailbox
|
| + * EOM. */
|
| +
|
| + /* if the address is an extended address, we can adjust the address here since the extended
|
| + * address will bypass the normal checks in legacy HIF layers */
|
| + if (pReq->Address == pDev->MailBoxInfo.MboxProp[HTC_MAILBOX].ExtendedAddress) {
|
| + pReq->Address += pDev->MailBoxInfo.MboxProp[HTC_MAILBOX].ExtendedSize - pReq->TotalLength;
|
| + }
|
| + }
|
| +
|
| + /* use legacy readwrite */
|
| + status = HIFReadWrite(pDev->HIFDevice,
|
| + pReq->Address,
|
| + DEV_GET_VIRT_DMA_INFO(pReq)->pVirtDmaBuffer,
|
| + pReq->TotalLength,
|
| + request,
|
| + (request & HIF_ASYNCHRONOUS) ? pIOPacket : NULL);
|
| +
|
| + } while (FALSE);
|
| +
|
| + if ((status != A_PENDING) && A_FAILED(status) && (request & HIF_ASYNCHRONOUS)) {
|
| + if (pIOPacket != NULL) {
|
| + AR6KFreeIOPacket(pDev,pIOPacket);
|
| + }
|
| + pReq->CompletionStatus = status;
|
| + pReq->CompletionRoutine(pReq);
|
| + status = A_OK;
|
| + }
|
| +
|
| + return status;
|
| +}
|
| +
|
| +
|
| +static void DevCleanupVirtualScatterSupport(AR6K_DEVICE *pDev)
|
| +{
|
| + HIF_SCATTER_REQ *pReq;
|
| +
|
| + while (1) {
|
| + pReq = DevAllocScatterReq((HIF_DEVICE *)pDev);
|
| + if (NULL == pReq) {
|
| + break;
|
| + }
|
| + A_FREE(pReq);
|
| + }
|
| +
|
| +}
|
| +
|
| + /* function to set up virtual scatter support if HIF layer has not implemented the interface */
|
| +static A_STATUS DevSetupVirtualScatterSupport(AR6K_DEVICE *pDev)
|
| +{
|
| + A_STATUS status = A_OK;
|
| + int bufferSize, sgreqSize;
|
| + int i;
|
| + DEV_SCATTER_DMA_VIRTUAL_INFO *pVirtualInfo;
|
| + HIF_SCATTER_REQ *pReq;
|
| +
|
| + bufferSize = sizeof(DEV_SCATTER_DMA_VIRTUAL_INFO) +
|
| + 2 * (A_GET_CACHE_LINE_BYTES()) + AR6K_MAX_TRANSFER_SIZE_PER_SCATTER;
|
| +
|
| + sgreqSize = sizeof(HIF_SCATTER_REQ) +
|
| + (AR6K_SCATTER_ENTRIES_PER_REQ - 1) * (sizeof(HIF_SCATTER_ITEM));
|
| +
|
| + for (i = 0; i < AR6K_SCATTER_REQS; i++) {
|
| + /* allocate the scatter request, buffer info and the actual virtual buffer itself */
|
| + pReq = (HIF_SCATTER_REQ *)A_MALLOC(sgreqSize + bufferSize);
|
| +
|
| + if (NULL == pReq) {
|
| + status = A_NO_MEMORY;
|
| + break;
|
| + }
|
| +
|
| + A_MEMZERO(pReq, sgreqSize);
|
| +
|
| + /* the virtual DMA starts after the scatter request struct */
|
| + pVirtualInfo = (DEV_SCATTER_DMA_VIRTUAL_INFO *)((A_UINT8 *)pReq + sgreqSize);
|
| + A_MEMZERO(pVirtualInfo, sizeof(DEV_SCATTER_DMA_VIRTUAL_INFO));
|
| +
|
| + pVirtualInfo->pVirtDmaBuffer = &pVirtualInfo->DataArea[0];
|
| + /* align buffer to cache line in case host controller can actually DMA this */
|
| + pVirtualInfo->pVirtDmaBuffer = A_ALIGN_TO_CACHE_LINE(pVirtualInfo->pVirtDmaBuffer);
|
| + /* store the structure in the private area */
|
| + pReq->HIFPrivate[0] = pVirtualInfo;
|
| + /* we emulate a DMA bounce interface */
|
| + pReq->ScatterMethod = HIF_SCATTER_DMA_BOUNCE;
|
| + pReq->pScatterBounceBuffer = pVirtualInfo->pVirtDmaBuffer;
|
| + /* free request to the list */
|
| + DevFreeScatterReq((HIF_DEVICE *)pDev,pReq);
|
| + }
|
| +
|
| + if (A_FAILED(status)) {
|
| + DevCleanupVirtualScatterSupport(pDev);
|
| + } else {
|
| + pDev->HifScatterInfo.pAllocateReqFunc = DevAllocScatterReq;
|
| + pDev->HifScatterInfo.pFreeReqFunc = DevFreeScatterReq;
|
| + pDev->HifScatterInfo.pReadWriteScatterFunc = DevReadWriteScatter;
|
| + pDev->HifScatterInfo.MaxScatterEntries = AR6K_SCATTER_ENTRIES_PER_REQ;
|
| + pDev->HifScatterInfo.MaxTransferSizePerScatterReq = AR6K_MAX_TRANSFER_SIZE_PER_SCATTER;
|
| + pDev->ScatterIsVirtual = TRUE;
|
| + }
|
| +
|
| + return status;
|
| +}
|
| +
|
| +
|
| +A_STATUS DevSetupMsgBundling(AR6K_DEVICE *pDev, int MaxMsgsPerTransfer)
|
| +{
|
| + A_STATUS status;
|
| +
|
| + status = HIFConfigureDevice(pDev->HIFDevice,
|
| + HIF_CONFIGURE_QUERY_SCATTER_REQUEST_SUPPORT,
|
| + &pDev->HifScatterInfo,
|
| + sizeof(pDev->HifScatterInfo));
|
| +
|
| + if (A_FAILED(status)) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_WARN,
|
| + ("AR6K: ** HIF layer does not support scatter requests (%d) \n",status));
|
| +
|
| + /* we can try to use a virtual DMA scatter mechanism using legacy HIFReadWrite() */
|
| + status = DevSetupVirtualScatterSupport(pDev);
|
| +
|
| + if (A_SUCCESS(status)) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("AR6K: virtual scatter transfers enabled (max scatter items:%d: maxlen:%d) \n",
|
| + DEV_GET_MAX_MSG_PER_BUNDLE(pDev), DEV_GET_MAX_BUNDLE_LENGTH(pDev)));
|
| + }
|
| +
|
| + } else {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("AR6K: HIF layer supports scatter requests (max scatter items:%d: maxlen:%d) \n",
|
| + DEV_GET_MAX_MSG_PER_BUNDLE(pDev), DEV_GET_MAX_BUNDLE_LENGTH(pDev)));
|
| + }
|
| +
|
| + if (A_SUCCESS(status)) {
|
| + /* for the recv path, the maximum number of bytes per recv bundle is just limited
|
| + * by the maximum transfer size at the HIF layer */
|
| + pDev->MaxRecvBundleSize = pDev->HifScatterInfo.MaxTransferSizePerScatterReq;
|
| +
|
| + /* for the send path, the max transfer size is limited by the existence and size of
|
| + * the extended mailbox address range */
|
| + if (pDev->MailBoxInfo.MboxProp[0].ExtendedAddress != 0) {
|
| + pDev->MaxSendBundleSize = pDev->MailBoxInfo.MboxProp[0].ExtendedSize;
|
| + } else {
|
| + /* legacy */
|
| + pDev->MaxSendBundleSize = AR6K_LEGACY_MAX_WRITE_LENGTH;
|
| + }
|
| +
|
| + if (pDev->MaxSendBundleSize > pDev->HifScatterInfo.MaxTransferSizePerScatterReq) {
|
| + /* limit send bundle size to what the HIF can support for scatter requests */
|
| + pDev->MaxSendBundleSize = pDev->HifScatterInfo.MaxTransferSizePerScatterReq;
|
| + }
|
| +
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ANY,
|
| + ("AR6K: max recv: %d max send: %d \n",
|
| + DEV_GET_MAX_BUNDLE_RECV_LENGTH(pDev), DEV_GET_MAX_BUNDLE_SEND_LENGTH(pDev)));
|
| +
|
| + }
|
| + return status;
|
| +}
|
| +
|
| +A_STATUS DevSubmitScatterRequest(AR6K_DEVICE *pDev, HIF_SCATTER_REQ *pScatterReq, A_BOOL Read, A_BOOL Async)
|
| +{
|
| + A_STATUS status;
|
| +
|
| + if (Read) {
|
| + /* read operation */
|
| + pScatterReq->Request = (Async) ? HIF_RD_ASYNC_BLOCK_FIX : HIF_RD_SYNC_BLOCK_FIX;
|
| + pScatterReq->Address = pDev->MailBoxInfo.MboxAddresses[HTC_MAILBOX];
|
| + A_ASSERT(pScatterReq->TotalLength <= (A_UINT32)DEV_GET_MAX_BUNDLE_RECV_LENGTH(pDev));
|
| + } else {
|
| + A_UINT32 mailboxWidth;
|
| +
|
| + /* write operation */
|
| + pScatterReq->Request = (Async) ? HIF_WR_ASYNC_BLOCK_INC : HIF_WR_SYNC_BLOCK_INC;
|
| + A_ASSERT(pScatterReq->TotalLength <= (A_UINT32)DEV_GET_MAX_BUNDLE_SEND_LENGTH(pDev));
|
| + if (pScatterReq->TotalLength > AR6K_LEGACY_MAX_WRITE_LENGTH) {
|
| + /* for large writes use the extended address */
|
| + pScatterReq->Address = pDev->MailBoxInfo.MboxProp[HTC_MAILBOX].ExtendedAddress;
|
| + mailboxWidth = pDev->MailBoxInfo.MboxProp[HTC_MAILBOX].ExtendedSize;
|
| + } else {
|
| + pScatterReq->Address = pDev->MailBoxInfo.MboxAddresses[HTC_MAILBOX];
|
| + mailboxWidth = AR6K_LEGACY_MAX_WRITE_LENGTH;
|
| + }
|
| +
|
| + if (!pDev->ScatterIsVirtual) {
|
| + /* we are passing this scatter list down to the HIF layer' scatter request handler, fixup the address
|
| + * so that the last byte falls on the EOM, we do this for those HIFs that support the
|
| + * scatter API */
|
| + pScatterReq->Address += (mailboxWidth - pScatterReq->TotalLength);
|
| + }
|
| +
|
| + }
|
| +
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_RECV | ATH_DEBUG_SEND,
|
| + ("DevSubmitScatterRequest, Entries: %d, Total Length: %d Mbox:0x%X (mode: %s : %s)\n",
|
| + pScatterReq->ValidScatterEntries,
|
| + pScatterReq->TotalLength,
|
| + pScatterReq->Address,
|
| + Async ? "ASYNC" : "SYNC",
|
| + (Read) ? "RD" : "WR"));
|
| +
|
| + status = DEV_PREPARE_SCATTER_OPERATION(pScatterReq);
|
| +
|
| + if (A_FAILED(status)) {
|
| + if (Async) {
|
| + pScatterReq->CompletionStatus = status;
|
| + pScatterReq->CompletionRoutine(pScatterReq);
|
| + return A_OK;
|
| + }
|
| + return status;
|
| + }
|
| +
|
| + status = pDev->HifScatterInfo.pReadWriteScatterFunc(pDev->ScatterIsVirtual ? pDev : pDev->HIFDevice,
|
| + pScatterReq);
|
| + if (!Async) {
|
| + /* in sync mode, we can touch the scatter request */
|
| + pScatterReq->CompletionStatus = status;
|
| + DEV_FINISH_SCATTER_OPERATION(pScatterReq);
|
| + } else {
|
| + if (status == A_PENDING) {
|
| + status = A_OK;
|
| + }
|
| + }
|
| +
|
| + return status;
|
| +}
|
| +
|
| +
|
| +#ifdef MBOXHW_UNIT_TEST
|
| +
|
| +
|
| +/* This is a mailbox hardware unit test that must be called in a schedulable context
|
| + * This test is very simple, it will send a list of buffers with a counting pattern
|
| + * and the target will invert the data and send the message back
|
| + *
|
| + * the unit test has the following constraints:
|
| + *
|
| + * The target has at least 8 buffers of 256 bytes each. The host will send
|
| + * the following pattern of buffers in rapid succession :
|
| + *
|
| + * 1 buffer - 128 bytes
|
| + * 1 buffer - 256 bytes
|
| + * 1 buffer - 512 bytes
|
| + * 1 buffer - 1024 bytes
|
| + *
|
| + * The host will send the buffers to one mailbox and wait for buffers to be reflected
|
| + * back from the same mailbox. The target sends the buffers FIFO order.
|
| + * Once the final buffer has been received for a mailbox, the next mailbox is tested.
|
| + *
|
| + *
|
| + * Note: To simplifythe test , we assume that the chosen buffer sizes
|
| + * will fall on a nice block pad
|
| + *
|
| + * It is expected that higher-order tests will be written to stress the mailboxes using
|
| + * a message-based protocol (with some performance timming) that can create more
|
| + * randomness in the packets sent over mailboxes.
|
| + *
|
| + * */
|
| +
|
| +#define A_ROUND_UP_PWR2(x, align) (((int) (x) + ((align)-1)) & ~((align)-1))
|
| +
|
| +#define BUFFER_BLOCK_PAD 128
|
| +
|
| +#if 0
|
| +#define BUFFER1 128
|
| +#define BUFFER2 256
|
| +#define BUFFER3 512
|
| +#define BUFFER4 1024
|
| +#endif
|
| +
|
| +#if 1
|
| +#define BUFFER1 80
|
| +#define BUFFER2 200
|
| +#define BUFFER3 444
|
| +#define BUFFER4 800
|
| +#endif
|
| +
|
| +#define TOTAL_BYTES (A_ROUND_UP_PWR2(BUFFER1,BUFFER_BLOCK_PAD) + \
|
| + A_ROUND_UP_PWR2(BUFFER2,BUFFER_BLOCK_PAD) + \
|
| + A_ROUND_UP_PWR2(BUFFER3,BUFFER_BLOCK_PAD) + \
|
| + A_ROUND_UP_PWR2(BUFFER4,BUFFER_BLOCK_PAD) )
|
| +
|
| +#define TEST_BYTES (BUFFER1 + BUFFER2 + BUFFER3 + BUFFER4)
|
| +
|
| +#define TEST_CREDITS_RECV_TIMEOUT 100
|
| +
|
| +static A_UINT8 g_Buffer[TOTAL_BYTES];
|
| +static A_UINT32 g_MailboxAddrs[AR6K_MAILBOXES];
|
| +static A_UINT32 g_BlockSizes[AR6K_MAILBOXES];
|
| +
|
| +#define BUFFER_PROC_LIST_DEPTH 4
|
| +
|
| +typedef struct _BUFFER_PROC_LIST{
|
| + A_UINT8 *pBuffer;
|
| + A_UINT32 length;
|
| +}BUFFER_PROC_LIST;
|
| +
|
| +
|
| +#define PUSH_BUFF_PROC_ENTRY(pList,len,pCurrpos) \
|
| +{ \
|
| + (pList)->pBuffer = (pCurrpos); \
|
| + (pList)->length = (len); \
|
| + (pCurrpos) += (len); \
|
| + (pList)++; \
|
| +}
|
| +
|
| +/* a simple and crude way to send different "message" sizes */
|
| +static void AssembleBufferList(BUFFER_PROC_LIST *pList)
|
| +{
|
| + A_UINT8 *pBuffer = g_Buffer;
|
| +
|
| +#if BUFFER_PROC_LIST_DEPTH < 4
|
| +#error "Buffer processing list depth is not deep enough!!"
|
| +#endif
|
| +
|
| + PUSH_BUFF_PROC_ENTRY(pList,BUFFER1,pBuffer);
|
| + PUSH_BUFF_PROC_ENTRY(pList,BUFFER2,pBuffer);
|
| + PUSH_BUFF_PROC_ENTRY(pList,BUFFER3,pBuffer);
|
| + PUSH_BUFF_PROC_ENTRY(pList,BUFFER4,pBuffer);
|
| +
|
| +}
|
| +
|
| +#define FILL_ZERO TRUE
|
| +#define FILL_COUNTING FALSE
|
| +static void InitBuffers(A_BOOL Zero)
|
| +{
|
| + A_UINT16 *pBuffer16 = (A_UINT16 *)g_Buffer;
|
| + int i;
|
| +
|
| + /* fill buffer with 16 bit counting pattern or zeros */
|
| + for (i = 0; i < (TOTAL_BYTES / 2) ; i++) {
|
| + if (!Zero) {
|
| + pBuffer16[i] = (A_UINT16)i;
|
| + } else {
|
| + pBuffer16[i] = 0;
|
| + }
|
| + }
|
| +}
|
| +
|
| +
|
| +static A_BOOL CheckOneBuffer(A_UINT16 *pBuffer16, int Length)
|
| +{
|
| + int i;
|
| + A_UINT16 startCount;
|
| + A_BOOL success = TRUE;
|
| +
|
| + /* get the starting count */
|
| + startCount = pBuffer16[0];
|
| + /* invert it, this is the expected value */
|
| + startCount = ~startCount;
|
| + /* scan the buffer and verify */
|
| + for (i = 0; i < (Length / 2) ; i++,startCount++) {
|
| + /* target will invert all the data */
|
| + if ((A_UINT16)pBuffer16[i] != (A_UINT16)~startCount) {
|
| + success = FALSE;
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Invalid Data Got:0x%X, Expecting:0x%X (offset:%d, total:%d) \n",
|
| + pBuffer16[i], ((A_UINT16)~startCount), i, Length));
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("0x%X 0x%X 0x%X 0x%X \n",
|
| + pBuffer16[i], pBuffer16[i + 1], pBuffer16[i + 2],pBuffer16[i+3]));
|
| + break;
|
| + }
|
| + }
|
| +
|
| + return success;
|
| +}
|
| +
|
| +static A_BOOL CheckBuffers(void)
|
| +{
|
| + int i;
|
| + A_BOOL success = TRUE;
|
| + BUFFER_PROC_LIST checkList[BUFFER_PROC_LIST_DEPTH];
|
| +
|
| + /* assemble the list */
|
| + AssembleBufferList(checkList);
|
| +
|
| + /* scan the buffers and verify */
|
| + for (i = 0; i < BUFFER_PROC_LIST_DEPTH ; i++) {
|
| + success = CheckOneBuffer((A_UINT16 *)checkList[i].pBuffer, checkList[i].length);
|
| + if (!success) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Buffer : 0x%X, Length:%d failed verify \n",
|
| + (A_UINT32)checkList[i].pBuffer, checkList[i].length));
|
| + break;
|
| + }
|
| + }
|
| +
|
| + return success;
|
| +}
|
| +
|
| + /* find the end marker for the last buffer we will be sending */
|
| +static A_UINT16 GetEndMarker(void)
|
| +{
|
| + A_UINT8 *pBuffer;
|
| + BUFFER_PROC_LIST checkList[BUFFER_PROC_LIST_DEPTH];
|
| +
|
| + /* fill up buffers with the normal counting pattern */
|
| + InitBuffers(FILL_COUNTING);
|
| +
|
| + /* assemble the list we will be sending down */
|
| + AssembleBufferList(checkList);
|
| + /* point to the last 2 bytes of the last buffer */
|
| + pBuffer = &(checkList[BUFFER_PROC_LIST_DEPTH - 1].pBuffer[(checkList[BUFFER_PROC_LIST_DEPTH - 1].length) - 2]);
|
| +
|
| + /* the last count in the last buffer is the marker */
|
| + return (A_UINT16)pBuffer[0] | ((A_UINT16)pBuffer[1] << 8);
|
| +}
|
| +
|
| +#define ATH_PRINT_OUT_ZONE ATH_DEBUG_ERR
|
| +
|
| +/* send the ordered buffers to the target */
|
| +static A_STATUS SendBuffers(AR6K_DEVICE *pDev, int mbox)
|
| +{
|
| + A_STATUS status = A_OK;
|
| + A_UINT32 request = HIF_WR_SYNC_BLOCK_INC;
|
| + BUFFER_PROC_LIST sendList[BUFFER_PROC_LIST_DEPTH];
|
| + int i;
|
| + int totalBytes = 0;
|
| + int paddedLength;
|
| + int totalwPadding = 0;
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("Sending buffers on mailbox : %d \n",mbox));
|
| +
|
| + /* fill buffer with counting pattern */
|
| + InitBuffers(FILL_COUNTING);
|
| +
|
| + /* assemble the order in which we send */
|
| + AssembleBufferList(sendList);
|
| +
|
| + for (i = 0; i < BUFFER_PROC_LIST_DEPTH; i++) {
|
| +
|
| + /* we are doing block transfers, so we need to pad everything to a block size */
|
| + paddedLength = (sendList[i].length + (g_BlockSizes[mbox] - 1)) &
|
| + (~(g_BlockSizes[mbox] - 1));
|
| +
|
| + /* send each buffer synchronously */
|
| + status = HIFReadWrite(pDev->HIFDevice,
|
| + g_MailboxAddrs[mbox],
|
| + sendList[i].pBuffer,
|
| + paddedLength,
|
| + request,
|
| + NULL);
|
| + if (status != A_OK) {
|
| + break;
|
| + }
|
| + totalBytes += sendList[i].length;
|
| + totalwPadding += paddedLength;
|
| + }
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("Sent %d bytes (%d padded bytes) to mailbox : %d \n",totalBytes,totalwPadding,mbox));
|
| +
|
| + return status;
|
| +}
|
| +
|
| +/* poll the mailbox credit counter until we get a credit or timeout */
|
| +static A_STATUS GetCredits(AR6K_DEVICE *pDev, int mbox, int *pCredits)
|
| +{
|
| + A_STATUS status = A_OK;
|
| + int timeout = TEST_CREDITS_RECV_TIMEOUT;
|
| + A_UINT8 credits = 0;
|
| + A_UINT32 address;
|
| +
|
| + while (TRUE) {
|
| +
|
| + /* Read the counter register to get credits, this auto-decrements */
|
| + address = COUNT_DEC_ADDRESS + (AR6K_MAILBOXES + mbox) * 4;
|
| + status = HIFReadWrite(pDev->HIFDevice, address, &credits, sizeof(credits),
|
| + HIF_RD_SYNC_BYTE_FIX, NULL);
|
| + if (status != A_OK) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR,
|
| + ("Unable to decrement the command credit count register (mbox=%d)\n",mbox));
|
| + status = A_ERROR;
|
| + break;
|
| + }
|
| +
|
| + if (credits) {
|
| + break;
|
| + }
|
| +
|
| + timeout--;
|
| +
|
| + if (timeout <= 0) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR,
|
| + (" Timeout reading credit registers (mbox=%d, address:0x%X) \n",mbox,address));
|
| + status = A_ERROR;
|
| + break;
|
| + }
|
| +
|
| + /* delay a little, target may not be ready */
|
| + A_MDELAY(1000);
|
| +
|
| + }
|
| +
|
| + if (status == A_OK) {
|
| + *pCredits = credits;
|
| + }
|
| +
|
| + return status;
|
| +}
|
| +
|
| +
|
| +/* wait for the buffers to come back */
|
| +static A_STATUS RecvBuffers(AR6K_DEVICE *pDev, int mbox)
|
| +{
|
| + A_STATUS status = A_OK;
|
| + A_UINT32 request = HIF_RD_SYNC_BLOCK_INC;
|
| + BUFFER_PROC_LIST recvList[BUFFER_PROC_LIST_DEPTH];
|
| + int curBuffer;
|
| + int credits;
|
| + int i;
|
| + int totalBytes = 0;
|
| + int paddedLength;
|
| + int totalwPadding = 0;
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("Waiting for buffers on mailbox : %d \n",mbox));
|
| +
|
| + /* zero the buffers */
|
| + InitBuffers(FILL_ZERO);
|
| +
|
| + /* assemble the order in which we should receive */
|
| + AssembleBufferList(recvList);
|
| +
|
| + curBuffer = 0;
|
| +
|
| + while (curBuffer < BUFFER_PROC_LIST_DEPTH) {
|
| +
|
| + /* get number of buffers that have been completed, this blocks
|
| + * until we get at least 1 credit or it times out */
|
| + status = GetCredits(pDev, mbox, &credits);
|
| +
|
| + if (status != A_OK) {
|
| + break;
|
| + }
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("Got %d messages on mailbox : %d \n",credits, mbox));
|
| +
|
| + /* get all the buffers that are sitting on the queue */
|
| + for (i = 0; i < credits; i++) {
|
| + A_ASSERT(curBuffer < BUFFER_PROC_LIST_DEPTH);
|
| + /* recv the current buffer synchronously, the buffers should come back in
|
| + * order... with padding applied by the target */
|
| + paddedLength = (recvList[curBuffer].length + (g_BlockSizes[mbox] - 1)) &
|
| + (~(g_BlockSizes[mbox] - 1));
|
| +
|
| + status = HIFReadWrite(pDev->HIFDevice,
|
| + g_MailboxAddrs[mbox],
|
| + recvList[curBuffer].pBuffer,
|
| + paddedLength,
|
| + request,
|
| + NULL);
|
| + if (status != A_OK) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Failed to read %d bytes on mailbox:%d : address:0x%X \n",
|
| + recvList[curBuffer].length, mbox, g_MailboxAddrs[mbox]));
|
| + break;
|
| + }
|
| +
|
| + totalwPadding += paddedLength;
|
| + totalBytes += recvList[curBuffer].length;
|
| + curBuffer++;
|
| + }
|
| +
|
| + if (status != A_OK) {
|
| + break;
|
| + }
|
| + /* go back and get some more */
|
| + credits = 0;
|
| + }
|
| +
|
| + if (totalBytes != TEST_BYTES) {
|
| + A_ASSERT(FALSE);
|
| + } else {
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("Got all buffers on mbox:%d total recv :%d (w/Padding : %d) \n",
|
| + mbox, totalBytes, totalwPadding));
|
| + }
|
| +
|
| + return status;
|
| +
|
| +
|
| +}
|
| +
|
| +static A_STATUS DoOneMboxHWTest(AR6K_DEVICE *pDev, int mbox)
|
| +{
|
| + A_STATUS status;
|
| +
|
| + do {
|
| + /* send out buffers */
|
| + status = SendBuffers(pDev,mbox);
|
| +
|
| + if (status != A_OK) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Sending buffers Failed : %d mbox:%d\n",status,mbox));
|
| + break;
|
| + }
|
| +
|
| + /* go get them, this will block */
|
| + status = RecvBuffers(pDev, mbox);
|
| +
|
| + if (status != A_OK) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Recv buffers Failed : %d mbox:%d\n",status,mbox));
|
| + break;
|
| + }
|
| +
|
| + /* check the returned data patterns */
|
| + if (!CheckBuffers()) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Buffer Verify Failed : mbox:%d\n",mbox));
|
| + status = A_ERROR;
|
| + break;
|
| + }
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, (" Send/Recv success! mailbox : %d \n",mbox));
|
| +
|
| + } while (FALSE);
|
| +
|
| + return status;
|
| +}
|
| +
|
| +/* here is where the test starts */
|
| +A_STATUS DoMboxHWTest(AR6K_DEVICE *pDev)
|
| +{
|
| + int i;
|
| + A_STATUS status;
|
| + int credits = 0;
|
| + A_UINT8 params[4];
|
| + int numBufs;
|
| + int bufferSize;
|
| + A_UINT16 temp;
|
| +
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, (" DoMboxHWTest START - \n"));
|
| +
|
| + do {
|
| + /* get the addresses for all 4 mailboxes */
|
| + status = HIFConfigureDevice(pDev->HIFDevice, HIF_DEVICE_GET_MBOX_ADDR,
|
| + g_MailboxAddrs, sizeof(g_MailboxAddrs));
|
| +
|
| + if (status != A_OK) {
|
| + A_ASSERT(FALSE);
|
| + break;
|
| + }
|
| +
|
| + /* get the block sizes */
|
| + status = HIFConfigureDevice(pDev->HIFDevice, HIF_DEVICE_GET_MBOX_BLOCK_SIZE,
|
| + g_BlockSizes, sizeof(g_BlockSizes));
|
| +
|
| + if (status != A_OK) {
|
| + A_ASSERT(FALSE);
|
| + break;
|
| + }
|
| +
|
| + /* note, the HIF layer usually reports mbox 0 to have a block size of
|
| + * 1, but our test wants to run in block-mode for all mailboxes, so we treat all mailboxes
|
| + * the same. */
|
| + g_BlockSizes[0] = g_BlockSizes[1];
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("Block Size to use: %d \n",g_BlockSizes[0]));
|
| +
|
| + if (g_BlockSizes[1] > BUFFER_BLOCK_PAD) {
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("%d Block size is too large for buffer pad %d\n",
|
| + g_BlockSizes[1], BUFFER_BLOCK_PAD));
|
| + break;
|
| + }
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("Waiting for target.... \n"));
|
| +
|
| + /* the target lets us know it is ready by giving us 1 credit on
|
| + * mailbox 0 */
|
| + status = GetCredits(pDev, 0, &credits);
|
| +
|
| + if (status != A_OK) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Failed to wait for target ready \n"));
|
| + break;
|
| + }
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("Target is ready ...\n"));
|
| +
|
| + /* read the first 4 scratch registers */
|
| + status = HIFReadWrite(pDev->HIFDevice,
|
| + SCRATCH_ADDRESS,
|
| + params,
|
| + 4,
|
| + HIF_RD_SYNC_BYTE_INC,
|
| + NULL);
|
| +
|
| + if (status != A_OK) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Failed to wait get parameters \n"));
|
| + break;
|
| + }
|
| +
|
| + numBufs = params[0];
|
| + bufferSize = (int)(((A_UINT16)params[2] << 8) | (A_UINT16)params[1]);
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE,
|
| + ("Target parameters: bufs per mailbox:%d, buffer size:%d bytes (total space: %d, minimum required space (w/padding): %d) \n",
|
| + numBufs, bufferSize, (numBufs * bufferSize), TOTAL_BYTES));
|
| +
|
| + if ((numBufs * bufferSize) < TOTAL_BYTES) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Not Enough buffer space to run test! need:%d, got:%d \n",
|
| + TOTAL_BYTES, (numBufs*bufferSize)));
|
| + status = A_ERROR;
|
| + break;
|
| + }
|
| +
|
| + temp = GetEndMarker();
|
| +
|
| + status = HIFReadWrite(pDev->HIFDevice,
|
| + SCRATCH_ADDRESS + 4,
|
| + (A_UINT8 *)&temp,
|
| + 2,
|
| + HIF_WR_SYNC_BYTE_INC,
|
| + NULL);
|
| +
|
| + if (status != A_OK) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Failed to write end marker \n"));
|
| + break;
|
| + }
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("End Marker: 0x%X \n",temp));
|
| +
|
| + temp = (A_UINT16)g_BlockSizes[1];
|
| + /* convert to a mask */
|
| + temp = temp - 1;
|
| + status = HIFReadWrite(pDev->HIFDevice,
|
| + SCRATCH_ADDRESS + 6,
|
| + (A_UINT8 *)&temp,
|
| + 2,
|
| + HIF_WR_SYNC_BYTE_INC,
|
| + NULL);
|
| +
|
| + if (status != A_OK) {
|
| + AR_DEBUG_PRINTF(ATH_DEBUG_ERR, ("Failed to write block mask \n"));
|
| + break;
|
| + }
|
| +
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, ("Set Block Mask: 0x%X \n",temp));
|
| +
|
| + /* execute the test on each mailbox */
|
| + for (i = 0; i < AR6K_MAILBOXES; i++) {
|
| + status = DoOneMboxHWTest(pDev, i);
|
| + if (status != A_OK) {
|
| + break;
|
| + }
|
| + }
|
| +
|
| + } while (FALSE);
|
| +
|
| + if (status == A_OK) {
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, (" DoMboxHWTest DONE - SUCCESS! - \n"));
|
| + } else {
|
| + AR_DEBUG_PRINTF(ATH_PRINT_OUT_ZONE, (" DoMboxHWTest DONE - FAILED! - \n"));
|
| + }
|
| + /* don't let HTC_Start continue, the target is actually not running any HTC code */
|
| + return A_ERROR;
|
| +}
|
| +#endif
|
| +
|
| +
|
| +
|
|
|