| /* |
| * Copyright (c) 2010 Broadcom Corporation |
| * |
| * Permission to use, copy, modify, and/or distribute this software for any |
| * purpose with or without fee is hereby granted, provided that the above |
| * copyright notice and this permission notice appear in all copies. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES |
| * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF |
| * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY |
| * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES |
| * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION |
| * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN |
| * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. |
| */ |
| #include <linux/delay.h> |
| #include <linux/pci.h> |
| |
| #include <defs.h> |
| #include <chipcommon.h> |
| #include <brcmu_utils.h> |
| #include <brcm_hw_ids.h> |
| #include "types.h" |
| #include "pub.h" |
| #include "pmu.h" |
| #include "srom.h" |
| #include "nicpci.h" |
| #include "aiutils.h" |
| |
| /* slow_clk_ctl */ |
| #define SCC_SS_MASK 0x00000007 /* slow clock source mask */ |
| #define SCC_SS_LPO 0x00000000 /* source of slow clock is LPO */ |
| #define SCC_SS_XTAL 0x00000001 /* source of slow clock is crystal */ |
| #define SCC_SS_PCI 0x00000002 /* source of slow clock is PCI */ |
| #define SCC_LF 0x00000200 /* LPOFreqSel, 1: 160Khz, 0: 32KHz */ |
| #define SCC_LP 0x00000400 /* LPOPowerDown, 1: LPO is disabled, |
| * 0: LPO is enabled |
| */ |
| #define SCC_FS 0x00000800 /* ForceSlowClk, 1: sb/cores running on slow clock, |
| * 0: power logic control |
| */ |
| #define SCC_IP 0x00001000 /* IgnorePllOffReq, 1/0: power logic ignores/honors |
| * PLL clock disable requests from core |
| */ |
| #define SCC_XC 0x00002000 /* XtalControlEn, 1/0: power logic does/doesn't |
| * disable crystal when appropriate |
| */ |
| #define SCC_XP 0x00004000 /* XtalPU (RO), 1/0: crystal running/disabled */ |
| #define SCC_CD_MASK 0xffff0000 /* ClockDivider (SlowClk = 1/(4+divisor)) */ |
| #define SCC_CD_SHIFT 16 |
| |
| /* system_clk_ctl */ |
| #define SYCC_IE 0x00000001 /* ILPen: Enable Idle Low Power */ |
| #define SYCC_AE 0x00000002 /* ALPen: Enable Active Low Power */ |
| #define SYCC_FP 0x00000004 /* ForcePLLOn */ |
| #define SYCC_AR 0x00000008 /* Force ALP (or HT if ALPen is not set */ |
| #define SYCC_HR 0x00000010 /* Force HT */ |
| #define SYCC_CD_MASK 0xffff0000 /* ClkDiv (ILP = 1/(4 * (divisor + 1)) */ |
| #define SYCC_CD_SHIFT 16 |
| |
| #define CST4329_SPROM_OTP_SEL_MASK 0x00000003 |
| #define CST4329_DEFCIS_SEL 0 /* OTP is powered up, use def. CIS, no SPROM */ |
| #define CST4329_SPROM_SEL 1 /* OTP is powered up, SPROM is present */ |
| #define CST4329_OTP_SEL 2 /* OTP is powered up, no SPROM */ |
| #define CST4329_OTP_PWRDN 3 /* OTP is powered down, SPROM is present */ |
| #define CST4329_SPI_SDIO_MODE_MASK 0x00000004 |
| #define CST4329_SPI_SDIO_MODE_SHIFT 2 |
| |
| /* 43224 chip-specific ChipControl register bits */ |
| #define CCTRL43224_GPIO_TOGGLE 0x8000 |
| #define CCTRL_43224A0_12MA_LED_DRIVE 0x00F000F0 /* 12 mA drive strength */ |
| #define CCTRL_43224B0_12MA_LED_DRIVE 0xF0 /* 12 mA drive strength for later 43224s */ |
| |
| /* 43236 Chip specific ChipStatus register bits */ |
| #define CST43236_SFLASH_MASK 0x00000040 |
| #define CST43236_OTP_MASK 0x00000080 |
| #define CST43236_HSIC_MASK 0x00000100 /* USB/HSIC */ |
| #define CST43236_BP_CLK 0x00000200 /* 120/96Mbps */ |
| #define CST43236_BOOT_MASK 0x00001800 |
| #define CST43236_BOOT_SHIFT 11 |
| #define CST43236_BOOT_FROM_SRAM 0 /* boot from SRAM, ARM in reset */ |
| #define CST43236_BOOT_FROM_ROM 1 /* boot from ROM */ |
| #define CST43236_BOOT_FROM_FLASH 2 /* boot from FLASH */ |
| #define CST43236_BOOT_FROM_INVALID 3 |
| |
| /* 4331 chip-specific ChipControl register bits */ |
| #define CCTRL4331_BT_COEXIST (1<<0) /* 0 disable */ |
| #define CCTRL4331_SECI (1<<1) /* 0 SECI is disabled (JATG functional) */ |
| #define CCTRL4331_EXT_LNA (1<<2) /* 0 disable */ |
| #define CCTRL4331_SPROM_GPIO13_15 (1<<3) /* sprom/gpio13-15 mux */ |
| #define CCTRL4331_EXTPA_EN (1<<4) /* 0 ext pa disable, 1 ext pa enabled */ |
| #define CCTRL4331_GPIOCLK_ON_SPROMCS (1<<5) /* set drive out GPIO_CLK on sprom_cs pin */ |
| #define CCTRL4331_PCIE_MDIO_ON_SPROMCS (1<<6) /* use sprom_cs pin as PCIE mdio interface */ |
| #define CCTRL4331_EXTPA_ON_GPIO2_5 (1<<7) /* aband extpa will be at gpio2/5 and sprom_dout */ |
| #define CCTRL4331_OVR_PIPEAUXCLKEN (1<<8) /* override core control on pipe_AuxClkEnable */ |
| #define CCTRL4331_OVR_PIPEAUXPWRDOWN (1<<9) /* override core control on pipe_AuxPowerDown */ |
| #define CCTRL4331_PCIE_AUXCLKEN (1<<10) /* pcie_auxclkenable */ |
| #define CCTRL4331_PCIE_PIPE_PLLDOWN (1<<11) /* pcie_pipe_pllpowerdown */ |
| #define CCTRL4331_BT_SHD0_ON_GPIO4 (1<<16) /* enable bt_shd0 at gpio4 */ |
| #define CCTRL4331_BT_SHD1_ON_GPIO5 (1<<17) /* enable bt_shd1 at gpio5 */ |
| |
| /* 4331 Chip specific ChipStatus register bits */ |
| #define CST4331_XTAL_FREQ 0x00000001 /* crystal frequency 20/40Mhz */ |
| #define CST4331_SPROM_PRESENT 0x00000002 |
| #define CST4331_OTP_PRESENT 0x00000004 |
| #define CST4331_LDO_RF 0x00000008 |
| #define CST4331_LDO_PAR 0x00000010 |
| |
| /* 4319 chip-specific ChipStatus register bits */ |
| #define CST4319_SPI_CPULESSUSB 0x00000001 |
| #define CST4319_SPI_CLK_POL 0x00000002 |
| #define CST4319_SPI_CLK_PH 0x00000008 |
| #define CST4319_SPROM_OTP_SEL_MASK 0x000000c0 /* gpio [7:6], SDIO CIS selection */ |
| #define CST4319_SPROM_OTP_SEL_SHIFT 6 |
| #define CST4319_DEFCIS_SEL 0x00000000 /* use default CIS, OTP is powered up */ |
| #define CST4319_SPROM_SEL 0x00000040 /* use SPROM, OTP is powered up */ |
| #define CST4319_OTP_SEL 0x00000080 /* use OTP, OTP is powered up */ |
| #define CST4319_OTP_PWRDN 0x000000c0 /* use SPROM, OTP is powered down */ |
| #define CST4319_SDIO_USB_MODE 0x00000100 /* gpio [8], sdio/usb mode */ |
| #define CST4319_REMAP_SEL_MASK 0x00000600 |
| #define CST4319_ILPDIV_EN 0x00000800 |
| #define CST4319_XTAL_PD_POL 0x00001000 |
| #define CST4319_LPO_SEL 0x00002000 |
| #define CST4319_RES_INIT_MODE 0x0000c000 |
| #define CST4319_PALDO_EXTPNP 0x00010000 /* PALDO is configured with external PNP */ |
| #define CST4319_CBUCK_MODE_MASK 0x00060000 |
| #define CST4319_CBUCK_MODE_BURST 0x00020000 |
| #define CST4319_CBUCK_MODE_LPBURST 0x00060000 |
| #define CST4319_RCAL_VALID 0x01000000 |
| #define CST4319_RCAL_VALUE_MASK 0x3e000000 |
| #define CST4319_RCAL_VALUE_SHIFT 25 |
| |
| /* 4336 chip-specific ChipStatus register bits */ |
| #define CST4336_SPI_MODE_MASK 0x00000001 |
| #define CST4336_SPROM_PRESENT 0x00000002 |
| #define CST4336_OTP_PRESENT 0x00000004 |
| #define CST4336_ARMREMAP_0 0x00000008 |
| #define CST4336_ILPDIV_EN_MASK 0x00000010 |
| #define CST4336_ILPDIV_EN_SHIFT 4 |
| #define CST4336_XTAL_PD_POL_MASK 0x00000020 |
| #define CST4336_XTAL_PD_POL_SHIFT 5 |
| #define CST4336_LPO_SEL_MASK 0x00000040 |
| #define CST4336_LPO_SEL_SHIFT 6 |
| #define CST4336_RES_INIT_MODE_MASK 0x00000180 |
| #define CST4336_RES_INIT_MODE_SHIFT 7 |
| #define CST4336_CBUCK_MODE_MASK 0x00000600 |
| #define CST4336_CBUCK_MODE_SHIFT 9 |
| |
| /* 4313 chip-specific ChipStatus register bits */ |
| #define CST4313_SPROM_PRESENT 1 |
| #define CST4313_OTP_PRESENT 2 |
| #define CST4313_SPROM_OTP_SEL_MASK 0x00000002 |
| #define CST4313_SPROM_OTP_SEL_SHIFT 0 |
| |
| /* 4313 Chip specific ChipControl register bits */ |
| #define CCTRL_4313_12MA_LED_DRIVE 0x00000007 /* 12 mA drive strengh for later 4313 */ |
| |
| #define BCM47162_DMP() ((sih->chip == BCM47162_CHIP_ID) && \ |
| (sih->chiprev == 0) && \ |
| (sii->coreid[sii->curidx] == MIPS74K_CORE_ID)) |
| |
| /* Manufacturer Ids */ |
| #define MFGID_ARM 0x43b |
| #define MFGID_BRCM 0x4bf |
| #define MFGID_MIPS 0x4a7 |
| |
| /* Enumeration ROM registers */ |
| #define ER_EROMENTRY 0x000 |
| #define ER_REMAPCONTROL 0xe00 |
| #define ER_REMAPSELECT 0xe04 |
| #define ER_MASTERSELECT 0xe10 |
| #define ER_ITCR 0xf00 |
| #define ER_ITIP 0xf04 |
| |
| /* Erom entries */ |
| #define ER_TAG 0xe |
| #define ER_TAG1 0x6 |
| #define ER_VALID 1 |
| #define ER_CI 0 |
| #define ER_MP 2 |
| #define ER_ADD 4 |
| #define ER_END 0xe |
| #define ER_BAD 0xffffffff |
| |
| /* EROM CompIdentA */ |
| #define CIA_MFG_MASK 0xfff00000 |
| #define CIA_MFG_SHIFT 20 |
| #define CIA_CID_MASK 0x000fff00 |
| #define CIA_CID_SHIFT 8 |
| #define CIA_CCL_MASK 0x000000f0 |
| #define CIA_CCL_SHIFT 4 |
| |
| /* EROM CompIdentB */ |
| #define CIB_REV_MASK 0xff000000 |
| #define CIB_REV_SHIFT 24 |
| #define CIB_NSW_MASK 0x00f80000 |
| #define CIB_NSW_SHIFT 19 |
| #define CIB_NMW_MASK 0x0007c000 |
| #define CIB_NMW_SHIFT 14 |
| #define CIB_NSP_MASK 0x00003e00 |
| #define CIB_NSP_SHIFT 9 |
| #define CIB_NMP_MASK 0x000001f0 |
| #define CIB_NMP_SHIFT 4 |
| |
| /* EROM AddrDesc */ |
| #define AD_ADDR_MASK 0xfffff000 |
| #define AD_SP_MASK 0x00000f00 |
| #define AD_SP_SHIFT 8 |
| #define AD_ST_MASK 0x000000c0 |
| #define AD_ST_SHIFT 6 |
| #define AD_ST_SLAVE 0x00000000 |
| #define AD_ST_BRIDGE 0x00000040 |
| #define AD_ST_SWRAP 0x00000080 |
| #define AD_ST_MWRAP 0x000000c0 |
| #define AD_SZ_MASK 0x00000030 |
| #define AD_SZ_SHIFT 4 |
| #define AD_SZ_4K 0x00000000 |
| #define AD_SZ_8K 0x00000010 |
| #define AD_SZ_16K 0x00000020 |
| #define AD_SZ_SZD 0x00000030 |
| #define AD_AG32 0x00000008 |
| #define AD_ADDR_ALIGN 0x00000fff |
| #define AD_SZ_BASE 0x00001000 /* 4KB */ |
| |
| /* EROM SizeDesc */ |
| #define SD_SZ_MASK 0xfffff000 |
| #define SD_SG32 0x00000008 |
| #define SD_SZ_ALIGN 0x00000fff |
| |
| #define PCI_CFG_GPIO_SCS 0x10 /* PCI config space bit 4 for 4306c0 slow clock source */ |
| #define PCI_CFG_GPIO_XTAL 0x40 /* PCI config space GPIO 14 for Xtal power-up */ |
| #define PCI_CFG_GPIO_PLL 0x80 /* PCI config space GPIO 15 for PLL power-down */ |
| |
| /* power control defines */ |
| #define PLL_DELAY 150 /* us pll on delay */ |
| #define FREF_DELAY 200 /* us fref change delay */ |
| #define XTAL_ON_DELAY 1000 /* us crystal power-on delay */ |
| |
| /* resetctrl */ |
| #define AIRC_RESET 1 |
| |
| struct aidmp { |
| u32 oobselina30; /* 0x000 */ |
| u32 oobselina74; /* 0x004 */ |
| u32 PAD[6]; |
| u32 oobselinb30; /* 0x020 */ |
| u32 oobselinb74; /* 0x024 */ |
| u32 PAD[6]; |
| u32 oobselinc30; /* 0x040 */ |
| u32 oobselinc74; /* 0x044 */ |
| u32 PAD[6]; |
| u32 oobselind30; /* 0x060 */ |
| u32 oobselind74; /* 0x064 */ |
| u32 PAD[38]; |
| u32 oobselouta30; /* 0x100 */ |
| u32 oobselouta74; /* 0x104 */ |
| u32 PAD[6]; |
| u32 oobseloutb30; /* 0x120 */ |
| u32 oobseloutb74; /* 0x124 */ |
| u32 PAD[6]; |
| u32 oobseloutc30; /* 0x140 */ |
| u32 oobseloutc74; /* 0x144 */ |
| u32 PAD[6]; |
| u32 oobseloutd30; /* 0x160 */ |
| u32 oobseloutd74; /* 0x164 */ |
| u32 PAD[38]; |
| u32 oobsynca; /* 0x200 */ |
| u32 oobseloutaen; /* 0x204 */ |
| u32 PAD[6]; |
| u32 oobsyncb; /* 0x220 */ |
| u32 oobseloutben; /* 0x224 */ |
| u32 PAD[6]; |
| u32 oobsyncc; /* 0x240 */ |
| u32 oobseloutcen; /* 0x244 */ |
| u32 PAD[6]; |
| u32 oobsyncd; /* 0x260 */ |
| u32 oobseloutden; /* 0x264 */ |
| u32 PAD[38]; |
| u32 oobaextwidth; /* 0x300 */ |
| u32 oobainwidth; /* 0x304 */ |
| u32 oobaoutwidth; /* 0x308 */ |
| u32 PAD[5]; |
| u32 oobbextwidth; /* 0x320 */ |
| u32 oobbinwidth; /* 0x324 */ |
| u32 oobboutwidth; /* 0x328 */ |
| u32 PAD[5]; |
| u32 oobcextwidth; /* 0x340 */ |
| u32 oobcinwidth; /* 0x344 */ |
| u32 oobcoutwidth; /* 0x348 */ |
| u32 PAD[5]; |
| u32 oobdextwidth; /* 0x360 */ |
| u32 oobdinwidth; /* 0x364 */ |
| u32 oobdoutwidth; /* 0x368 */ |
| u32 PAD[37]; |
| u32 ioctrlset; /* 0x400 */ |
| u32 ioctrlclear; /* 0x404 */ |
| u32 ioctrl; /* 0x408 */ |
| u32 PAD[61]; |
| u32 iostatus; /* 0x500 */ |
| u32 PAD[127]; |
| u32 ioctrlwidth; /* 0x700 */ |
| u32 iostatuswidth; /* 0x704 */ |
| u32 PAD[62]; |
| u32 resetctrl; /* 0x800 */ |
| u32 resetstatus; /* 0x804 */ |
| u32 resetreadid; /* 0x808 */ |
| u32 resetwriteid; /* 0x80c */ |
| u32 PAD[60]; |
| u32 errlogctrl; /* 0x900 */ |
| u32 errlogdone; /* 0x904 */ |
| u32 errlogstatus; /* 0x908 */ |
| u32 errlogaddrlo; /* 0x90c */ |
| u32 errlogaddrhi; /* 0x910 */ |
| u32 errlogid; /* 0x914 */ |
| u32 errloguser; /* 0x918 */ |
| u32 errlogflags; /* 0x91c */ |
| u32 PAD[56]; |
| u32 intstatus; /* 0xa00 */ |
| u32 PAD[127]; |
| u32 config; /* 0xe00 */ |
| u32 PAD[63]; |
| u32 itcr; /* 0xf00 */ |
| u32 PAD[3]; |
| u32 itipooba; /* 0xf10 */ |
| u32 itipoobb; /* 0xf14 */ |
| u32 itipoobc; /* 0xf18 */ |
| u32 itipoobd; /* 0xf1c */ |
| u32 PAD[4]; |
| u32 itipoobaout; /* 0xf30 */ |
| u32 itipoobbout; /* 0xf34 */ |
| u32 itipoobcout; /* 0xf38 */ |
| u32 itipoobdout; /* 0xf3c */ |
| u32 PAD[4]; |
| u32 itopooba; /* 0xf50 */ |
| u32 itopoobb; /* 0xf54 */ |
| u32 itopoobc; /* 0xf58 */ |
| u32 itopoobd; /* 0xf5c */ |
| u32 PAD[4]; |
| u32 itopoobain; /* 0xf70 */ |
| u32 itopoobbin; /* 0xf74 */ |
| u32 itopoobcin; /* 0xf78 */ |
| u32 itopoobdin; /* 0xf7c */ |
| u32 PAD[4]; |
| u32 itopreset; /* 0xf90 */ |
| u32 PAD[15]; |
| u32 peripherialid4; /* 0xfd0 */ |
| u32 peripherialid5; /* 0xfd4 */ |
| u32 peripherialid6; /* 0xfd8 */ |
| u32 peripherialid7; /* 0xfdc */ |
| u32 peripherialid0; /* 0xfe0 */ |
| u32 peripherialid1; /* 0xfe4 */ |
| u32 peripherialid2; /* 0xfe8 */ |
| u32 peripherialid3; /* 0xfec */ |
| u32 componentid0; /* 0xff0 */ |
| u32 componentid1; /* 0xff4 */ |
| u32 componentid2; /* 0xff8 */ |
| u32 componentid3; /* 0xffc */ |
| }; |
| |
| /* EROM parsing */ |
| |
| static u32 |
| get_erom_ent(struct si_pub *sih, u32 **eromptr, u32 mask, u32 match) |
| { |
| u32 ent; |
| uint inv = 0, nom = 0; |
| |
| while (true) { |
| ent = R_REG(*eromptr); |
| (*eromptr)++; |
| |
| if (mask == 0) |
| break; |
| |
| if ((ent & ER_VALID) == 0) { |
| inv++; |
| continue; |
| } |
| |
| if (ent == (ER_END | ER_VALID)) |
| break; |
| |
| if ((ent & mask) == match) |
| break; |
| |
| nom++; |
| } |
| |
| SI_VMSG(("%s: Returning ent 0x%08x\n", __func__, ent)); |
| if (inv + nom) { |
| SI_VMSG((" after %d invalid and %d non-matching entries\n", |
| inv, nom)); |
| } |
| return ent; |
| } |
| |
| static u32 |
| get_asd(struct si_pub *sih, u32 **eromptr, uint sp, uint ad, uint st, |
| u32 *addrl, u32 *addrh, u32 *sizel, u32 *sizeh) |
| { |
| u32 asd, sz, szd; |
| |
| asd = get_erom_ent(sih, eromptr, ER_VALID, ER_VALID); |
| if (((asd & ER_TAG1) != ER_ADD) || |
| (((asd & AD_SP_MASK) >> AD_SP_SHIFT) != sp) || |
| ((asd & AD_ST_MASK) != st)) { |
| /* This is not what we want, "push" it back */ |
| (*eromptr)--; |
| return 0; |
| } |
| *addrl = asd & AD_ADDR_MASK; |
| if (asd & AD_AG32) |
| *addrh = get_erom_ent(sih, eromptr, 0, 0); |
| else |
| *addrh = 0; |
| *sizeh = 0; |
| sz = asd & AD_SZ_MASK; |
| if (sz == AD_SZ_SZD) { |
| szd = get_erom_ent(sih, eromptr, 0, 0); |
| *sizel = szd & SD_SZ_MASK; |
| if (szd & SD_SG32) |
| *sizeh = get_erom_ent(sih, eromptr, 0, 0); |
| } else |
| *sizel = AD_SZ_BASE << (sz >> AD_SZ_SHIFT); |
| |
| SI_VMSG((" SP %d, ad %d: st = %d, 0x%08x_0x%08x @ 0x%08x_0x%08x\n", |
| sp, ad, st, *sizeh, *sizel, *addrh, *addrl)); |
| |
| return asd; |
| } |
| |
| static void ai_hwfixup(struct si_info *sii) |
| { |
| } |
| |
| /* parse the enumeration rom to identify all cores */ |
| void ai_scan(struct si_pub *sih, void *regs) |
| { |
| struct si_info *sii = SI_INFO(sih); |
| chipcregs_t *cc = (chipcregs_t *) regs; |
| u32 erombase, *eromptr, *eromlim; |
| |
| erombase = R_REG(&cc->eromptr); |
| |
| switch (sih->bustype) { |
| case SI_BUS: |
| eromptr = (u32 *) REG_MAP(erombase, SI_CORE_SIZE); |
| break; |
| |
| case PCI_BUS: |
| /* Set wrappers address */ |
| sii->curwrap = (void *)((unsigned long)regs + SI_CORE_SIZE); |
| |
| /* Now point the window at the erom */ |
| pci_write_config_dword(sii->pbus, PCI_BAR0_WIN, erombase); |
| eromptr = regs; |
| break; |
| |
| case SPI_BUS: |
| case SDIO_BUS: |
| eromptr = (u32 *)(unsigned long)erombase; |
| break; |
| |
| default: |
| SI_ERROR(("Don't know how to do AXI enumertion on bus %d\n", |
| sih->bustype)); |
| return; |
| } |
| eromlim = eromptr + (ER_REMAPCONTROL / sizeof(u32)); |
| |
| SI_VMSG(("ai_scan: regs = 0x%p, erombase = 0x%08x, eromptr = 0x%p, eromlim = 0x%p\n", regs, erombase, eromptr, eromlim)); |
| while (eromptr < eromlim) { |
| u32 cia, cib, cid, mfg, crev, nmw, nsw, nmp, nsp; |
| u32 mpd, asd, addrl, addrh, sizel, sizeh; |
| u32 *base; |
| uint i, j, idx; |
| bool br; |
| |
| br = false; |
| |
| /* Grok a component */ |
| cia = get_erom_ent(sih, &eromptr, ER_TAG, ER_CI); |
| if (cia == (ER_END | ER_VALID)) { |
| SI_VMSG(("Found END of erom after %d cores\n", |
| sii->numcores)); |
| ai_hwfixup(sii); |
| return; |
| } |
| base = eromptr - 1; |
| cib = get_erom_ent(sih, &eromptr, 0, 0); |
| |
| if ((cib & ER_TAG) != ER_CI) { |
| SI_ERROR(("CIA not followed by CIB\n")); |
| goto error; |
| } |
| |
| cid = (cia & CIA_CID_MASK) >> CIA_CID_SHIFT; |
| mfg = (cia & CIA_MFG_MASK) >> CIA_MFG_SHIFT; |
| crev = (cib & CIB_REV_MASK) >> CIB_REV_SHIFT; |
| nmw = (cib & CIB_NMW_MASK) >> CIB_NMW_SHIFT; |
| nsw = (cib & CIB_NSW_MASK) >> CIB_NSW_SHIFT; |
| nmp = (cib & CIB_NMP_MASK) >> CIB_NMP_SHIFT; |
| nsp = (cib & CIB_NSP_MASK) >> CIB_NSP_SHIFT; |
| |
| SI_VMSG(("Found component 0x%04x/0x%04x rev %d at erom addr 0x%p, with nmw = %d, " "nsw = %d, nmp = %d & nsp = %d\n", mfg, cid, crev, base, nmw, nsw, nmp, nsp)); |
| |
| if (((mfg == MFGID_ARM) && (cid == DEF_AI_COMP)) || (nsp == 0)) |
| continue; |
| if ((nmw + nsw == 0)) { |
| /* A component which is not a core */ |
| if (cid == OOB_ROUTER_CORE_ID) { |
| asd = get_asd(sih, &eromptr, 0, 0, AD_ST_SLAVE, |
| &addrl, &addrh, &sizel, &sizeh); |
| if (asd != 0) { |
| sii->oob_router = addrl; |
| } |
| } |
| continue; |
| } |
| |
| idx = sii->numcores; |
| /* sii->eromptr[idx] = base; */ |
| sii->cia[idx] = cia; |
| sii->cib[idx] = cib; |
| sii->coreid[idx] = cid; |
| |
| for (i = 0; i < nmp; i++) { |
| mpd = get_erom_ent(sih, &eromptr, ER_VALID, ER_VALID); |
| if ((mpd & ER_TAG) != ER_MP) { |
| SI_ERROR(("Not enough MP entries for component 0x%x\n", cid)); |
| goto error; |
| } |
| SI_VMSG((" Master port %d, mp: %d id: %d\n", i, |
| (mpd & MPD_MP_MASK) >> MPD_MP_SHIFT, |
| (mpd & MPD_MUI_MASK) >> MPD_MUI_SHIFT)); |
| } |
| |
| /* First Slave Address Descriptor should be port 0: |
| * the main register space for the core |
| */ |
| asd = |
| get_asd(sih, &eromptr, 0, 0, AD_ST_SLAVE, &addrl, &addrh, |
| &sizel, &sizeh); |
| if (asd == 0) { |
| /* Try again to see if it is a bridge */ |
| asd = |
| get_asd(sih, &eromptr, 0, 0, AD_ST_BRIDGE, &addrl, |
| &addrh, &sizel, &sizeh); |
| if (asd != 0) |
| br = true; |
| else if ((addrh != 0) || (sizeh != 0) |
| || (sizel != SI_CORE_SIZE)) { |
| SI_ERROR(("First Slave ASD for core 0x%04x malformed " "(0x%08x)\n", cid, asd)); |
| goto error; |
| } |
| } |
| sii->coresba[idx] = addrl; |
| sii->coresba_size[idx] = sizel; |
| /* Get any more ASDs in port 0 */ |
| j = 1; |
| do { |
| asd = |
| get_asd(sih, &eromptr, 0, j, AD_ST_SLAVE, &addrl, |
| &addrh, &sizel, &sizeh); |
| if ((asd != 0) && (j == 1) && (sizel == SI_CORE_SIZE)) { |
| sii->coresba2[idx] = addrl; |
| sii->coresba2_size[idx] = sizel; |
| } |
| j++; |
| } while (asd != 0); |
| |
| /* Go through the ASDs for other slave ports */ |
| for (i = 1; i < nsp; i++) { |
| j = 0; |
| do { |
| asd = |
| get_asd(sih, &eromptr, i, j++, AD_ST_SLAVE, |
| &addrl, &addrh, &sizel, &sizeh); |
| } while (asd != 0); |
| if (j == 0) { |
| SI_ERROR((" SP %d has no address descriptors\n", |
| i)); |
| goto error; |
| } |
| } |
| |
| /* Now get master wrappers */ |
| for (i = 0; i < nmw; i++) { |
| asd = |
| get_asd(sih, &eromptr, i, 0, AD_ST_MWRAP, &addrl, |
| &addrh, &sizel, &sizeh); |
| if (asd == 0) { |
| SI_ERROR(("Missing descriptor for MW %d\n", i)); |
| goto error; |
| } |
| if ((sizeh != 0) || (sizel != SI_CORE_SIZE)) { |
| SI_ERROR(("Master wrapper %d is not 4KB\n", i)); |
| goto error; |
| } |
| if (i == 0) |
| sii->wrapba[idx] = addrl; |
| } |
| |
| /* And finally slave wrappers */ |
| for (i = 0; i < nsw; i++) { |
| uint fwp = (nsp == 1) ? 0 : 1; |
| asd = |
| get_asd(sih, &eromptr, fwp + i, 0, AD_ST_SWRAP, |
| &addrl, &addrh, &sizel, &sizeh); |
| if (asd == 0) { |
| SI_ERROR(("Missing descriptor for SW %d\n", i)); |
| goto error; |
| } |
| if ((sizeh != 0) || (sizel != SI_CORE_SIZE)) { |
| SI_ERROR(("Slave wrapper %d is not 4KB\n", i)); |
| goto error; |
| } |
| if ((nmw == 0) && (i == 0)) |
| sii->wrapba[idx] = addrl; |
| } |
| |
| /* Don't record bridges */ |
| if (br) |
| continue; |
| |
| /* Done with core */ |
| sii->numcores++; |
| } |
| |
| SI_ERROR(("Reached end of erom without finding END")); |
| |
| error: |
| sii->numcores = 0; |
| return; |
| } |
| |
| /* This function changes the logical "focus" to the indicated core. |
| * Return the current core's virtual address. |
| */ |
| void *ai_setcoreidx(struct si_pub *sih, uint coreidx) |
| { |
| struct si_info *sii = SI_INFO(sih); |
| u32 addr = sii->coresba[coreidx]; |
| u32 wrap = sii->wrapba[coreidx]; |
| void *regs; |
| |
| if (coreidx >= sii->numcores) |
| return NULL; |
| |
| switch (sih->bustype) { |
| case SI_BUS: |
| /* map new one */ |
| if (!sii->regs[coreidx]) { |
| sii->regs[coreidx] = REG_MAP(addr, SI_CORE_SIZE); |
| } |
| sii->curmap = regs = sii->regs[coreidx]; |
| if (!sii->wrappers[coreidx]) { |
| sii->wrappers[coreidx] = REG_MAP(wrap, SI_CORE_SIZE); |
| } |
| sii->curwrap = sii->wrappers[coreidx]; |
| break; |
| |
| case PCI_BUS: |
| /* point bar0 window */ |
| pci_write_config_dword(sii->pbus, PCI_BAR0_WIN, addr); |
| regs = sii->curmap; |
| /* point bar0 2nd 4KB window */ |
| pci_write_config_dword(sii->pbus, PCI_BAR0_WIN2, wrap); |
| break; |
| |
| case SPI_BUS: |
| case SDIO_BUS: |
| sii->curmap = regs = (void *)(unsigned long)addr; |
| sii->curwrap = (void *)(unsigned long)wrap; |
| break; |
| |
| default: |
| regs = NULL; |
| break; |
| } |
| |
| sii->curmap = regs; |
| sii->curidx = coreidx; |
| |
| return regs; |
| } |
| |
| /* Return the number of address spaces in current core */ |
| int ai_numaddrspaces(struct si_pub *sih) |
| { |
| return 2; |
| } |
| |
| /* Return the address of the nth address space in the current core */ |
| u32 ai_addrspace(struct si_pub *sih, uint asidx) |
| { |
| struct si_info *sii; |
| uint cidx; |
| |
| sii = SI_INFO(sih); |
| cidx = sii->curidx; |
| |
| if (asidx == 0) |
| return sii->coresba[cidx]; |
| else if (asidx == 1) |
| return sii->coresba2[cidx]; |
| else { |
| SI_ERROR(("%s: Need to parse the erom again to find addr space %d\n", __func__, asidx)); |
| return 0; |
| } |
| } |
| |
| /* Return the size of the nth address space in the current core */ |
| u32 ai_addrspacesize(struct si_pub *sih, uint asidx) |
| { |
| struct si_info *sii; |
| uint cidx; |
| |
| sii = SI_INFO(sih); |
| cidx = sii->curidx; |
| |
| if (asidx == 0) |
| return sii->coresba_size[cidx]; |
| else if (asidx == 1) |
| return sii->coresba2_size[cidx]; |
| else { |
| SI_ERROR(("%s: Need to parse the erom again to find addr space %d\n", __func__, asidx)); |
| return 0; |
| } |
| } |
| |
| uint ai_flag(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| struct aidmp *ai; |
| |
| sii = SI_INFO(sih); |
| if (BCM47162_DMP()) { |
| SI_ERROR(("%s: Attempting to read MIPS DMP registers on 47162a0", __func__)); |
| return sii->curidx; |
| } |
| ai = sii->curwrap; |
| |
| return R_REG(&ai->oobselouta30) & 0x1f; |
| } |
| |
| void ai_setint(struct si_pub *sih, int siflag) |
| { |
| } |
| |
| uint ai_corevendor(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| u32 cia; |
| |
| sii = SI_INFO(sih); |
| cia = sii->cia[sii->curidx]; |
| return (cia & CIA_MFG_MASK) >> CIA_MFG_SHIFT; |
| } |
| |
| uint ai_corerev(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| u32 cib; |
| |
| sii = SI_INFO(sih); |
| cib = sii->cib[sii->curidx]; |
| return (cib & CIB_REV_MASK) >> CIB_REV_SHIFT; |
| } |
| |
| bool ai_iscoreup(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| struct aidmp *ai; |
| |
| sii = SI_INFO(sih); |
| ai = sii->curwrap; |
| |
| return (((R_REG(&ai->ioctrl) & (SICF_FGC | SICF_CLOCK_EN)) == |
| SICF_CLOCK_EN) |
| && ((R_REG(&ai->resetctrl) & AIRC_RESET) == 0)); |
| } |
| |
| void ai_core_cflags_wo(struct si_pub *sih, u32 mask, u32 val) |
| { |
| struct si_info *sii; |
| struct aidmp *ai; |
| u32 w; |
| |
| sii = SI_INFO(sih); |
| |
| if (BCM47162_DMP()) { |
| SI_ERROR(("%s: Accessing MIPS DMP register (ioctrl) on 47162a0", |
| __func__)); |
| return; |
| } |
| |
| ai = sii->curwrap; |
| |
| if (mask || val) { |
| w = ((R_REG(&ai->ioctrl) & ~mask) | val); |
| W_REG(&ai->ioctrl, w); |
| } |
| } |
| |
| u32 ai_core_cflags(struct si_pub *sih, u32 mask, u32 val) |
| { |
| struct si_info *sii; |
| struct aidmp *ai; |
| u32 w; |
| |
| sii = SI_INFO(sih); |
| if (BCM47162_DMP()) { |
| SI_ERROR(("%s: Accessing MIPS DMP register (ioctrl) on 47162a0", |
| __func__)); |
| return 0; |
| } |
| |
| ai = sii->curwrap; |
| |
| if (mask || val) { |
| w = ((R_REG(&ai->ioctrl) & ~mask) | val); |
| W_REG(&ai->ioctrl, w); |
| } |
| |
| return R_REG(&ai->ioctrl); |
| } |
| |
| u32 ai_core_sflags(struct si_pub *sih, u32 mask, u32 val) |
| { |
| struct si_info *sii; |
| struct aidmp *ai; |
| u32 w; |
| |
| sii = SI_INFO(sih); |
| if (BCM47162_DMP()) { |
| SI_ERROR(("%s: Accessing MIPS DMP register (iostatus) on 47162a0", __func__)); |
| return 0; |
| } |
| |
| ai = sii->curwrap; |
| |
| if (mask || val) { |
| w = ((R_REG(&ai->iostatus) & ~mask) | val); |
| W_REG(&ai->iostatus, w); |
| } |
| |
| return R_REG(&ai->iostatus); |
| } |
| |
| /* *************** from siutils.c ************** */ |
| /* local prototypes */ |
| static struct si_info *ai_doattach(struct si_info *sii, void *regs, |
| uint bustype, void *sdh, char **vars, |
| uint *varsz); |
| static bool ai_buscore_prep(struct si_info *sii, uint bustype); |
| static bool ai_buscore_setup(struct si_info *sii, chipcregs_t *cc, uint bustype, |
| u32 savewin, uint *origidx, void *regs); |
| static void ai_nvram_process(struct si_info *sii, char *pvars); |
| |
| /* dev path concatenation util */ |
| static char *ai_devpathvar(struct si_pub *sih, char *var, int len, |
| const char *name); |
| static bool _ai_clkctl_cc(struct si_info *sii, uint mode); |
| static bool ai_ispcie(struct si_info *sii); |
| |
| /* global variable to indicate reservation/release of gpio's */ |
| static u32 ai_gpioreservation; |
| |
| /* |
| * Allocate a si handle. |
| * devid - pci device id (used to determine chip#) |
| * osh - opaque OS handle |
| * regs - virtual address of initial core registers |
| * bustype - pci/sb/sdio/etc |
| * vars - pointer to a pointer area for "environment" variables |
| * varsz - pointer to int to return the size of the vars |
| */ |
| struct si_pub *ai_attach(void *regs, uint bustype, |
| void *sdh, char **vars, uint *varsz) |
| { |
| struct si_info *sii; |
| |
| /* alloc struct si_info */ |
| sii = kmalloc(sizeof(struct si_info), GFP_ATOMIC); |
| if (sii == NULL) { |
| SI_ERROR(("si_attach: malloc failed!\n")); |
| return NULL; |
| } |
| |
| if (ai_doattach(sii, regs, bustype, sdh, vars, varsz) == |
| NULL) { |
| kfree(sii); |
| return NULL; |
| } |
| sii->vars = vars ? *vars : NULL; |
| sii->varsz = varsz ? *varsz : 0; |
| |
| return (struct si_pub *) sii; |
| } |
| |
| /* global kernel resource */ |
| static struct si_info ksii; |
| |
| static bool ai_buscore_prep(struct si_info *sii, uint bustype) |
| { |
| /* kludge to enable the clock on the 4306 which lacks a slowclock */ |
| if (bustype == PCI_BUS && !ai_ispcie(sii)) |
| ai_clkctl_xtal(&sii->pub, XTAL | PLL, ON); |
| return true; |
| } |
| |
| static bool ai_buscore_setup(struct si_info *sii, chipcregs_t *cc, uint bustype, |
| u32 savewin, uint *origidx, void *regs) |
| { |
| bool pci, pcie; |
| uint i; |
| uint pciidx, pcieidx, pcirev, pcierev; |
| |
| cc = ai_setcoreidx(&sii->pub, SI_CC_IDX); |
| |
| /* get chipcommon rev */ |
| sii->pub.ccrev = (int)ai_corerev(&sii->pub); |
| |
| /* get chipcommon chipstatus */ |
| if (sii->pub.ccrev >= 11) |
| sii->pub.chipst = R_REG(&cc->chipstatus); |
| |
| /* get chipcommon capabilites */ |
| sii->pub.cccaps = R_REG(&cc->capabilities); |
| /* get chipcommon extended capabilities */ |
| |
| if (sii->pub.ccrev >= 35) |
| sii->pub.cccaps_ext = R_REG(&cc->capabilities_ext); |
| |
| /* get pmu rev and caps */ |
| if (sii->pub.cccaps & CC_CAP_PMU) { |
| sii->pub.pmucaps = R_REG(&cc->pmucapabilities); |
| sii->pub.pmurev = sii->pub.pmucaps & PCAP_REV_MASK; |
| } |
| |
| /* figure out bus/orignal core idx */ |
| sii->pub.buscoretype = NODEV_CORE_ID; |
| sii->pub.buscorerev = NOREV; |
| sii->pub.buscoreidx = BADIDX; |
| |
| pci = pcie = false; |
| pcirev = pcierev = NOREV; |
| pciidx = pcieidx = BADIDX; |
| |
| for (i = 0; i < sii->numcores; i++) { |
| uint cid, crev; |
| |
| ai_setcoreidx(&sii->pub, i); |
| cid = ai_coreid(&sii->pub); |
| crev = ai_corerev(&sii->pub); |
| |
| /* Display cores found */ |
| SI_VMSG(("CORE[%d]: id 0x%x rev %d base 0x%x regs 0x%p\n", |
| i, cid, crev, sii->coresba[i], sii->regs[i])); |
| |
| if (bustype == PCI_BUS) { |
| if (cid == PCI_CORE_ID) { |
| pciidx = i; |
| pcirev = crev; |
| pci = true; |
| } else if (cid == PCIE_CORE_ID) { |
| pcieidx = i; |
| pcierev = crev; |
| pcie = true; |
| } |
| } |
| |
| /* find the core idx before entering this func. */ |
| if ((savewin && (savewin == sii->coresba[i])) || |
| (regs == sii->regs[i])) |
| *origidx = i; |
| } |
| |
| if (pci && pcie) { |
| if (ai_ispcie(sii)) |
| pci = false; |
| else |
| pcie = false; |
| } |
| if (pci) { |
| sii->pub.buscoretype = PCI_CORE_ID; |
| sii->pub.buscorerev = pcirev; |
| sii->pub.buscoreidx = pciidx; |
| } else if (pcie) { |
| sii->pub.buscoretype = PCIE_CORE_ID; |
| sii->pub.buscorerev = pcierev; |
| sii->pub.buscoreidx = pcieidx; |
| } |
| |
| SI_VMSG(("Buscore id/type/rev %d/0x%x/%d\n", sii->pub.buscoreidx, |
| sii->pub.buscoretype, sii->pub.buscorerev)); |
| |
| /* fixup necessary chip/core configurations */ |
| if (sii->pub.bustype == PCI_BUS) { |
| if (SI_FAST(sii)) { |
| if (!sii->pch) { |
| sii->pch = (void *)pcicore_init( |
| &sii->pub, sii->pbus, |
| (void *)PCIEREGS(sii)); |
| if (sii->pch == NULL) |
| return false; |
| } |
| } |
| if (ai_pci_fixcfg(&sii->pub)) { |
| SI_ERROR(("si_doattach: si_pci_fixcfg failed\n")); |
| return false; |
| } |
| } |
| |
| /* return to the original core */ |
| ai_setcoreidx(&sii->pub, *origidx); |
| |
| return true; |
| } |
| |
| static __used void ai_nvram_process(struct si_info *sii, char *pvars) |
| { |
| uint w = 0; |
| |
| /* get boardtype and boardrev */ |
| switch (sii->pub.bustype) { |
| case PCI_BUS: |
| /* do a pci config read to get subsystem id and subvendor id */ |
| pci_read_config_dword(sii->pbus, PCI_SUBSYSTEM_VENDOR_ID, &w); |
| /* Let nvram variables override subsystem Vend/ID */ |
| sii->pub.boardvendor = (u16)ai_getdevpathintvar(&sii->pub, |
| "boardvendor"); |
| if (sii->pub.boardvendor == 0) |
| sii->pub.boardvendor = w & 0xffff; |
| else |
| SI_ERROR(("Overriding boardvendor: 0x%x instead of " |
| "0x%x\n", sii->pub.boardvendor, w & 0xffff)); |
| sii->pub.boardtype = (u16)ai_getdevpathintvar(&sii->pub, |
| "boardtype"); |
| if (sii->pub.boardtype == 0) |
| sii->pub.boardtype = (w >> 16) & 0xffff; |
| else |
| SI_ERROR(("Overriding boardtype: 0x%x instead of 0x%x\n" |
| , sii->pub.boardtype, (w >> 16) & 0xffff)); |
| break; |
| |
| sii->pub.boardvendor = getintvar(pvars, "manfid"); |
| sii->pub.boardtype = getintvar(pvars, "prodid"); |
| break; |
| |
| case SI_BUS: |
| case JTAG_BUS: |
| sii->pub.boardvendor = PCI_VENDOR_ID_BROADCOM; |
| sii->pub.boardtype = getintvar(pvars, "prodid"); |
| if (pvars == NULL || (sii->pub.boardtype == 0)) { |
| sii->pub.boardtype = getintvar(NULL, "boardtype"); |
| if (sii->pub.boardtype == 0) |
| sii->pub.boardtype = 0xffff; |
| } |
| break; |
| } |
| |
| if (sii->pub.boardtype == 0) { |
| SI_ERROR(("si_doattach: unknown board type\n")); |
| } |
| |
| sii->pub.boardflags = getintvar(pvars, "boardflags"); |
| } |
| |
| static struct si_info *ai_doattach(struct si_info *sii, |
| void *regs, uint bustype, void *pbus, |
| char **vars, uint *varsz) |
| { |
| struct si_pub *sih = &sii->pub; |
| u32 w, savewin; |
| chipcregs_t *cc; |
| char *pvars = NULL; |
| uint socitype; |
| uint origidx; |
| |
| memset((unsigned char *) sii, 0, sizeof(struct si_info)); |
| |
| savewin = 0; |
| |
| sih->buscoreidx = BADIDX; |
| |
| sii->curmap = regs; |
| sii->pbus = pbus; |
| |
| /* check to see if we are a si core mimic'ing a pci core */ |
| if (bustype == PCI_BUS) { |
| pci_read_config_dword(sii->pbus, PCI_SPROM_CONTROL, &w); |
| if (w == 0xffffffff) { |
| SI_ERROR(("%s: incoming bus is PCI but it's a lie, " |
| " switching to SI devid:0x%x\n", |
| __func__, devid)); |
| bustype = SI_BUS; |
| } |
| } |
| |
| /* find Chipcommon address */ |
| if (bustype == PCI_BUS) { |
| pci_read_config_dword(sii->pbus, PCI_BAR0_WIN, &savewin); |
| if (!GOODCOREADDR(savewin, SI_ENUM_BASE)) |
| savewin = SI_ENUM_BASE; |
| pci_write_config_dword(sii->pbus, PCI_BAR0_WIN, |
| SI_ENUM_BASE); |
| cc = (chipcregs_t *) regs; |
| } else { |
| cc = (chipcregs_t *) REG_MAP(SI_ENUM_BASE, SI_CORE_SIZE); |
| } |
| |
| sih->bustype = bustype; |
| |
| /* bus/core/clk setup for register access */ |
| if (!ai_buscore_prep(sii, bustype)) { |
| SI_ERROR(("si_doattach: si_core_clk_prep failed %d\n", |
| bustype)); |
| return NULL; |
| } |
| |
| /* |
| * ChipID recognition. |
| * We assume we can read chipid at offset 0 from the regs arg. |
| * If we add other chiptypes (or if we need to support old sdio |
| * hosts w/o chipcommon), some way of recognizing them needs to |
| * be added here. |
| */ |
| w = R_REG(&cc->chipid); |
| socitype = (w & CID_TYPE_MASK) >> CID_TYPE_SHIFT; |
| /* Might as wll fill in chip id rev & pkg */ |
| sih->chip = w & CID_ID_MASK; |
| sih->chiprev = (w & CID_REV_MASK) >> CID_REV_SHIFT; |
| sih->chippkg = (w & CID_PKG_MASK) >> CID_PKG_SHIFT; |
| |
| sih->issim = IS_SIM(sih->chippkg); |
| |
| /* scan for cores */ |
| if (socitype == SOCI_AI) { |
| SI_MSG(("Found chip type AI (0x%08x)\n", w)); |
| /* pass chipc address instead of original core base */ |
| ai_scan(&sii->pub, (void *)cc); |
| } else { |
| SI_ERROR(("Found chip of unknown type (0x%08x)\n", w)); |
| return NULL; |
| } |
| /* no cores found, bail out */ |
| if (sii->numcores == 0) { |
| SI_ERROR(("si_doattach: could not find any cores\n")); |
| return NULL; |
| } |
| /* bus/core/clk setup */ |
| origidx = SI_CC_IDX; |
| if (!ai_buscore_setup(sii, cc, bustype, savewin, &origidx, regs)) { |
| SI_ERROR(("si_doattach: si_buscore_setup failed\n")); |
| goto exit; |
| } |
| |
| /* Init nvram from sprom/otp if they exist */ |
| if (srom_var_init |
| (&sii->pub, bustype, regs, vars, varsz)) { |
| SI_ERROR(("si_doattach: srom_var_init failed: bad srom\n")); |
| goto exit; |
| } |
| pvars = vars ? *vars : NULL; |
| ai_nvram_process(sii, pvars); |
| |
| /* === NVRAM, clock is ready === */ |
| cc = (chipcregs_t *) ai_setcore(sih, CC_CORE_ID, 0); |
| W_REG(&cc->gpiopullup, 0); |
| W_REG(&cc->gpiopulldown, 0); |
| ai_setcoreidx(sih, origidx); |
| |
| /* PMU specific initializations */ |
| if (PMUCTL_ENAB(sih)) { |
| u32 xtalfreq; |
| si_pmu_init(sih); |
| si_pmu_chip_init(sih); |
| xtalfreq = getintvar(pvars, "xtalfreq"); |
| /* If xtalfreq var not available, try to measure it */ |
| if (xtalfreq == 0) |
| xtalfreq = si_pmu_measure_alpclk(sih); |
| si_pmu_pll_init(sih, xtalfreq); |
| si_pmu_res_init(sih); |
| si_pmu_swreg_init(sih); |
| } |
| |
| /* setup the GPIO based LED powersave register */ |
| w = getintvar(pvars, "leddc"); |
| if (w == 0) |
| w = DEFAULT_GPIOTIMERVAL; |
| ai_corereg(sih, SI_CC_IDX, offsetof(chipcregs_t, gpiotimerval), ~0, w); |
| |
| if (PCIE(sii)) { |
| pcicore_attach(sii->pch, pvars, SI_DOATTACH); |
| } |
| |
| if (sih->chip == BCM43224_CHIP_ID) { |
| /* |
| * enable 12 mA drive strenth for 43224 and |
| * set chipControl register bit 15 |
| */ |
| if (sih->chiprev == 0) { |
| SI_MSG(("Applying 43224A0 WARs\n")); |
| ai_corereg(sih, SI_CC_IDX, |
| offsetof(chipcregs_t, chipcontrol), |
| CCTRL43224_GPIO_TOGGLE, |
| CCTRL43224_GPIO_TOGGLE); |
| si_pmu_chipcontrol(sih, 0, CCTRL_43224A0_12MA_LED_DRIVE, |
| CCTRL_43224A0_12MA_LED_DRIVE); |
| } |
| if (sih->chiprev >= 1) { |
| SI_MSG(("Applying 43224B0+ WARs\n")); |
| si_pmu_chipcontrol(sih, 0, CCTRL_43224B0_12MA_LED_DRIVE, |
| CCTRL_43224B0_12MA_LED_DRIVE); |
| } |
| } |
| |
| if (sih->chip == BCM4313_CHIP_ID) { |
| /* |
| * enable 12 mA drive strenth for 4313 and |
| * set chipControl register bit 1 |
| */ |
| SI_MSG(("Applying 4313 WARs\n")); |
| si_pmu_chipcontrol(sih, 0, CCTRL_4313_12MA_LED_DRIVE, |
| CCTRL_4313_12MA_LED_DRIVE); |
| } |
| |
| return sii; |
| exit: |
| if (sih->bustype == PCI_BUS) { |
| if (sii->pch) |
| pcicore_deinit(sii->pch); |
| sii->pch = NULL; |
| } |
| |
| return NULL; |
| } |
| |
| /* may be called with core in reset */ |
| void ai_detach(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| uint idx; |
| |
| struct si_pub *si_local = NULL; |
| memcpy(&si_local, &sih, sizeof(struct si_pub **)); |
| |
| sii = SI_INFO(sih); |
| |
| if (sii == NULL) |
| return; |
| |
| if (sih->bustype == SI_BUS) |
| for (idx = 0; idx < SI_MAXCORES; idx++) |
| if (sii->regs[idx]) { |
| iounmap(sii->regs[idx]); |
| sii->regs[idx] = NULL; |
| } |
| |
| if (sih->bustype == PCI_BUS) { |
| if (sii->pch) |
| pcicore_deinit(sii->pch); |
| sii->pch = NULL; |
| } |
| |
| if (sii != &ksii) |
| kfree(sii); |
| } |
| |
| /* register driver interrupt disabling and restoring callback functions */ |
| void |
| ai_register_intr_callback(struct si_pub *sih, void *intrsoff_fn, |
| void *intrsrestore_fn, |
| void *intrsenabled_fn, void *intr_arg) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| sii->intr_arg = intr_arg; |
| sii->intrsoff_fn = (si_intrsoff_t) intrsoff_fn; |
| sii->intrsrestore_fn = (si_intrsrestore_t) intrsrestore_fn; |
| sii->intrsenabled_fn = (si_intrsenabled_t) intrsenabled_fn; |
| /* save current core id. when this function called, the current core |
| * must be the core which provides driver functions(il, et, wl, etc.) |
| */ |
| sii->dev_coreid = sii->coreid[sii->curidx]; |
| } |
| |
| void ai_deregister_intr_callback(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| sii->intrsoff_fn = NULL; |
| } |
| |
| uint ai_coreid(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| return sii->coreid[sii->curidx]; |
| } |
| |
| uint ai_coreidx(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| return sii->curidx; |
| } |
| |
| bool ai_backplane64(struct si_pub *sih) |
| { |
| return (sih->cccaps & CC_CAP_BKPLN64) != 0; |
| } |
| |
| /* return index of coreid or BADIDX if not found */ |
| uint ai_findcoreidx(struct si_pub *sih, uint coreid, uint coreunit) |
| { |
| struct si_info *sii; |
| uint found; |
| uint i; |
| |
| sii = SI_INFO(sih); |
| |
| found = 0; |
| |
| for (i = 0; i < sii->numcores; i++) |
| if (sii->coreid[i] == coreid) { |
| if (found == coreunit) |
| return i; |
| found++; |
| } |
| |
| return BADIDX; |
| } |
| |
| /* |
| * This function changes logical "focus" to the indicated core; |
| * must be called with interrupts off. |
| * Moreover, callers should keep interrupts off during switching |
| * out of and back to d11 core. |
| */ |
| void *ai_setcore(struct si_pub *sih, uint coreid, uint coreunit) |
| { |
| uint idx; |
| |
| idx = ai_findcoreidx(sih, coreid, coreunit); |
| if (!GOODIDX(idx)) |
| return NULL; |
| |
| return ai_setcoreidx(sih, idx); |
| } |
| |
| /* Turn off interrupt as required by ai_setcore, before switch core */ |
| void *ai_switch_core(struct si_pub *sih, uint coreid, uint *origidx, |
| uint *intr_val) |
| { |
| void *cc; |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| |
| if (SI_FAST(sii)) { |
| /* Overloading the origidx variable to remember the coreid, |
| * this works because the core ids cannot be confused with |
| * core indices. |
| */ |
| *origidx = coreid; |
| if (coreid == CC_CORE_ID) |
| return (void *)CCREGS_FAST(sii); |
| else if (coreid == sih->buscoretype) |
| return (void *)PCIEREGS(sii); |
| } |
| INTR_OFF(sii, *intr_val); |
| *origidx = sii->curidx; |
| cc = ai_setcore(sih, coreid, 0); |
| return cc; |
| } |
| |
| /* restore coreidx and restore interrupt */ |
| void ai_restore_core(struct si_pub *sih, uint coreid, uint intr_val) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| if (SI_FAST(sii) |
| && ((coreid == CC_CORE_ID) || (coreid == sih->buscoretype))) |
| return; |
| |
| ai_setcoreidx(sih, coreid); |
| INTR_RESTORE(sii, intr_val); |
| } |
| |
| void ai_write_wrapperreg(struct si_pub *sih, u32 offset, u32 val) |
| { |
| struct si_info *sii = SI_INFO(sih); |
| u32 *w = (u32 *) sii->curwrap; |
| W_REG(w + (offset / 4), val); |
| return; |
| } |
| |
| /* |
| * Switch to 'coreidx', issue a single arbitrary 32bit register mask&set |
| * operation, switch back to the original core, and return the new value. |
| * |
| * When using the silicon backplane, no fiddling with interrupts or core |
| * switches is needed. |
| * |
| * Also, when using pci/pcie, we can optimize away the core switching for pci |
| * registers and (on newer pci cores) chipcommon registers. |
| */ |
| uint ai_corereg(struct si_pub *sih, uint coreidx, uint regoff, uint mask, |
| uint val) |
| { |
| uint origidx = 0; |
| u32 *r = NULL; |
| uint w; |
| uint intr_val = 0; |
| bool fast = false; |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| |
| if (coreidx >= SI_MAXCORES) |
| return 0; |
| |
| if (sih->bustype == SI_BUS) { |
| /* If internal bus, we can always get at everything */ |
| fast = true; |
| /* map if does not exist */ |
| if (!sii->regs[coreidx]) { |
| sii->regs[coreidx] = REG_MAP(sii->coresba[coreidx], |
| SI_CORE_SIZE); |
| } |
| r = (u32 *) ((unsigned char *) sii->regs[coreidx] + regoff); |
| } else if (sih->bustype == PCI_BUS) { |
| /* |
| * If pci/pcie, we can get at pci/pcie regs |
| * and on newer cores to chipc |
| */ |
| if ((sii->coreid[coreidx] == CC_CORE_ID) && SI_FAST(sii)) { |
| /* Chipc registers are mapped at 12KB */ |
| |
| fast = true; |
| r = (u32 *) ((char *)sii->curmap + |
| PCI_16KB0_CCREGS_OFFSET + regoff); |
| } else if (sii->pub.buscoreidx == coreidx) { |
| /* |
| * pci registers are at either in the last 2KB of |
| * an 8KB window or, in pcie and pci rev 13 at 8KB |
| */ |
| fast = true; |
| if (SI_FAST(sii)) |
| r = (u32 *) ((char *)sii->curmap + |
| PCI_16KB0_PCIREGS_OFFSET + |
| regoff); |
| else |
| r = (u32 *) ((char *)sii->curmap + |
| ((regoff >= SBCONFIGOFF) ? |
| PCI_BAR0_PCISBR_OFFSET : |
| PCI_BAR0_PCIREGS_OFFSET) + |
| regoff); |
| } |
| } |
| |
| if (!fast) { |
| INTR_OFF(sii, intr_val); |
| |
| /* save current core index */ |
| origidx = ai_coreidx(&sii->pub); |
| |
| /* switch core */ |
| r = (u32 *) ((unsigned char *) ai_setcoreidx(&sii->pub, coreidx) |
| + regoff); |
| } |
| |
| /* mask and set */ |
| if (mask || val) { |
| w = (R_REG(r) & ~mask) | val; |
| W_REG(r, w); |
| } |
| |
| /* readback */ |
| w = R_REG(r); |
| |
| if (!fast) { |
| /* restore core index */ |
| if (origidx != coreidx) |
| ai_setcoreidx(&sii->pub, origidx); |
| |
| INTR_RESTORE(sii, intr_val); |
| } |
| |
| return w; |
| } |
| |
| void ai_core_disable(struct si_pub *sih, u32 bits) |
| { |
| struct si_info *sii; |
| u32 dummy; |
| struct aidmp *ai; |
| |
| sii = SI_INFO(sih); |
| |
| ai = sii->curwrap; |
| |
| /* if core is already in reset, just return */ |
| if (R_REG(&ai->resetctrl) & AIRC_RESET) |
| return; |
| |
| W_REG(&ai->ioctrl, bits); |
| dummy = R_REG(&ai->ioctrl); |
| udelay(10); |
| |
| W_REG(&ai->resetctrl, AIRC_RESET); |
| udelay(1); |
| } |
| |
| /* reset and re-enable a core |
| * inputs: |
| * bits - core specific bits that are set during and after reset sequence |
| * resetbits - core specific bits that are set only during reset sequence |
| */ |
| void ai_core_reset(struct si_pub *sih, u32 bits, u32 resetbits) |
| { |
| struct si_info *sii; |
| struct aidmp *ai; |
| u32 dummy; |
| |
| sii = SI_INFO(sih); |
| ai = sii->curwrap; |
| |
| /* |
| * Must do the disable sequence first to work |
| * for arbitrary current core state. |
| */ |
| ai_core_disable(sih, (bits | resetbits)); |
| |
| /* |
| * Now do the initialization sequence. |
| */ |
| W_REG(&ai->ioctrl, (bits | SICF_FGC | SICF_CLOCK_EN)); |
| dummy = R_REG(&ai->ioctrl); |
| W_REG(&ai->resetctrl, 0); |
| udelay(1); |
| |
| W_REG(&ai->ioctrl, (bits | SICF_CLOCK_EN)); |
| dummy = R_REG(&ai->ioctrl); |
| udelay(1); |
| } |
| |
| /* return the slow clock source - LPO, XTAL, or PCI */ |
| static uint ai_slowclk_src(struct si_info *sii) |
| { |
| chipcregs_t *cc; |
| u32 val; |
| |
| if (sii->pub.ccrev < 6) { |
| if (sii->pub.bustype == PCI_BUS) { |
| pci_read_config_dword(sii->pbus, PCI_GPIO_OUT, |
| &val); |
| if (val & PCI_CFG_GPIO_SCS) |
| return SCC_SS_PCI; |
| } |
| return SCC_SS_XTAL; |
| } else if (sii->pub.ccrev < 10) { |
| cc = (chipcregs_t *) ai_setcoreidx(&sii->pub, sii->curidx); |
| return R_REG(&cc->slow_clk_ctl) & SCC_SS_MASK; |
| } else /* Insta-clock */ |
| return SCC_SS_XTAL; |
| } |
| |
| /* |
| * return the ILP (slowclock) min or max frequency |
| * precondition: we've established the chip has dynamic clk control |
| */ |
| static uint ai_slowclk_freq(struct si_info *sii, bool max_freq, chipcregs_t *cc) |
| { |
| u32 slowclk; |
| uint div; |
| |
| slowclk = ai_slowclk_src(sii); |
| if (sii->pub.ccrev < 6) { |
| if (slowclk == SCC_SS_PCI) |
| return max_freq ? (PCIMAXFREQ / 64) |
| : (PCIMINFREQ / 64); |
| else |
| return max_freq ? (XTALMAXFREQ / 32) |
| : (XTALMINFREQ / 32); |
| } else if (sii->pub.ccrev < 10) { |
| div = 4 * |
| (((R_REG(&cc->slow_clk_ctl) & SCC_CD_MASK) >> |
| SCC_CD_SHIFT) + 1); |
| if (slowclk == SCC_SS_LPO) |
| return max_freq ? LPOMAXFREQ : LPOMINFREQ; |
| else if (slowclk == SCC_SS_XTAL) |
| return max_freq ? (XTALMAXFREQ / div) |
| : (XTALMINFREQ / div); |
| else if (slowclk == SCC_SS_PCI) |
| return max_freq ? (PCIMAXFREQ / div) |
| : (PCIMINFREQ / div); |
| } else { |
| /* Chipc rev 10 is InstaClock */ |
| div = R_REG(&cc->system_clk_ctl) >> SYCC_CD_SHIFT; |
| div = 4 * (div + 1); |
| return max_freq ? XTALMAXFREQ : (XTALMINFREQ / div); |
| } |
| return 0; |
| } |
| |
| static void ai_clkctl_setdelay(struct si_info *sii, void *chipcregs) |
| { |
| chipcregs_t *cc = (chipcregs_t *) chipcregs; |
| uint slowmaxfreq, pll_delay, slowclk; |
| uint pll_on_delay, fref_sel_delay; |
| |
| pll_delay = PLL_DELAY; |
| |
| /* |
| * If the slow clock is not sourced by the xtal then |
| * add the xtal_on_delay since the xtal will also be |
| * powered down by dynamic clk control logic. |
| */ |
| |
| slowclk = ai_slowclk_src(sii); |
| if (slowclk != SCC_SS_XTAL) |
| pll_delay += XTAL_ON_DELAY; |
| |
| /* Starting with 4318 it is ILP that is used for the delays */ |
| slowmaxfreq = |
| ai_slowclk_freq(sii, (sii->pub.ccrev >= 10) ? false : true, cc); |
| |
| pll_on_delay = ((slowmaxfreq * pll_delay) + 999999) / 1000000; |
| fref_sel_delay = ((slowmaxfreq * FREF_DELAY) + 999999) / 1000000; |
| |
| W_REG(&cc->pll_on_delay, pll_on_delay); |
| W_REG(&cc->fref_sel_delay, fref_sel_delay); |
| } |
| |
| /* initialize power control delay registers */ |
| void ai_clkctl_init(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| uint origidx = 0; |
| chipcregs_t *cc; |
| bool fast; |
| |
| if (!CCCTL_ENAB(sih)) |
| return; |
| |
| sii = SI_INFO(sih); |
| fast = SI_FAST(sii); |
| if (!fast) { |
| origidx = sii->curidx; |
| cc = (chipcregs_t *) ai_setcore(sih, CC_CORE_ID, 0); |
| if (cc == NULL) |
| return; |
| } else { |
| cc = (chipcregs_t *) CCREGS_FAST(sii); |
| if (cc == NULL) |
| return; |
| } |
| |
| /* set all Instaclk chip ILP to 1 MHz */ |
| if (sih->ccrev >= 10) |
| SET_REG(&cc->system_clk_ctl, SYCC_CD_MASK, |
| (ILP_DIV_1MHZ << SYCC_CD_SHIFT)); |
| |
| ai_clkctl_setdelay(sii, (void *)cc); |
| |
| if (!fast) |
| ai_setcoreidx(sih, origidx); |
| } |
| |
| /* |
| * return the value suitable for writing to the |
| * dot11 core FAST_PWRUP_DELAY register |
| */ |
| u16 ai_clkctl_fast_pwrup_delay(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| uint origidx = 0; |
| chipcregs_t *cc; |
| uint slowminfreq; |
| u16 fpdelay; |
| uint intr_val = 0; |
| bool fast; |
| |
| sii = SI_INFO(sih); |
| if (PMUCTL_ENAB(sih)) { |
| INTR_OFF(sii, intr_val); |
| fpdelay = si_pmu_fast_pwrup_delay(sih); |
| INTR_RESTORE(sii, intr_val); |
| return fpdelay; |
| } |
| |
| if (!CCCTL_ENAB(sih)) |
| return 0; |
| |
| fast = SI_FAST(sii); |
| fpdelay = 0; |
| if (!fast) { |
| origidx = sii->curidx; |
| INTR_OFF(sii, intr_val); |
| cc = (chipcregs_t *) ai_setcore(sih, CC_CORE_ID, 0); |
| if (cc == NULL) |
| goto done; |
| } else { |
| cc = (chipcregs_t *) CCREGS_FAST(sii); |
| if (cc == NULL) |
| goto done; |
| } |
| |
| slowminfreq = ai_slowclk_freq(sii, false, cc); |
| fpdelay = (((R_REG(&cc->pll_on_delay) + 2) * 1000000) + |
| (slowminfreq - 1)) / slowminfreq; |
| |
| done: |
| if (!fast) { |
| ai_setcoreidx(sih, origidx); |
| INTR_RESTORE(sii, intr_val); |
| } |
| return fpdelay; |
| } |
| |
| /* turn primary xtal and/or pll off/on */ |
| int ai_clkctl_xtal(struct si_pub *sih, uint what, bool on) |
| { |
| struct si_info *sii; |
| u32 in, out, outen; |
| |
| sii = SI_INFO(sih); |
| |
| switch (sih->bustype) { |
| |
| case PCI_BUS: |
| /* pcie core doesn't have any mapping to control the xtal pu */ |
| if (PCIE(sii)) |
| return -1; |
| |
| pci_read_config_dword(sii->pbus, PCI_GPIO_IN, &in); |
| pci_read_config_dword(sii->pbus, PCI_GPIO_OUT, &out); |
| pci_read_config_dword(sii->pbus, PCI_GPIO_OUTEN, &outen); |
| |
| /* |
| * Avoid glitching the clock if GPRS is already using it. |
| * We can't actually read the state of the PLLPD so we infer it |
| * by the value of XTAL_PU which *is* readable via gpioin. |
| */ |
| if (on && (in & PCI_CFG_GPIO_XTAL)) |
| return 0; |
| |
| if (what & XTAL) |
| outen |= PCI_CFG_GPIO_XTAL; |
| if (what & PLL) |
| outen |= PCI_CFG_GPIO_PLL; |
| |
| if (on) { |
| /* turn primary xtal on */ |
| if (what & XTAL) { |
| out |= PCI_CFG_GPIO_XTAL; |
| if (what & PLL) |
| out |= PCI_CFG_GPIO_PLL; |
| pci_write_config_dword(sii->pbus, |
| PCI_GPIO_OUT, out); |
| pci_write_config_dword(sii->pbus, |
| PCI_GPIO_OUTEN, outen); |
| udelay(XTAL_ON_DELAY); |
| } |
| |
| /* turn pll on */ |
| if (what & PLL) { |
| out &= ~PCI_CFG_GPIO_PLL; |
| pci_write_config_dword(sii->pbus, |
| PCI_GPIO_OUT, out); |
| mdelay(2); |
| } |
| } else { |
| if (what & XTAL) |
| out &= ~PCI_CFG_GPIO_XTAL; |
| if (what & PLL) |
| out |= PCI_CFG_GPIO_PLL; |
| pci_write_config_dword(sii->pbus, |
| PCI_GPIO_OUT, out); |
| pci_write_config_dword(sii->pbus, |
| PCI_GPIO_OUTEN, outen); |
| } |
| |
| default: |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| /* |
| * clock control policy function throught chipcommon |
| * |
| * set dynamic clk control mode (forceslow, forcefast, dynamic) |
| * returns true if we are forcing fast clock |
| * this is a wrapper over the next internal function |
| * to allow flexible policy settings for outside caller |
| */ |
| bool ai_clkctl_cc(struct si_pub *sih, uint mode) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| |
| /* chipcommon cores prior to rev6 don't support dynamic clock control */ |
| if (sih->ccrev < 6) |
| return false; |
| |
| if (PCI_FORCEHT(sii)) |
| return mode == CLK_FAST; |
| |
| return _ai_clkctl_cc(sii, mode); |
| } |
| |
| /* clk control mechanism through chipcommon, no policy checking */ |
| static bool _ai_clkctl_cc(struct si_info *sii, uint mode) |
| { |
| uint origidx = 0; |
| chipcregs_t *cc; |
| u32 scc; |
| uint intr_val = 0; |
| bool fast = SI_FAST(sii); |
| |
| /* chipcommon cores prior to rev6 don't support dynamic clock control */ |
| if (sii->pub.ccrev < 6) |
| return false; |
| |
| if (!fast) { |
| INTR_OFF(sii, intr_val); |
| origidx = sii->curidx; |
| |
| if ((sii->pub.bustype == SI_BUS) && |
| ai_setcore(&sii->pub, MIPS33_CORE_ID, 0) && |
| (ai_corerev(&sii->pub) <= 7) && (sii->pub.ccrev >= 10)) |
| goto done; |
| |
| cc = (chipcregs_t *) ai_setcore(&sii->pub, CC_CORE_ID, 0); |
| } else { |
| cc = (chipcregs_t *) CCREGS_FAST(sii); |
| if (cc == NULL) |
| goto done; |
| } |
| |
| if (!CCCTL_ENAB(&sii->pub) && (sii->pub.ccrev < 20)) |
| goto done; |
| |
| switch (mode) { |
| case CLK_FAST: /* FORCEHT, fast (pll) clock */ |
| if (sii->pub.ccrev < 10) { |
| /* |
| * don't forget to force xtal back |
| * on before we clear SCC_DYN_XTAL.. |
| */ |
| ai_clkctl_xtal(&sii->pub, XTAL, ON); |
| SET_REG(&cc->slow_clk_ctl, |
| (SCC_XC | SCC_FS | SCC_IP), SCC_IP); |
| } else if (sii->pub.ccrev < 20) { |
| OR_REG(&cc->system_clk_ctl, SYCC_HR); |
| } else { |
| OR_REG(&cc->clk_ctl_st, CCS_FORCEHT); |
| } |
| |
| /* wait for the PLL */ |
| if (PMUCTL_ENAB(&sii->pub)) { |
| u32 htavail = CCS_HTAVAIL; |
| SPINWAIT(((R_REG(&cc->clk_ctl_st) & htavail) |
| == 0), PMU_MAX_TRANSITION_DLY); |
| } else { |
| udelay(PLL_DELAY); |
| } |
| break; |
| |
| case CLK_DYNAMIC: /* enable dynamic clock control */ |
| if (sii->pub.ccrev < 10) { |
| scc = R_REG(&cc->slow_clk_ctl); |
| scc &= ~(SCC_FS | SCC_IP | SCC_XC); |
| if ((scc & SCC_SS_MASK) != SCC_SS_XTAL) |
| scc |= SCC_XC; |
| W_REG(&cc->slow_clk_ctl, scc); |
| |
| /* |
| * for dynamic control, we have to |
| * release our xtal_pu "force on" |
| */ |
| if (scc & SCC_XC) |
| ai_clkctl_xtal(&sii->pub, XTAL, OFF); |
| } else if (sii->pub.ccrev < 20) { |
| /* Instaclock */ |
| AND_REG(&cc->system_clk_ctl, ~SYCC_HR); |
| } else { |
| AND_REG(&cc->clk_ctl_st, ~CCS_FORCEHT); |
| } |
| break; |
| |
| default: |
| break; |
| } |
| |
| done: |
| if (!fast) { |
| ai_setcoreidx(&sii->pub, origidx); |
| INTR_RESTORE(sii, intr_val); |
| } |
| return mode == CLK_FAST; |
| } |
| |
| /* Build device path. Support SI, PCI, and JTAG for now. */ |
| int ai_devpath(struct si_pub *sih, char *path, int size) |
| { |
| int slen; |
| |
| if (!path || size <= 0) |
| return -1; |
| |
| switch (sih->bustype) { |
| case SI_BUS: |
| case JTAG_BUS: |
| slen = snprintf(path, (size_t) size, "sb/%u/", ai_coreidx(sih)); |
| break; |
| case PCI_BUS: |
| slen = snprintf(path, (size_t) size, "pci/%u/%u/", |
| ((struct pci_dev *)((SI_INFO(sih))->pbus))->bus->number, |
| PCI_SLOT( |
| ((struct pci_dev *)((SI_INFO(sih))->pbus))->devfn)); |
| break; |
| |
| default: |
| slen = -1; |
| break; |
| } |
| |
| if (slen < 0 || slen >= size) { |
| path[0] = '\0'; |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| /* Get a variable, but only if it has a devpath prefix */ |
| char *ai_getdevpathvar(struct si_pub *sih, const char *name) |
| { |
| char varname[SI_DEVPATH_BUFSZ + 32]; |
| |
| ai_devpathvar(sih, varname, sizeof(varname), name); |
| |
| return getvar(NULL, varname); |
| } |
| |
| /* Get a variable, but only if it has a devpath prefix */ |
| int ai_getdevpathintvar(struct si_pub *sih, const char *name) |
| { |
| #if defined(BCMBUSTYPE) && (BCMBUSTYPE == SI_BUS) |
| return getintvar(NULL, name); |
| #else |
| char varname[SI_DEVPATH_BUFSZ + 32]; |
| |
| ai_devpathvar(sih, varname, sizeof(varname), name); |
| |
| return getintvar(NULL, varname); |
| #endif |
| } |
| |
| char *ai_getnvramflvar(struct si_pub *sih, const char *name) |
| { |
| return getvar(NULL, name); |
| } |
| |
| /* Concatenate the dev path with a varname into the given 'var' buffer |
| * and return the 'var' pointer. Nothing is done to the arguments if |
| * len == 0 or var is NULL, var is still returned. On overflow, the |
| * first char will be set to '\0'. |
| */ |
| static char *ai_devpathvar(struct si_pub *sih, char *var, int len, |
| const char *name) |
| { |
| uint path_len; |
| |
| if (!var || len <= 0) |
| return var; |
| |
| if (ai_devpath(sih, var, len) == 0) { |
| path_len = strlen(var); |
| |
| if (strlen(name) + 1 > (uint) (len - path_len)) |
| var[0] = '\0'; |
| else |
| strncpy(var + path_len, name, len - path_len - 1); |
| } |
| |
| return var; |
| } |
| |
| /* return true if PCIE capability exists in the pci config space */ |
| static bool ai_ispcie(struct si_info *sii) |
| { |
| u8 cap_ptr; |
| |
| if (sii->pub.bustype != PCI_BUS) |
| return false; |
| |
| cap_ptr = |
| pcicore_find_pci_capability(sii->pbus, PCI_CAP_ID_EXP, NULL, |
| NULL); |
| if (!cap_ptr) |
| return false; |
| |
| return true; |
| } |
| |
| bool ai_pci_war16165(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| |
| return PCI(sii) && (sih->buscorerev <= 10); |
| } |
| |
| void ai_pci_up(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| |
| /* if not pci bus, we're done */ |
| if (sih->bustype != PCI_BUS) |
| return; |
| |
| if (PCI_FORCEHT(sii)) |
| _ai_clkctl_cc(sii, CLK_FAST); |
| |
| if (PCIE(sii)) |
| pcicore_up(sii->pch, SI_PCIUP); |
| |
| } |
| |
| /* Unconfigure and/or apply various WARs when system is going to sleep mode */ |
| void ai_pci_sleep(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| |
| pcicore_sleep(sii->pch); |
| } |
| |
| /* Unconfigure and/or apply various WARs when going down */ |
| void ai_pci_down(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| |
| /* if not pci bus, we're done */ |
| if (sih->bustype != PCI_BUS) |
| return; |
| |
| /* release FORCEHT since chip is going to "down" state */ |
| if (PCI_FORCEHT(sii)) |
| _ai_clkctl_cc(sii, CLK_DYNAMIC); |
| |
| pcicore_down(sii->pch, SI_PCIDOWN); |
| } |
| |
| /* |
| * Configure the pci core for pci client (NIC) action |
| * coremask is the bitvec of cores by index to be enabled. |
| */ |
| void ai_pci_setup(struct si_pub *sih, uint coremask) |
| { |
| struct si_info *sii; |
| void *regs = NULL; |
| u32 siflag = 0, w; |
| uint idx = 0; |
| |
| sii = SI_INFO(sih); |
| |
| if (sii->pub.bustype != PCI_BUS) |
| return; |
| |
| if (PCI(sii)) { |
| /* get current core index */ |
| idx = sii->curidx; |
| |
| /* we interrupt on this backplane flag number */ |
| siflag = ai_flag(sih); |
| |
| /* switch over to pci core */ |
| regs = ai_setcoreidx(sih, sii->pub.buscoreidx); |
| } |
| |
| /* |
| * Enable sb->pci interrupts. Assume |
| * PCI rev 2.3 support was added in pci core rev 6 and things changed.. |
| */ |
| if (PCIE(sii) || (PCI(sii) && ((sii->pub.buscorerev) >= 6))) { |
| /* pci config write to set this core bit in PCIIntMask */ |
| pci_read_config_dword(sii->pbus, PCI_INT_MASK, &w); |
| w |= (coremask << PCI_SBIM_SHIFT); |
| pci_write_config_dword(sii->pbus, PCI_INT_MASK, w); |
| } else { |
| /* set sbintvec bit for our flag number */ |
| ai_setint(sih, siflag); |
| } |
| |
| if (PCI(sii)) { |
| pcicore_pci_setup(sii->pch, regs); |
| |
| /* switch back to previous core */ |
| ai_setcoreidx(sih, idx); |
| } |
| } |
| |
| /* |
| * Fixup SROMless PCI device's configuration. |
| * The current core may be changed upon return. |
| */ |
| int ai_pci_fixcfg(struct si_pub *sih) |
| { |
| uint origidx; |
| void *regs = NULL; |
| |
| struct si_info *sii = SI_INFO(sih); |
| |
| /* Fixup PI in SROM shadow area to enable the correct PCI core access */ |
| /* save the current index */ |
| origidx = ai_coreidx(&sii->pub); |
| |
| /* check 'pi' is correct and fix it if not */ |
| regs = ai_setcore(&sii->pub, sii->pub.buscoretype, 0); |
| pcicore_fixcfg(sii->pch, regs); |
| |
| /* restore the original index */ |
| ai_setcoreidx(&sii->pub, origidx); |
| |
| pcicore_hwup(sii->pch); |
| return 0; |
| } |
| |
| /* mask&set gpiocontrol bits */ |
| u32 ai_gpiocontrol(struct si_pub *sih, u32 mask, u32 val, u8 priority) |
| { |
| uint regoff; |
| |
| regoff = 0; |
| |
| /* gpios could be shared on router platforms |
| * ignore reservation if it's high priority (e.g., test apps) |
| */ |
| if ((priority != GPIO_HI_PRIORITY) && |
| (sih->bustype == SI_BUS) && (val || mask)) { |
| mask = priority ? (ai_gpioreservation & mask) : |
| ((ai_gpioreservation | mask) & ~(ai_gpioreservation)); |
| val &= mask; |
| } |
| |
| regoff = offsetof(chipcregs_t, gpiocontrol); |
| return ai_corereg(sih, SI_CC_IDX, regoff, mask, val); |
| } |
| |
| void ai_chipcontrl_epa4331(struct si_pub *sih, bool on) |
| { |
| struct si_info *sii; |
| chipcregs_t *cc; |
| uint origidx; |
| u32 val; |
| |
| sii = SI_INFO(sih); |
| origidx = ai_coreidx(sih); |
| |
| cc = (chipcregs_t *) ai_setcore(sih, CC_CORE_ID, 0); |
| |
| val = R_REG(&cc->chipcontrol); |
| |
| if (on) { |
| if (sih->chippkg == 9 || sih->chippkg == 0xb) { |
| /* Ext PA Controls for 4331 12x9 Package */ |
| W_REG(&cc->chipcontrol, val | |
| (CCTRL4331_EXTPA_EN | |
| CCTRL4331_EXTPA_ON_GPIO2_5)); |
| } else { |
| /* Ext PA Controls for 4331 12x12 Package */ |
| W_REG(&cc->chipcontrol, |
| val | (CCTRL4331_EXTPA_EN)); |
| } |
| } else { |
| val &= ~(CCTRL4331_EXTPA_EN | CCTRL4331_EXTPA_ON_GPIO2_5); |
| W_REG(&cc->chipcontrol, val); |
| } |
| |
| ai_setcoreidx(sih, origidx); |
| } |
| |
| /* Enable BT-COEX & Ex-PA for 4313 */ |
| void ai_epa_4313war(struct si_pub *sih) |
| { |
| struct si_info *sii; |
| chipcregs_t *cc; |
| uint origidx; |
| |
| sii = SI_INFO(sih); |
| origidx = ai_coreidx(sih); |
| |
| cc = (chipcregs_t *) ai_setcore(sih, CC_CORE_ID, 0); |
| |
| /* EPA Fix */ |
| W_REG(&cc->gpiocontrol, |
| R_REG(&cc->gpiocontrol) | GPIO_CTRL_EPA_EN_MASK); |
| |
| ai_setcoreidx(sih, origidx); |
| } |
| |
| /* check if the device is removed */ |
| bool ai_deviceremoved(struct si_pub *sih) |
| { |
| u32 w; |
| struct si_info *sii; |
| |
| sii = SI_INFO(sih); |
| |
| switch (sih->bustype) { |
| case PCI_BUS: |
| pci_read_config_dword(sii->pbus, PCI_VENDOR_ID, &w); |
| if ((w & 0xFFFF) != PCI_VENDOR_ID_BROADCOM) |
| return true; |
| break; |
| } |
| return false; |
| } |
| |
| bool ai_is_sprom_available(struct si_pub *sih) |
| { |
| if (sih->ccrev >= 31) { |
| struct si_info *sii; |
| uint origidx; |
| chipcregs_t *cc; |
| u32 sromctrl; |
| |
| if ((sih->cccaps & CC_CAP_SROM) == 0) |
| return false; |
| |
| sii = SI_INFO(sih); |
| origidx = sii->curidx; |
| cc = ai_setcoreidx(sih, SI_CC_IDX); |
| sromctrl = R_REG(&cc->sromcontrol); |
| ai_setcoreidx(sih, origidx); |
| return sromctrl & SRC_PRESENT; |
| } |
| |
| switch (sih->chip) { |
| case BCM4313_CHIP_ID: |
| return (sih->chipst & CST4313_SPROM_PRESENT) != 0; |
| default: |
| return true; |
| } |
| } |
| |
| bool ai_is_otp_disabled(struct si_pub *sih) |
| { |
| switch (sih->chip) { |
| case BCM4313_CHIP_ID: |
| return (sih->chipst & CST4313_OTP_PRESENT) == 0; |
| /* These chips always have their OTP on */ |
| case BCM43224_CHIP_ID: |
| case BCM43225_CHIP_ID: |
| default: |
| return false; |
| } |
| } |