|  | /* | 
|  | * Low-Level PCI Support for the SH7780 | 
|  | * | 
|  | *  Copyright (C) 2005 - 2010  Paul Mundt | 
|  | * | 
|  | * This file is subject to the terms and conditions of the GNU General Public | 
|  | * License.  See the file "COPYING" in the main directory of this archive | 
|  | * for more details. | 
|  | */ | 
|  | #include <linux/types.h> | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/pci.h> | 
|  | #include <linux/interrupt.h> | 
|  | #include <linux/timer.h> | 
|  | #include <linux/irq.h> | 
|  | #include <linux/errno.h> | 
|  | #include <linux/delay.h> | 
|  | #include <linux/log2.h> | 
|  | #include "pci-sh4.h" | 
|  | #include <asm/mmu.h> | 
|  | #include <asm/sizes.h> | 
|  |  | 
|  | static struct resource sh7785_pci_resources[] = { | 
|  | { | 
|  | .name	= "PCI IO", | 
|  | .start	= 0x1000, | 
|  | .end	= SZ_4M - 1, | 
|  | .flags	= IORESOURCE_IO, | 
|  | }, { | 
|  | .name	= "PCI MEM 0", | 
|  | .start	= 0xfd000000, | 
|  | .end	= 0xfd000000 + SZ_16M - 1, | 
|  | .flags	= IORESOURCE_MEM, | 
|  | }, { | 
|  | .name	= "PCI MEM 1", | 
|  | .start	= 0x10000000, | 
|  | .end	= 0x10000000 + SZ_64M - 1, | 
|  | .flags	= IORESOURCE_MEM, | 
|  | }, { | 
|  | /* | 
|  | * 32-bit only resources must be last. | 
|  | */ | 
|  | .name	= "PCI MEM 2", | 
|  | .start	= 0xc0000000, | 
|  | .end	= 0xc0000000 + SZ_512M - 1, | 
|  | .flags	= IORESOURCE_MEM | IORESOURCE_MEM_32BIT, | 
|  | }, | 
|  | }; | 
|  |  | 
|  | static struct pci_channel sh7780_pci_controller = { | 
|  | .pci_ops	= &sh4_pci_ops, | 
|  | .resources	= sh7785_pci_resources, | 
|  | .nr_resources	= ARRAY_SIZE(sh7785_pci_resources), | 
|  | .io_offset	= 0, | 
|  | .mem_offset	= 0, | 
|  | .io_map_base	= 0xfe200000, | 
|  | .serr_irq	= evt2irq(0xa00), | 
|  | .err_irq	= evt2irq(0xaa0), | 
|  | }; | 
|  |  | 
|  | struct pci_errors { | 
|  | unsigned int	mask; | 
|  | const char	*str; | 
|  | } pci_arbiter_errors[] = { | 
|  | { SH4_PCIAINT_MBKN,	"master broken" }, | 
|  | { SH4_PCIAINT_TBTO,	"target bus time out" }, | 
|  | { SH4_PCIAINT_MBTO,	"master bus time out" }, | 
|  | { SH4_PCIAINT_TABT,	"target abort" }, | 
|  | { SH4_PCIAINT_MABT,	"master abort" }, | 
|  | { SH4_PCIAINT_RDPE,	"read data parity error" }, | 
|  | { SH4_PCIAINT_WDPE,	"write data parity error" }, | 
|  | }, pci_interrupt_errors[] = { | 
|  | { SH4_PCIINT_MLCK,	"master lock error" }, | 
|  | { SH4_PCIINT_TABT,	"target-target abort" }, | 
|  | { SH4_PCIINT_TRET,	"target retry time out" }, | 
|  | { SH4_PCIINT_MFDE,	"master function disable erorr" }, | 
|  | { SH4_PCIINT_PRTY,	"address parity error" }, | 
|  | { SH4_PCIINT_SERR,	"SERR" }, | 
|  | { SH4_PCIINT_TWDP,	"data parity error for target write" }, | 
|  | { SH4_PCIINT_TRDP,	"PERR detected for target read" }, | 
|  | { SH4_PCIINT_MTABT,	"target abort for master" }, | 
|  | { SH4_PCIINT_MMABT,	"master abort for master" }, | 
|  | { SH4_PCIINT_MWPD,	"master write data parity error" }, | 
|  | { SH4_PCIINT_MRPD,	"master read data parity error" }, | 
|  | }; | 
|  |  | 
|  | static irqreturn_t sh7780_pci_err_irq(int irq, void *dev_id) | 
|  | { | 
|  | struct pci_channel *hose = dev_id; | 
|  | unsigned long addr; | 
|  | unsigned int status; | 
|  | unsigned int cmd; | 
|  | int i; | 
|  |  | 
|  | addr = __raw_readl(hose->reg_base + SH4_PCIALR); | 
|  |  | 
|  | /* | 
|  | * Handle status errors. | 
|  | */ | 
|  | status = __raw_readw(hose->reg_base + PCI_STATUS); | 
|  | if (status & (PCI_STATUS_PARITY | | 
|  | PCI_STATUS_DETECTED_PARITY | | 
|  | PCI_STATUS_SIG_TARGET_ABORT | | 
|  | PCI_STATUS_REC_TARGET_ABORT | | 
|  | PCI_STATUS_REC_MASTER_ABORT)) { | 
|  | cmd = pcibios_handle_status_errors(addr, status, hose); | 
|  | if (likely(cmd)) | 
|  | __raw_writew(cmd, hose->reg_base + PCI_STATUS); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Handle arbiter errors. | 
|  | */ | 
|  | status = __raw_readl(hose->reg_base + SH4_PCIAINT); | 
|  | for (i = cmd = 0; i < ARRAY_SIZE(pci_arbiter_errors); i++) { | 
|  | if (status & pci_arbiter_errors[i].mask) { | 
|  | printk(KERN_DEBUG "PCI: %s, addr=%08lx\n", | 
|  | pci_arbiter_errors[i].str, addr); | 
|  | cmd |= pci_arbiter_errors[i].mask; | 
|  | } | 
|  | } | 
|  | __raw_writel(cmd, hose->reg_base + SH4_PCIAINT); | 
|  |  | 
|  | /* | 
|  | * Handle the remaining PCI errors. | 
|  | */ | 
|  | status = __raw_readl(hose->reg_base + SH4_PCIINT); | 
|  | for (i = cmd = 0; i < ARRAY_SIZE(pci_interrupt_errors); i++) { | 
|  | if (status & pci_interrupt_errors[i].mask) { | 
|  | printk(KERN_DEBUG "PCI: %s, addr=%08lx\n", | 
|  | pci_interrupt_errors[i].str, addr); | 
|  | cmd |= pci_interrupt_errors[i].mask; | 
|  | } | 
|  | } | 
|  | __raw_writel(cmd, hose->reg_base + SH4_PCIINT); | 
|  |  | 
|  | return IRQ_HANDLED; | 
|  | } | 
|  |  | 
|  | static irqreturn_t sh7780_pci_serr_irq(int irq, void *dev_id) | 
|  | { | 
|  | struct pci_channel *hose = dev_id; | 
|  |  | 
|  | printk(KERN_DEBUG "PCI: system error received: "); | 
|  | pcibios_report_status(PCI_STATUS_SIG_SYSTEM_ERROR, 1); | 
|  | printk("\n"); | 
|  |  | 
|  | /* Deassert SERR */ | 
|  | __raw_writel(SH4_PCIINTM_SDIM, hose->reg_base + SH4_PCIINTM); | 
|  |  | 
|  | /* Back off the IRQ for awhile */ | 
|  | disable_irq_nosync(irq); | 
|  | hose->serr_timer.expires = jiffies + HZ; | 
|  | add_timer(&hose->serr_timer); | 
|  |  | 
|  | return IRQ_HANDLED; | 
|  | } | 
|  |  | 
|  | static int __init sh7780_pci_setup_irqs(struct pci_channel *hose) | 
|  | { | 
|  | int ret; | 
|  |  | 
|  | /* Clear out PCI arbiter IRQs */ | 
|  | __raw_writel(0, hose->reg_base + SH4_PCIAINT); | 
|  |  | 
|  | /* Clear all error conditions */ | 
|  | __raw_writew(PCI_STATUS_DETECTED_PARITY  | \ | 
|  | PCI_STATUS_SIG_SYSTEM_ERROR | \ | 
|  | PCI_STATUS_REC_MASTER_ABORT | \ | 
|  | PCI_STATUS_REC_TARGET_ABORT | \ | 
|  | PCI_STATUS_SIG_TARGET_ABORT | \ | 
|  | PCI_STATUS_PARITY, hose->reg_base + PCI_STATUS); | 
|  |  | 
|  | ret = request_irq(hose->serr_irq, sh7780_pci_serr_irq, IRQF_DISABLED, | 
|  | "PCI SERR interrupt", hose); | 
|  | if (unlikely(ret)) { | 
|  | printk(KERN_ERR "PCI: Failed hooking SERR IRQ\n"); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * The PCI ERR IRQ needs to be IRQF_SHARED since all of the power | 
|  | * down IRQ vectors are routed through the ERR IRQ vector. We | 
|  | * only request_irq() once as there is only a single masking | 
|  | * source for multiple events. | 
|  | */ | 
|  | ret = request_irq(hose->err_irq, sh7780_pci_err_irq, IRQF_SHARED, | 
|  | "PCI ERR interrupt", hose); | 
|  | if (unlikely(ret)) { | 
|  | free_irq(hose->serr_irq, hose); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Unmask all of the arbiter IRQs. */ | 
|  | __raw_writel(SH4_PCIAINT_MBKN | SH4_PCIAINT_TBTO | SH4_PCIAINT_MBTO | \ | 
|  | SH4_PCIAINT_TABT | SH4_PCIAINT_MABT | SH4_PCIAINT_RDPE | \ | 
|  | SH4_PCIAINT_WDPE, hose->reg_base + SH4_PCIAINTM); | 
|  |  | 
|  | /* Unmask all of the PCI IRQs */ | 
|  | __raw_writel(SH4_PCIINTM_TTADIM  | SH4_PCIINTM_TMTOIM  | \ | 
|  | SH4_PCIINTM_MDEIM   | SH4_PCIINTM_APEDIM  | \ | 
|  | SH4_PCIINTM_SDIM    | SH4_PCIINTM_DPEITWM | \ | 
|  | SH4_PCIINTM_PEDITRM | SH4_PCIINTM_TADIMM  | \ | 
|  | SH4_PCIINTM_MADIMM  | SH4_PCIINTM_MWPDIM  | \ | 
|  | SH4_PCIINTM_MRDPEIM, hose->reg_base + SH4_PCIINTM); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static inline void __init sh7780_pci_teardown_irqs(struct pci_channel *hose) | 
|  | { | 
|  | free_irq(hose->err_irq, hose); | 
|  | free_irq(hose->serr_irq, hose); | 
|  | } | 
|  |  | 
|  | static void __init sh7780_pci66_init(struct pci_channel *hose) | 
|  | { | 
|  | unsigned int tmp; | 
|  |  | 
|  | if (!pci_is_66mhz_capable(hose, 0, 0)) | 
|  | return; | 
|  |  | 
|  | /* Enable register access */ | 
|  | tmp = __raw_readl(hose->reg_base + SH4_PCICR); | 
|  | tmp |= SH4_PCICR_PREFIX; | 
|  | __raw_writel(tmp, hose->reg_base + SH4_PCICR); | 
|  |  | 
|  | /* Enable 66MHz operation */ | 
|  | tmp = __raw_readw(hose->reg_base + PCI_STATUS); | 
|  | tmp |= PCI_STATUS_66MHZ; | 
|  | __raw_writew(tmp, hose->reg_base + PCI_STATUS); | 
|  |  | 
|  | /* Done */ | 
|  | tmp = __raw_readl(hose->reg_base + SH4_PCICR); | 
|  | tmp |= SH4_PCICR_PREFIX | SH4_PCICR_CFIN; | 
|  | __raw_writel(tmp, hose->reg_base + SH4_PCICR); | 
|  | } | 
|  |  | 
|  | static int __init sh7780_pci_init(void) | 
|  | { | 
|  | struct pci_channel *chan = &sh7780_pci_controller; | 
|  | phys_addr_t memphys; | 
|  | size_t memsize; | 
|  | unsigned int id; | 
|  | const char *type; | 
|  | int ret, i; | 
|  |  | 
|  | printk(KERN_NOTICE "PCI: Starting intialization.\n"); | 
|  |  | 
|  | chan->reg_base = 0xfe040000; | 
|  |  | 
|  | /* Enable CPU access to the PCIC registers. */ | 
|  | __raw_writel(PCIECR_ENBL, PCIECR); | 
|  |  | 
|  | /* Reset */ | 
|  | __raw_writel(SH4_PCICR_PREFIX | SH4_PCICR_PRST, | 
|  | chan->reg_base + SH4_PCICR); | 
|  |  | 
|  | /* | 
|  | * Wait for it to come back up. The spec says to allow for up to | 
|  | * 1 second after toggling the reset pin, but in practice 100ms | 
|  | * is more than enough. | 
|  | */ | 
|  | mdelay(100); | 
|  |  | 
|  | id = __raw_readw(chan->reg_base + PCI_VENDOR_ID); | 
|  | if (id != PCI_VENDOR_ID_RENESAS) { | 
|  | printk(KERN_ERR "PCI: Unknown vendor ID 0x%04x.\n", id); | 
|  | return -ENODEV; | 
|  | } | 
|  |  | 
|  | id = __raw_readw(chan->reg_base + PCI_DEVICE_ID); | 
|  | type = (id == PCI_DEVICE_ID_RENESAS_SH7763) ? "SH7763" : | 
|  | (id == PCI_DEVICE_ID_RENESAS_SH7780) ? "SH7780" : | 
|  | (id == PCI_DEVICE_ID_RENESAS_SH7781) ? "SH7781" : | 
|  | (id == PCI_DEVICE_ID_RENESAS_SH7785) ? "SH7785" : | 
|  | NULL; | 
|  | if (unlikely(!type)) { | 
|  | printk(KERN_ERR "PCI: Found an unsupported Renesas host " | 
|  | "controller, device id 0x%04x.\n", id); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | printk(KERN_NOTICE "PCI: Found a Renesas %s host " | 
|  | "controller, revision %d.\n", type, | 
|  | __raw_readb(chan->reg_base + PCI_REVISION_ID)); | 
|  |  | 
|  | /* | 
|  | * Now throw it in to register initialization mode and | 
|  | * start the real work. | 
|  | */ | 
|  | __raw_writel(SH4_PCICR_PREFIX, chan->reg_base + SH4_PCICR); | 
|  |  | 
|  | memphys = __pa(memory_start); | 
|  | memsize = roundup_pow_of_two(memory_end - memory_start); | 
|  |  | 
|  | /* | 
|  | * If there's more than 512MB of memory, we need to roll over to | 
|  | * LAR1/LSR1. | 
|  | */ | 
|  | if (memsize > SZ_512M) { | 
|  | __raw_writel(memphys + SZ_512M, chan->reg_base + SH4_PCILAR1); | 
|  | __raw_writel((((memsize - SZ_512M) - SZ_1M) & 0x1ff00000) | 1, | 
|  | chan->reg_base + SH4_PCILSR1); | 
|  | memsize = SZ_512M; | 
|  | } else { | 
|  | /* | 
|  | * Otherwise just zero it out and disable it. | 
|  | */ | 
|  | __raw_writel(0, chan->reg_base + SH4_PCILAR1); | 
|  | __raw_writel(0, chan->reg_base + SH4_PCILSR1); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * LAR0/LSR0 covers up to the first 512MB, which is enough to | 
|  | * cover all of lowmem on most platforms. | 
|  | */ | 
|  | __raw_writel(memphys, chan->reg_base + SH4_PCILAR0); | 
|  | __raw_writel(((memsize - SZ_1M) & 0x1ff00000) | 1, | 
|  | chan->reg_base + SH4_PCILSR0); | 
|  |  | 
|  | /* | 
|  | * Hook up the ERR and SERR IRQs. | 
|  | */ | 
|  | ret = sh7780_pci_setup_irqs(chan); | 
|  | if (unlikely(ret)) | 
|  | return ret; | 
|  |  | 
|  | /* | 
|  | * Disable the cache snoop controller for non-coherent DMA. | 
|  | */ | 
|  | __raw_writel(0, chan->reg_base + SH7780_PCICSCR0); | 
|  | __raw_writel(0, chan->reg_base + SH7780_PCICSAR0); | 
|  | __raw_writel(0, chan->reg_base + SH7780_PCICSCR1); | 
|  | __raw_writel(0, chan->reg_base + SH7780_PCICSAR1); | 
|  |  | 
|  | /* | 
|  | * Setup the memory BARs | 
|  | */ | 
|  | for (i = 1; i < chan->nr_resources; i++) { | 
|  | struct resource *res = chan->resources + i; | 
|  | resource_size_t size; | 
|  |  | 
|  | if (unlikely(res->flags & IORESOURCE_IO)) | 
|  | continue; | 
|  |  | 
|  | /* | 
|  | * Make sure we're in the right physical addressing mode | 
|  | * for dealing with the resource. | 
|  | */ | 
|  | if ((res->flags & IORESOURCE_MEM_32BIT) && __in_29bit_mode()) { | 
|  | chan->nr_resources--; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | size = resource_size(res); | 
|  |  | 
|  | /* | 
|  | * The MBMR mask is calculated in units of 256kB, which | 
|  | * keeps things pretty simple. | 
|  | */ | 
|  | __raw_writel(((roundup_pow_of_two(size) / SZ_256K) - 1) << 18, | 
|  | chan->reg_base + SH7780_PCIMBMR(i - 1)); | 
|  | __raw_writel(res->start, chan->reg_base + SH7780_PCIMBR(i - 1)); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * And I/O. | 
|  | */ | 
|  | __raw_writel(0, chan->reg_base + PCI_BASE_ADDRESS_0); | 
|  | __raw_writel(0, chan->reg_base + SH7780_PCIIOBR); | 
|  | __raw_writel(0, chan->reg_base + SH7780_PCIIOBMR); | 
|  |  | 
|  | __raw_writew(PCI_COMMAND_SERR   | PCI_COMMAND_WAIT   | \ | 
|  | PCI_COMMAND_PARITY | PCI_COMMAND_MASTER | \ | 
|  | PCI_COMMAND_MEMORY, chan->reg_base + PCI_COMMAND); | 
|  |  | 
|  | /* | 
|  | * Initialization mode complete, release the control register and | 
|  | * enable round robin mode to stop device overruns/starvation. | 
|  | */ | 
|  | __raw_writel(SH4_PCICR_PREFIX | SH4_PCICR_CFIN | SH4_PCICR_FTO, | 
|  | chan->reg_base + SH4_PCICR); | 
|  |  | 
|  | ret = register_pci_controller(chan); | 
|  | if (unlikely(ret)) | 
|  | goto err; | 
|  |  | 
|  | sh7780_pci66_init(chan); | 
|  |  | 
|  | printk(KERN_NOTICE "PCI: Running at %dMHz.\n", | 
|  | (__raw_readw(chan->reg_base + PCI_STATUS) & PCI_STATUS_66MHZ) ? | 
|  | 66 : 33); | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err: | 
|  | sh7780_pci_teardown_irqs(chan); | 
|  | return ret; | 
|  | } | 
|  | arch_initcall(sh7780_pci_init); |