|  | #ifndef _ASM_IA64_PCI_H | 
|  | #define _ASM_IA64_PCI_H | 
|  |  | 
|  | #include <linux/mm.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/spinlock.h> | 
|  | #include <linux/string.h> | 
|  | #include <linux/types.h> | 
|  |  | 
|  | #include <asm/io.h> | 
|  | #include <asm/scatterlist.h> | 
|  | #include <asm/hw_irq.h> | 
|  |  | 
|  | /* | 
|  | * Can be used to override the logic in pci_scan_bus for skipping already-configured bus | 
|  | * numbers - to be used for buggy BIOSes or architectures with incomplete PCI setup by the | 
|  | * loader. | 
|  | */ | 
|  | #define pcibios_assign_all_busses()     0 | 
|  | #define pcibios_scan_all_fns(a, b)	0 | 
|  |  | 
|  | #define PCIBIOS_MIN_IO		0x1000 | 
|  | #define PCIBIOS_MIN_MEM		0x10000000 | 
|  |  | 
|  | void pcibios_config_init(void); | 
|  |  | 
|  | struct pci_dev; | 
|  |  | 
|  | /* | 
|  | * PCI_DMA_BUS_IS_PHYS should be set to 1 if there is _necessarily_ a direct | 
|  | * correspondence between device bus addresses and CPU physical addresses. | 
|  | * Platforms with a hardware I/O MMU _must_ turn this off to suppress the | 
|  | * bounce buffer handling code in the block and network device layers. | 
|  | * Platforms with separate bus address spaces _must_ turn this off and provide | 
|  | * a device DMA mapping implementation that takes care of the necessary | 
|  | * address translation. | 
|  | * | 
|  | * For now, the ia64 platforms which may have separate/multiple bus address | 
|  | * spaces all have I/O MMUs which support the merging of physically | 
|  | * discontiguous buffers, so we can use that as the sole factor to determine | 
|  | * the setting of PCI_DMA_BUS_IS_PHYS. | 
|  | */ | 
|  | extern unsigned long ia64_max_iommu_merge_mask; | 
|  | #define PCI_DMA_BUS_IS_PHYS	(ia64_max_iommu_merge_mask == ~0UL) | 
|  |  | 
|  | static inline void | 
|  | pcibios_set_master (struct pci_dev *dev) | 
|  | { | 
|  | /* No special bus mastering setup handling */ | 
|  | } | 
|  |  | 
|  | static inline void | 
|  | pcibios_penalize_isa_irq (int irq, int active) | 
|  | { | 
|  | /* We don't do dynamic PCI IRQ allocation */ | 
|  | } | 
|  |  | 
|  | #include <asm-generic/pci-dma-compat.h> | 
|  |  | 
|  | /* pci_unmap_{single,page} is not a nop, thus... */ | 
|  | #define DECLARE_PCI_UNMAP_ADDR(ADDR_NAME)	\ | 
|  | dma_addr_t ADDR_NAME; | 
|  | #define DECLARE_PCI_UNMAP_LEN(LEN_NAME)		\ | 
|  | __u32 LEN_NAME; | 
|  | #define pci_unmap_addr(PTR, ADDR_NAME)			\ | 
|  | ((PTR)->ADDR_NAME) | 
|  | #define pci_unmap_addr_set(PTR, ADDR_NAME, VAL)		\ | 
|  | (((PTR)->ADDR_NAME) = (VAL)) | 
|  | #define pci_unmap_len(PTR, LEN_NAME)			\ | 
|  | ((PTR)->LEN_NAME) | 
|  | #define pci_unmap_len_set(PTR, LEN_NAME, VAL)		\ | 
|  | (((PTR)->LEN_NAME) = (VAL)) | 
|  |  | 
|  | #ifdef CONFIG_PCI | 
|  | static inline void pci_dma_burst_advice(struct pci_dev *pdev, | 
|  | enum pci_dma_burst_strategy *strat, | 
|  | unsigned long *strategy_parameter) | 
|  | { | 
|  | unsigned long cacheline_size; | 
|  | u8 byte; | 
|  |  | 
|  | pci_read_config_byte(pdev, PCI_CACHE_LINE_SIZE, &byte); | 
|  | if (byte == 0) | 
|  | cacheline_size = 1024; | 
|  | else | 
|  | cacheline_size = (int) byte * 4; | 
|  |  | 
|  | *strat = PCI_DMA_BURST_MULTIPLE; | 
|  | *strategy_parameter = cacheline_size; | 
|  | } | 
|  | #endif | 
|  |  | 
|  | #define HAVE_PCI_MMAP | 
|  | extern int pci_mmap_page_range (struct pci_dev *dev, struct vm_area_struct *vma, | 
|  | enum pci_mmap_state mmap_state, int write_combine); | 
|  | #define HAVE_PCI_LEGACY | 
|  | extern int pci_mmap_legacy_page_range(struct pci_bus *bus, | 
|  | struct vm_area_struct *vma); | 
|  | extern ssize_t pci_read_legacy_io(struct kobject *kobj, | 
|  | struct bin_attribute *bin_attr, | 
|  | char *buf, loff_t off, size_t count); | 
|  | extern ssize_t pci_write_legacy_io(struct kobject *kobj, | 
|  | struct bin_attribute *bin_attr, | 
|  | char *buf, loff_t off, size_t count); | 
|  | extern int pci_mmap_legacy_mem(struct kobject *kobj, | 
|  | struct bin_attribute *attr, | 
|  | struct vm_area_struct *vma); | 
|  |  | 
|  | #define pci_get_legacy_mem platform_pci_get_legacy_mem | 
|  | #define pci_legacy_read platform_pci_legacy_read | 
|  | #define pci_legacy_write platform_pci_legacy_write | 
|  |  | 
|  | struct pci_window { | 
|  | struct resource resource; | 
|  | u64 offset; | 
|  | }; | 
|  |  | 
|  | struct pci_controller { | 
|  | void *acpi_handle; | 
|  | void *iommu; | 
|  | int segment; | 
|  | int node;		/* nearest node with memory or -1 for global allocation */ | 
|  |  | 
|  | unsigned int windows; | 
|  | struct pci_window *window; | 
|  |  | 
|  | void *platform_data; | 
|  | }; | 
|  |  | 
|  | #define PCI_CONTROLLER(busdev) ((struct pci_controller *) busdev->sysdata) | 
|  | #define pci_domain_nr(busdev)    (PCI_CONTROLLER(busdev)->segment) | 
|  |  | 
|  | extern struct pci_ops pci_root_ops; | 
|  |  | 
|  | static inline int pci_proc_domain(struct pci_bus *bus) | 
|  | { | 
|  | return (pci_domain_nr(bus) != 0); | 
|  | } | 
|  |  | 
|  | extern void pcibios_resource_to_bus(struct pci_dev *dev, | 
|  | struct pci_bus_region *region, struct resource *res); | 
|  |  | 
|  | extern void pcibios_bus_to_resource(struct pci_dev *dev, | 
|  | struct resource *res, struct pci_bus_region *region); | 
|  |  | 
|  | static inline struct resource * | 
|  | pcibios_select_root(struct pci_dev *pdev, struct resource *res) | 
|  | { | 
|  | struct resource *root = NULL; | 
|  |  | 
|  | if (res->flags & IORESOURCE_IO) | 
|  | root = &ioport_resource; | 
|  | if (res->flags & IORESOURCE_MEM) | 
|  | root = &iomem_resource; | 
|  |  | 
|  | return root; | 
|  | } | 
|  |  | 
|  | #define pcibios_scan_all_fns(a, b)	0 | 
|  |  | 
|  | #define HAVE_ARCH_PCI_GET_LEGACY_IDE_IRQ | 
|  | static inline int pci_get_legacy_ide_irq(struct pci_dev *dev, int channel) | 
|  | { | 
|  | return channel ? isa_irq_to_vector(15) : isa_irq_to_vector(14); | 
|  | } | 
|  |  | 
|  | #endif /* _ASM_IA64_PCI_H */ |