|  | /* | 
|  | * Machine vector for IA-64. | 
|  | * | 
|  | * Copyright (C) 1999 Silicon Graphics, Inc. | 
|  | * Copyright (C) Srinivasa Thirumalachar <sprasad@engr.sgi.com> | 
|  | * Copyright (C) Vijay Chander <vijay@engr.sgi.com> | 
|  | * Copyright (C) 1999-2001, 2003-2004 Hewlett-Packard Co. | 
|  | *	David Mosberger-Tang <davidm@hpl.hp.com> | 
|  | */ | 
|  | #ifndef _ASM_IA64_MACHVEC_H | 
|  | #define _ASM_IA64_MACHVEC_H | 
|  |  | 
|  | #include <linux/config.h> | 
|  | #include <linux/types.h> | 
|  |  | 
|  | /* forward declarations: */ | 
|  | struct device; | 
|  | struct pt_regs; | 
|  | struct scatterlist; | 
|  | struct page; | 
|  | struct mm_struct; | 
|  | struct pci_bus; | 
|  |  | 
|  | typedef void ia64_mv_setup_t (char **); | 
|  | typedef void ia64_mv_cpu_init_t (void); | 
|  | typedef void ia64_mv_irq_init_t (void); | 
|  | typedef void ia64_mv_send_ipi_t (int, int, int, int); | 
|  | typedef void ia64_mv_timer_interrupt_t (int, void *, struct pt_regs *); | 
|  | typedef void ia64_mv_global_tlb_purge_t (unsigned long, unsigned long, unsigned long); | 
|  | typedef void ia64_mv_tlb_migrate_finish_t (struct mm_struct *); | 
|  | typedef unsigned int ia64_mv_local_vector_to_irq (u8); | 
|  | typedef char *ia64_mv_pci_get_legacy_mem_t (struct pci_bus *); | 
|  | typedef int ia64_mv_pci_legacy_read_t (struct pci_bus *, u16 port, u32 *val, | 
|  | u8 size); | 
|  | typedef int ia64_mv_pci_legacy_write_t (struct pci_bus *, u16 port, u32 val, | 
|  | u8 size); | 
|  |  | 
|  | /* DMA-mapping interface: */ | 
|  | typedef void ia64_mv_dma_init (void); | 
|  | typedef void *ia64_mv_dma_alloc_coherent (struct device *, size_t, dma_addr_t *, int); | 
|  | typedef void ia64_mv_dma_free_coherent (struct device *, size_t, void *, dma_addr_t); | 
|  | typedef dma_addr_t ia64_mv_dma_map_single (struct device *, void *, size_t, int); | 
|  | typedef void ia64_mv_dma_unmap_single (struct device *, dma_addr_t, size_t, int); | 
|  | typedef int ia64_mv_dma_map_sg (struct device *, struct scatterlist *, int, int); | 
|  | typedef void ia64_mv_dma_unmap_sg (struct device *, struct scatterlist *, int, int); | 
|  | typedef void ia64_mv_dma_sync_single_for_cpu (struct device *, dma_addr_t, size_t, int); | 
|  | typedef void ia64_mv_dma_sync_sg_for_cpu (struct device *, struct scatterlist *, int, int); | 
|  | typedef void ia64_mv_dma_sync_single_for_device (struct device *, dma_addr_t, size_t, int); | 
|  | typedef void ia64_mv_dma_sync_sg_for_device (struct device *, struct scatterlist *, int, int); | 
|  | typedef int ia64_mv_dma_mapping_error (dma_addr_t dma_addr); | 
|  | typedef int ia64_mv_dma_supported (struct device *, u64); | 
|  |  | 
|  | /* | 
|  | * WARNING: The legacy I/O space is _architected_.  Platforms are | 
|  | * expected to follow this architected model (see Section 10.7 in the | 
|  | * IA-64 Architecture Software Developer's Manual).  Unfortunately, | 
|  | * some broken machines do not follow that model, which is why we have | 
|  | * to make the inX/outX operations part of the machine vector. | 
|  | * Platform designers should follow the architected model whenever | 
|  | * possible. | 
|  | */ | 
|  | typedef unsigned int ia64_mv_inb_t (unsigned long); | 
|  | typedef unsigned int ia64_mv_inw_t (unsigned long); | 
|  | typedef unsigned int ia64_mv_inl_t (unsigned long); | 
|  | typedef void ia64_mv_outb_t (unsigned char, unsigned long); | 
|  | typedef void ia64_mv_outw_t (unsigned short, unsigned long); | 
|  | typedef void ia64_mv_outl_t (unsigned int, unsigned long); | 
|  | typedef void ia64_mv_mmiowb_t (void); | 
|  | typedef unsigned char ia64_mv_readb_t (const volatile void __iomem *); | 
|  | typedef unsigned short ia64_mv_readw_t (const volatile void __iomem *); | 
|  | typedef unsigned int ia64_mv_readl_t (const volatile void __iomem *); | 
|  | typedef unsigned long ia64_mv_readq_t (const volatile void __iomem *); | 
|  | typedef unsigned char ia64_mv_readb_relaxed_t (const volatile void __iomem *); | 
|  | typedef unsigned short ia64_mv_readw_relaxed_t (const volatile void __iomem *); | 
|  | typedef unsigned int ia64_mv_readl_relaxed_t (const volatile void __iomem *); | 
|  | typedef unsigned long ia64_mv_readq_relaxed_t (const volatile void __iomem *); | 
|  |  | 
|  | static inline void | 
|  | machvec_noop (void) | 
|  | { | 
|  | } | 
|  |  | 
|  | static inline void | 
|  | machvec_noop_mm (struct mm_struct *mm) | 
|  | { | 
|  | } | 
|  |  | 
|  | extern void machvec_setup (char **); | 
|  | extern void machvec_timer_interrupt (int, void *, struct pt_regs *); | 
|  | extern void machvec_dma_sync_single (struct device *, dma_addr_t, size_t, int); | 
|  | extern void machvec_dma_sync_sg (struct device *, struct scatterlist *, int, int); | 
|  | extern void machvec_tlb_migrate_finish (struct mm_struct *); | 
|  |  | 
|  | # if defined (CONFIG_IA64_HP_SIM) | 
|  | #  include <asm/machvec_hpsim.h> | 
|  | # elif defined (CONFIG_IA64_DIG) | 
|  | #  include <asm/machvec_dig.h> | 
|  | # elif defined (CONFIG_IA64_HP_ZX1) | 
|  | #  include <asm/machvec_hpzx1.h> | 
|  | # elif defined (CONFIG_IA64_HP_ZX1_SWIOTLB) | 
|  | #  include <asm/machvec_hpzx1_swiotlb.h> | 
|  | # elif defined (CONFIG_IA64_SGI_SN2) | 
|  | #  include <asm/machvec_sn2.h> | 
|  | # elif defined (CONFIG_IA64_GENERIC) | 
|  |  | 
|  | # ifdef MACHVEC_PLATFORM_HEADER | 
|  | #  include MACHVEC_PLATFORM_HEADER | 
|  | # else | 
|  | #  define platform_name		ia64_mv.name | 
|  | #  define platform_setup	ia64_mv.setup | 
|  | #  define platform_cpu_init	ia64_mv.cpu_init | 
|  | #  define platform_irq_init	ia64_mv.irq_init | 
|  | #  define platform_send_ipi	ia64_mv.send_ipi | 
|  | #  define platform_timer_interrupt	ia64_mv.timer_interrupt | 
|  | #  define platform_global_tlb_purge	ia64_mv.global_tlb_purge | 
|  | #  define platform_tlb_migrate_finish	ia64_mv.tlb_migrate_finish | 
|  | #  define platform_dma_init		ia64_mv.dma_init | 
|  | #  define platform_dma_alloc_coherent	ia64_mv.dma_alloc_coherent | 
|  | #  define platform_dma_free_coherent	ia64_mv.dma_free_coherent | 
|  | #  define platform_dma_map_single	ia64_mv.dma_map_single | 
|  | #  define platform_dma_unmap_single	ia64_mv.dma_unmap_single | 
|  | #  define platform_dma_map_sg		ia64_mv.dma_map_sg | 
|  | #  define platform_dma_unmap_sg		ia64_mv.dma_unmap_sg | 
|  | #  define platform_dma_sync_single_for_cpu ia64_mv.dma_sync_single_for_cpu | 
|  | #  define platform_dma_sync_sg_for_cpu	ia64_mv.dma_sync_sg_for_cpu | 
|  | #  define platform_dma_sync_single_for_device ia64_mv.dma_sync_single_for_device | 
|  | #  define platform_dma_sync_sg_for_device ia64_mv.dma_sync_sg_for_device | 
|  | #  define platform_dma_mapping_error		ia64_mv.dma_mapping_error | 
|  | #  define platform_dma_supported	ia64_mv.dma_supported | 
|  | #  define platform_local_vector_to_irq	ia64_mv.local_vector_to_irq | 
|  | #  define platform_pci_get_legacy_mem	ia64_mv.pci_get_legacy_mem | 
|  | #  define platform_pci_legacy_read	ia64_mv.pci_legacy_read | 
|  | #  define platform_pci_legacy_write	ia64_mv.pci_legacy_write | 
|  | #  define platform_inb		ia64_mv.inb | 
|  | #  define platform_inw		ia64_mv.inw | 
|  | #  define platform_inl		ia64_mv.inl | 
|  | #  define platform_outb		ia64_mv.outb | 
|  | #  define platform_outw		ia64_mv.outw | 
|  | #  define platform_outl		ia64_mv.outl | 
|  | #  define platform_mmiowb	ia64_mv.mmiowb | 
|  | #  define platform_readb        ia64_mv.readb | 
|  | #  define platform_readw        ia64_mv.readw | 
|  | #  define platform_readl        ia64_mv.readl | 
|  | #  define platform_readq        ia64_mv.readq | 
|  | #  define platform_readb_relaxed        ia64_mv.readb_relaxed | 
|  | #  define platform_readw_relaxed        ia64_mv.readw_relaxed | 
|  | #  define platform_readl_relaxed        ia64_mv.readl_relaxed | 
|  | #  define platform_readq_relaxed        ia64_mv.readq_relaxed | 
|  | # endif | 
|  |  | 
|  | /* __attribute__((__aligned__(16))) is required to make size of the | 
|  | * structure multiple of 16 bytes. | 
|  | * This will fillup the holes created because of section 3.3.1 in | 
|  | * Software Conventions guide. | 
|  | */ | 
|  | struct ia64_machine_vector { | 
|  | const char *name; | 
|  | ia64_mv_setup_t *setup; | 
|  | ia64_mv_cpu_init_t *cpu_init; | 
|  | ia64_mv_irq_init_t *irq_init; | 
|  | ia64_mv_send_ipi_t *send_ipi; | 
|  | ia64_mv_timer_interrupt_t *timer_interrupt; | 
|  | ia64_mv_global_tlb_purge_t *global_tlb_purge; | 
|  | ia64_mv_tlb_migrate_finish_t *tlb_migrate_finish; | 
|  | ia64_mv_dma_init *dma_init; | 
|  | ia64_mv_dma_alloc_coherent *dma_alloc_coherent; | 
|  | ia64_mv_dma_free_coherent *dma_free_coherent; | 
|  | ia64_mv_dma_map_single *dma_map_single; | 
|  | ia64_mv_dma_unmap_single *dma_unmap_single; | 
|  | ia64_mv_dma_map_sg *dma_map_sg; | 
|  | ia64_mv_dma_unmap_sg *dma_unmap_sg; | 
|  | ia64_mv_dma_sync_single_for_cpu *dma_sync_single_for_cpu; | 
|  | ia64_mv_dma_sync_sg_for_cpu *dma_sync_sg_for_cpu; | 
|  | ia64_mv_dma_sync_single_for_device *dma_sync_single_for_device; | 
|  | ia64_mv_dma_sync_sg_for_device *dma_sync_sg_for_device; | 
|  | ia64_mv_dma_mapping_error *dma_mapping_error; | 
|  | ia64_mv_dma_supported *dma_supported; | 
|  | ia64_mv_local_vector_to_irq *local_vector_to_irq; | 
|  | ia64_mv_pci_get_legacy_mem_t *pci_get_legacy_mem; | 
|  | ia64_mv_pci_legacy_read_t *pci_legacy_read; | 
|  | ia64_mv_pci_legacy_write_t *pci_legacy_write; | 
|  | ia64_mv_inb_t *inb; | 
|  | ia64_mv_inw_t *inw; | 
|  | ia64_mv_inl_t *inl; | 
|  | ia64_mv_outb_t *outb; | 
|  | ia64_mv_outw_t *outw; | 
|  | ia64_mv_outl_t *outl; | 
|  | ia64_mv_mmiowb_t *mmiowb; | 
|  | ia64_mv_readb_t *readb; | 
|  | ia64_mv_readw_t *readw; | 
|  | ia64_mv_readl_t *readl; | 
|  | ia64_mv_readq_t *readq; | 
|  | ia64_mv_readb_relaxed_t *readb_relaxed; | 
|  | ia64_mv_readw_relaxed_t *readw_relaxed; | 
|  | ia64_mv_readl_relaxed_t *readl_relaxed; | 
|  | ia64_mv_readq_relaxed_t *readq_relaxed; | 
|  | } __attribute__((__aligned__(16))); /* align attrib? see above comment */ | 
|  |  | 
|  | #define MACHVEC_INIT(name)			\ | 
|  | {						\ | 
|  | #name,					\ | 
|  | platform_setup,				\ | 
|  | platform_cpu_init,			\ | 
|  | platform_irq_init,			\ | 
|  | platform_send_ipi,			\ | 
|  | platform_timer_interrupt,		\ | 
|  | platform_global_tlb_purge,		\ | 
|  | platform_tlb_migrate_finish,		\ | 
|  | platform_dma_init,			\ | 
|  | platform_dma_alloc_coherent,		\ | 
|  | platform_dma_free_coherent,		\ | 
|  | platform_dma_map_single,		\ | 
|  | platform_dma_unmap_single,		\ | 
|  | platform_dma_map_sg,			\ | 
|  | platform_dma_unmap_sg,			\ | 
|  | platform_dma_sync_single_for_cpu,	\ | 
|  | platform_dma_sync_sg_for_cpu,		\ | 
|  | platform_dma_sync_single_for_device,	\ | 
|  | platform_dma_sync_sg_for_device,	\ | 
|  | platform_dma_mapping_error,			\ | 
|  | platform_dma_supported,			\ | 
|  | platform_local_vector_to_irq,		\ | 
|  | platform_pci_get_legacy_mem,		\ | 
|  | platform_pci_legacy_read,		\ | 
|  | platform_pci_legacy_write,		\ | 
|  | platform_inb,				\ | 
|  | platform_inw,				\ | 
|  | platform_inl,				\ | 
|  | platform_outb,				\ | 
|  | platform_outw,				\ | 
|  | platform_outl,				\ | 
|  | platform_mmiowb,			\ | 
|  | platform_readb,				\ | 
|  | platform_readw,				\ | 
|  | platform_readl,				\ | 
|  | platform_readq,				\ | 
|  | platform_readb_relaxed,			\ | 
|  | platform_readw_relaxed,			\ | 
|  | platform_readl_relaxed,			\ | 
|  | platform_readq_relaxed,			\ | 
|  | } | 
|  |  | 
|  | extern struct ia64_machine_vector ia64_mv; | 
|  | extern void machvec_init (const char *name); | 
|  |  | 
|  | # else | 
|  | #  error Unknown configuration.  Update asm-ia64/machvec.h. | 
|  | # endif /* CONFIG_IA64_GENERIC */ | 
|  |  | 
|  | /* | 
|  | * Declare default routines which aren't declared anywhere else: | 
|  | */ | 
|  | extern ia64_mv_dma_init			swiotlb_init; | 
|  | extern ia64_mv_dma_alloc_coherent	swiotlb_alloc_coherent; | 
|  | extern ia64_mv_dma_free_coherent	swiotlb_free_coherent; | 
|  | extern ia64_mv_dma_map_single		swiotlb_map_single; | 
|  | extern ia64_mv_dma_unmap_single		swiotlb_unmap_single; | 
|  | extern ia64_mv_dma_map_sg		swiotlb_map_sg; | 
|  | extern ia64_mv_dma_unmap_sg		swiotlb_unmap_sg; | 
|  | extern ia64_mv_dma_sync_single_for_cpu	swiotlb_sync_single_for_cpu; | 
|  | extern ia64_mv_dma_sync_sg_for_cpu	swiotlb_sync_sg_for_cpu; | 
|  | extern ia64_mv_dma_sync_single_for_device swiotlb_sync_single_for_device; | 
|  | extern ia64_mv_dma_sync_sg_for_device	swiotlb_sync_sg_for_device; | 
|  | extern ia64_mv_dma_mapping_error	swiotlb_dma_mapping_error; | 
|  | extern ia64_mv_dma_supported		swiotlb_dma_supported; | 
|  |  | 
|  | /* | 
|  | * Define default versions so we can extend machvec for new platforms without having | 
|  | * to update the machvec files for all existing platforms. | 
|  | */ | 
|  | #ifndef platform_setup | 
|  | # define platform_setup			machvec_setup | 
|  | #endif | 
|  | #ifndef platform_cpu_init | 
|  | # define platform_cpu_init		machvec_noop | 
|  | #endif | 
|  | #ifndef platform_irq_init | 
|  | # define platform_irq_init		machvec_noop | 
|  | #endif | 
|  |  | 
|  | #ifndef platform_send_ipi | 
|  | # define platform_send_ipi		ia64_send_ipi	/* default to architected version */ | 
|  | #endif | 
|  | #ifndef platform_timer_interrupt | 
|  | # define platform_timer_interrupt 	machvec_timer_interrupt | 
|  | #endif | 
|  | #ifndef platform_global_tlb_purge | 
|  | # define platform_global_tlb_purge	ia64_global_tlb_purge /* default to architected version */ | 
|  | #endif | 
|  | #ifndef platform_tlb_migrate_finish | 
|  | # define platform_tlb_migrate_finish	machvec_noop_mm | 
|  | #endif | 
|  | #ifndef platform_dma_init | 
|  | # define platform_dma_init		swiotlb_init | 
|  | #endif | 
|  | #ifndef platform_dma_alloc_coherent | 
|  | # define platform_dma_alloc_coherent	swiotlb_alloc_coherent | 
|  | #endif | 
|  | #ifndef platform_dma_free_coherent | 
|  | # define platform_dma_free_coherent	swiotlb_free_coherent | 
|  | #endif | 
|  | #ifndef platform_dma_map_single | 
|  | # define platform_dma_map_single	swiotlb_map_single | 
|  | #endif | 
|  | #ifndef platform_dma_unmap_single | 
|  | # define platform_dma_unmap_single	swiotlb_unmap_single | 
|  | #endif | 
|  | #ifndef platform_dma_map_sg | 
|  | # define platform_dma_map_sg		swiotlb_map_sg | 
|  | #endif | 
|  | #ifndef platform_dma_unmap_sg | 
|  | # define platform_dma_unmap_sg		swiotlb_unmap_sg | 
|  | #endif | 
|  | #ifndef platform_dma_sync_single_for_cpu | 
|  | # define platform_dma_sync_single_for_cpu	swiotlb_sync_single_for_cpu | 
|  | #endif | 
|  | #ifndef platform_dma_sync_sg_for_cpu | 
|  | # define platform_dma_sync_sg_for_cpu		swiotlb_sync_sg_for_cpu | 
|  | #endif | 
|  | #ifndef platform_dma_sync_single_for_device | 
|  | # define platform_dma_sync_single_for_device	swiotlb_sync_single_for_device | 
|  | #endif | 
|  | #ifndef platform_dma_sync_sg_for_device | 
|  | # define platform_dma_sync_sg_for_device	swiotlb_sync_sg_for_device | 
|  | #endif | 
|  | #ifndef platform_dma_mapping_error | 
|  | # define platform_dma_mapping_error		swiotlb_dma_mapping_error | 
|  | #endif | 
|  | #ifndef platform_dma_supported | 
|  | # define  platform_dma_supported	swiotlb_dma_supported | 
|  | #endif | 
|  | #ifndef platform_local_vector_to_irq | 
|  | # define platform_local_vector_to_irq	__ia64_local_vector_to_irq | 
|  | #endif | 
|  | #ifndef platform_pci_get_legacy_mem | 
|  | # define platform_pci_get_legacy_mem	ia64_pci_get_legacy_mem | 
|  | #endif | 
|  | #ifndef platform_pci_legacy_read | 
|  | # define platform_pci_legacy_read	ia64_pci_legacy_read | 
|  | #endif | 
|  | #ifndef platform_pci_legacy_write | 
|  | # define platform_pci_legacy_write	ia64_pci_legacy_write | 
|  | #endif | 
|  | #ifndef platform_inb | 
|  | # define platform_inb		__ia64_inb | 
|  | #endif | 
|  | #ifndef platform_inw | 
|  | # define platform_inw		__ia64_inw | 
|  | #endif | 
|  | #ifndef platform_inl | 
|  | # define platform_inl		__ia64_inl | 
|  | #endif | 
|  | #ifndef platform_outb | 
|  | # define platform_outb		__ia64_outb | 
|  | #endif | 
|  | #ifndef platform_outw | 
|  | # define platform_outw		__ia64_outw | 
|  | #endif | 
|  | #ifndef platform_outl | 
|  | # define platform_outl		__ia64_outl | 
|  | #endif | 
|  | #ifndef platform_mmiowb | 
|  | # define platform_mmiowb	__ia64_mmiowb | 
|  | #endif | 
|  | #ifndef platform_readb | 
|  | # define platform_readb		__ia64_readb | 
|  | #endif | 
|  | #ifndef platform_readw | 
|  | # define platform_readw		__ia64_readw | 
|  | #endif | 
|  | #ifndef platform_readl | 
|  | # define platform_readl		__ia64_readl | 
|  | #endif | 
|  | #ifndef platform_readq | 
|  | # define platform_readq		__ia64_readq | 
|  | #endif | 
|  | #ifndef platform_readb_relaxed | 
|  | # define platform_readb_relaxed	__ia64_readb_relaxed | 
|  | #endif | 
|  | #ifndef platform_readw_relaxed | 
|  | # define platform_readw_relaxed	__ia64_readw_relaxed | 
|  | #endif | 
|  | #ifndef platform_readl_relaxed | 
|  | # define platform_readl_relaxed	__ia64_readl_relaxed | 
|  | #endif | 
|  | #ifndef platform_readq_relaxed | 
|  | # define platform_readq_relaxed	__ia64_readq_relaxed | 
|  | #endif | 
|  |  | 
|  | #endif /* _ASM_IA64_MACHVEC_H */ |