| /* | 
 |  * Copyright (C) 2006 Benjamin Herrenschmidt, IBM Corporation | 
 |  * | 
 |  * Provide default implementations of the DMA mapping callbacks for | 
 |  * busses using the iommu infrastructure | 
 |  */ | 
 |  | 
 | #include <asm/iommu.h> | 
 |  | 
 | /* | 
 |  * Generic iommu implementation | 
 |  */ | 
 |  | 
 | /* Allocates a contiguous real buffer and creates mappings over it. | 
 |  * Returns the virtual address of the buffer and sets dma_handle | 
 |  * to the dma address (mapping) of the first page. | 
 |  */ | 
 | static void *dma_iommu_alloc_coherent(struct device *dev, size_t size, | 
 | 				      dma_addr_t *dma_handle, gfp_t flag) | 
 | { | 
 | 	return iommu_alloc_coherent(dev, get_iommu_table_base(dev), size, | 
 | 				    dma_handle, device_to_mask(dev), flag, | 
 | 				    dev_to_node(dev)); | 
 | } | 
 |  | 
 | static void dma_iommu_free_coherent(struct device *dev, size_t size, | 
 | 				    void *vaddr, dma_addr_t dma_handle) | 
 | { | 
 | 	iommu_free_coherent(get_iommu_table_base(dev), size, vaddr, dma_handle); | 
 | } | 
 |  | 
 | /* Creates TCEs for a user provided buffer.  The user buffer must be | 
 |  * contiguous real kernel storage (not vmalloc).  The address passed here | 
 |  * comprises a page address and offset into that page. The dma_addr_t | 
 |  * returned will point to the same byte within the page as was passed in. | 
 |  */ | 
 | static dma_addr_t dma_iommu_map_page(struct device *dev, struct page *page, | 
 | 				     unsigned long offset, size_t size, | 
 | 				     enum dma_data_direction direction, | 
 | 				     struct dma_attrs *attrs) | 
 | { | 
 | 	return iommu_map_page(dev, get_iommu_table_base(dev), page, offset, | 
 | 			      size, device_to_mask(dev), direction, attrs); | 
 | } | 
 |  | 
 |  | 
 | static void dma_iommu_unmap_page(struct device *dev, dma_addr_t dma_handle, | 
 | 				 size_t size, enum dma_data_direction direction, | 
 | 				 struct dma_attrs *attrs) | 
 | { | 
 | 	iommu_unmap_page(get_iommu_table_base(dev), dma_handle, size, direction, | 
 | 			 attrs); | 
 | } | 
 |  | 
 |  | 
 | static int dma_iommu_map_sg(struct device *dev, struct scatterlist *sglist, | 
 | 			    int nelems, enum dma_data_direction direction, | 
 | 			    struct dma_attrs *attrs) | 
 | { | 
 | 	return iommu_map_sg(dev, get_iommu_table_base(dev), sglist, nelems, | 
 | 			    device_to_mask(dev), direction, attrs); | 
 | } | 
 |  | 
 | static void dma_iommu_unmap_sg(struct device *dev, struct scatterlist *sglist, | 
 | 		int nelems, enum dma_data_direction direction, | 
 | 		struct dma_attrs *attrs) | 
 | { | 
 | 	iommu_unmap_sg(get_iommu_table_base(dev), sglist, nelems, direction, | 
 | 		       attrs); | 
 | } | 
 |  | 
 | /* We support DMA to/from any memory page via the iommu */ | 
 | static int dma_iommu_dma_supported(struct device *dev, u64 mask) | 
 | { | 
 | 	struct iommu_table *tbl = get_iommu_table_base(dev); | 
 |  | 
 | 	if (!tbl) { | 
 | 		dev_info(dev, "Warning: IOMMU dma not supported: mask 0x%08llx" | 
 | 			", table unavailable\n", mask); | 
 | 		return 0; | 
 | 	} | 
 |  | 
 | 	if ((tbl->it_offset + tbl->it_size) > (mask >> IOMMU_PAGE_SHIFT)) { | 
 | 		dev_info(dev, "Warning: IOMMU window too big for device mask\n"); | 
 | 		dev_info(dev, "mask: 0x%08llx, table end: 0x%08lx\n", | 
 | 				mask, (tbl->it_offset + tbl->it_size) << | 
 | 				IOMMU_PAGE_SHIFT); | 
 | 		return 0; | 
 | 	} else | 
 | 		return 1; | 
 | } | 
 |  | 
 | struct dma_map_ops dma_iommu_ops = { | 
 | 	.alloc_coherent	= dma_iommu_alloc_coherent, | 
 | 	.free_coherent	= dma_iommu_free_coherent, | 
 | 	.map_sg		= dma_iommu_map_sg, | 
 | 	.unmap_sg	= dma_iommu_unmap_sg, | 
 | 	.dma_supported	= dma_iommu_dma_supported, | 
 | 	.map_page	= dma_iommu_map_page, | 
 | 	.unmap_page	= dma_iommu_unmap_page, | 
 | }; | 
 | EXPORT_SYMBOL(dma_iommu_ops); |