#ifndef _X8664_DMA_MAPPING_H #define _X8664_DMA_MAPPING_H 1 /* * IOMMU interface. See Documentation/DMA-mapping.txt and DMA-API.txt for * documentation. */ #include #include #include #include extern dma_addr_t bad_dma_address; #define dma_mapping_error(x) \ (swiotlb ? swiotlb_dma_mapping_error(x) : ((x) == bad_dma_address)) void *dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, unsigned gfp); void dma_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle); #ifdef CONFIG_GART_IOMMU extern dma_addr_t dma_map_single(struct device *hwdev, void *ptr, size_t size, int direction); extern void dma_unmap_single(struct device *dev, dma_addr_t addr,size_t size, int direction); #else /* No IOMMU */ static inline dma_addr_t dma_map_single(struct device *hwdev, void *ptr, size_t size, int direction) { dma_addr_t addr; if (direction == DMA_NONE) out_of_line_bug(); addr = virt_to_bus(ptr); if ((addr+size) & ~*hwdev->dma_mask) out_of_line_bug(); return addr; } static inline void dma_unmap_single(struct device *hwdev, dma_addr_t dma_addr, size_t size, int direction) { if (direction == DMA_NONE) out_of_line_bug(); /* Nothing to do */ } #endif #define dma_map_page(dev,page,offset,size,dir) \ dma_map_single((dev), page_address(page)+(offset), (size), (dir)) static inline void dma_sync_single_for_cpu(struct device *hwdev, dma_addr_t dma_handle, size_t size, int direction) { if (direction == DMA_NONE) out_of_line_bug(); if (swiotlb) return swiotlb_sync_single_for_cpu(hwdev,dma_handle,size,direction); flush_write_buffers(); } static inline void dma_sync_single_for_device(struct device *hwdev, dma_addr_t dma_handle, size_t size, int direction) { if (direction == DMA_NONE) out_of_line_bug(); if (swiotlb) return swiotlb_sync_single_for_device(hwdev,dma_handle,size,direction); flush_write_buffers(); } static inline void dma_sync_sg_for_cpu(struct device *hwdev, struct scatterlist *sg, int nelems, int direction) { if (direction == DMA_NONE) out_of_line_bug(); if (swiotlb) return swiotlb_sync_sg_for_cpu(hwdev,sg,nelems,direction); flush_write_buffers(); } static inline void dma_sync_sg_for_device(struct device *hwdev, struct scatterlist *sg, int nelems, int direction) { if (direction == DMA_NONE) out_of_line_bug(); if (swiotlb) return swiotlb_sync_sg_for_device(hwdev,sg,nelems,direction); flush_write_buffers(); } extern int dma_map_sg(struct device *hwdev, struct scatterlist *sg, int nents, int direction); extern void dma_unmap_sg(struct device *hwdev, struct scatterlist *sg, int nents, int direction); #define dma_unmap_page dma_unmap_single extern int dma_supported(struct device *hwdev, u64 mask); extern int dma_get_cache_alignment(void); #define dma_is_consistent(h) 1 static inline int dma_set_mask(struct device *dev, u64 mask) { if (!dev->dma_mask || !dma_supported(dev, mask)) return -EIO; *dev->dma_mask = mask; return 0; } static inline void dma_cache_sync(void *vaddr, size_t size, enum dma_data_direction dir) { flush_write_buffers(); } #endif