···342342}343343EXPORT_SYMBOL(dma_map_single);344344345345+/*346346+ * see if a mapped address was really a "safe" buffer and if so, copy347347+ * the data from the safe buffer back to the unsafe buffer and free up348348+ * the safe buffer. (basically return things back to the way they349349+ * should be)350350+ */351351+void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size,352352+ enum dma_data_direction dir)353353+{354354+ dev_dbg(dev, "%s(ptr=%p,size=%d,dir=%x)\n",355355+ __func__, (void *) dma_addr, size, dir);356356+357357+ unmap_single(dev, dma_addr, size, dir);358358+}359359+EXPORT_SYMBOL(dma_unmap_single);360360+345361dma_addr_t dma_map_page(struct device *dev, struct page *page,346362 unsigned long offset, size_t size, enum dma_data_direction dir)347363{···382366 * the safe buffer. (basically return things back to the way they383367 * should be)384368 */385385-386386-void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size,369369+void dma_unmap_page(struct device *dev, dma_addr_t dma_addr, size_t size,387370 enum dma_data_direction dir)388371{389372 dev_dbg(dev, "%s(ptr=%p,size=%d,dir=%x)\n",···390375391376 unmap_single(dev, dma_addr, size, dir);392377}393393-EXPORT_SYMBOL(dma_unmap_single);378378+EXPORT_SYMBOL(dma_unmap_page);394379395380int dmabounce_sync_for_cpu(struct device *dev, dma_addr_t addr,396381 unsigned long off, size_t sz, enum dma_data_direction dir)