blob: 285218a08e8f89cdeee49fe5d3c2acfece1df5ff [file] [log] [blame]
/*
* Copyright (C) 2013 Google, Inc.
*
* This software is licensed under the terms of the GNU General Public
* License version 2, as published by the Free Software Foundation, and
* may be copied, distributed, and modified under those terms.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
*/
#include <linux/dma-buf.h>
#include <linux/highmem.h>
#include <linux/memblock.h>
#include <linux/slab.h>
struct adf_memblock_pdata {
phys_addr_t base;
};
static struct sg_table *adf_memblock_map(struct dma_buf_attachment *attach,
enum dma_data_direction direction)
{
struct adf_memblock_pdata *pdata = attach->dmabuf->priv;
unsigned long pfn = PFN_DOWN(pdata->base);
struct page *page = pfn_to_page(pfn);
struct sg_table *table;
int nents, ret;
table = kzalloc(sizeof(*table), GFP_KERNEL);
if (!table)
return ERR_PTR(-ENOMEM);
ret = sg_alloc_table(table, 1, GFP_KERNEL);
if (ret < 0)
goto err_alloc;
sg_set_page(table->sgl, page, attach->dmabuf->size, 0);
nents = dma_map_sg(attach->dev, table->sgl, 1, direction);
if (!nents) {
ret = -EINVAL;
goto err_map;
}
return table;
err_map:
sg_free_table(table);
err_alloc:
kfree(table);
return ERR_PTR(ret);
}
static void adf_memblock_unmap(struct dma_buf_attachment *attach,
struct sg_table *table, enum dma_data_direction direction)
{
dma_unmap_sg(attach->dev, table->sgl, 1, direction);
sg_free_table(table);
}
static void __init_memblock adf_memblock_release(struct dma_buf *buf)
{
struct adf_memblock_pdata *pdata = buf->priv;
int err = memblock_free(pdata->base, buf->size);
if (err < 0)
pr_warn("%s: freeing memblock failed: %d\n", __func__, err);
kfree(pdata);
}
static void *adf_memblock_do_kmap(struct dma_buf *buf, unsigned long pgoffset,
bool atomic)
{
struct adf_memblock_pdata *pdata = buf->priv;
unsigned long pfn = PFN_DOWN(pdata->base) + pgoffset;
struct page *page = pfn_to_page(pfn);
if (atomic)
return kmap_atomic(page);
else
return kmap(page);
}
static void *adf_memblock_kmap_atomic(struct dma_buf *buf,
unsigned long pgoffset)
{
return adf_memblock_do_kmap(buf, pgoffset, true);
}
static void adf_memblock_kunmap_atomic(struct dma_buf *buf,
unsigned long pgoffset, void *vaddr)
{
kunmap_atomic(vaddr);
}
static void *adf_memblock_kmap(struct dma_buf *buf, unsigned long pgoffset)
{
return adf_memblock_do_kmap(buf, pgoffset, false);
}
static void adf_memblock_kunmap(struct dma_buf *buf, unsigned long pgoffset,
void *vaddr)
{
kunmap(vaddr);
}
static int adf_memblock_mmap(struct dma_buf *buf, struct vm_area_struct *vma)
{
struct adf_memblock_pdata *pdata = buf->priv;
return remap_pfn_range(vma, vma->vm_start, PFN_DOWN(pdata->base),
vma->vm_end - vma->vm_start, vma->vm_page_prot);
}
struct dma_buf_ops adf_memblock_ops = {
.map_dma_buf = adf_memblock_map,
.unmap_dma_buf = adf_memblock_unmap,
.release = adf_memblock_release,
.kmap_atomic = adf_memblock_kmap_atomic,
.kunmap_atomic = adf_memblock_kunmap_atomic,
.kmap = adf_memblock_kmap,
.kunmap = adf_memblock_kunmap,
.mmap = adf_memblock_mmap,
};
/**
* adf_memblock_export - export a memblock reserved area as a dma-buf
*
* @base: base physical address
* @size: memblock size
* @flags: mode flags for the dma-buf's file
*
* @base and @size must be page-aligned.
*
* Returns a dma-buf on success or ERR_PTR(-errno) on failure.
*/
struct dma_buf *adf_memblock_export(phys_addr_t base, size_t size, int flags)
{
struct adf_memblock_pdata *pdata;
struct dma_buf *buf;
DEFINE_DMA_BUF_EXPORT_INFO(exp_info);
if (PAGE_ALIGN(base) != base || PAGE_ALIGN(size) != size)
return ERR_PTR(-EINVAL);
pdata = kzalloc(sizeof(*pdata), GFP_KERNEL);
if (!pdata)
return ERR_PTR(-ENOMEM);
pdata->base = base;
exp_info.ops = &adf_memblock_ops;
exp_info.size = size;
exp_info.flags = flags;
exp_info.priv = pdata;
buf = dma_buf_export(&exp_info);
if (IS_ERR(buf))
kfree(pdata);
return buf;
}
EXPORT_SYMBOL(adf_memblock_export);