123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778 |
- From 2c58080407554e1bac8fd50d23cb02420524caed Mon Sep 17 00:00:00 2001
- From: Felix Fietkau <nbd@nbd.name>
- Date: Mon, 12 Aug 2013 12:50:22 +0200
- Subject: [PATCH] MIPS: partially inline dma ops
- Several DMA ops are no-op on many platforms, and the indirection through
- the mips_dma_map_ops function table is causing the compiler to emit
- unnecessary code.
- Inlining visibly improves network performance in my tests (on a 24Kc
- based system), and also slightly reduces code size of a few drivers.
- Signed-off-by: Felix Fietkau <nbd@nbd.name>
- ---
- arch/mips/Kconfig | 4 +
- arch/mips/include/asm/dma-mapping.h | 360 +++++++++++++++++++++++++++++++++++-
- arch/mips/mm/dma-default.c | 163 ++--------------
- 3 files changed, 373 insertions(+), 154 deletions(-)
- --- a/arch/mips/Kconfig
- +++ b/arch/mips/Kconfig
- @@ -1618,6 +1618,7 @@ config CPU_CAVIUM_OCTEON
- select USB_EHCI_BIG_ENDIAN_MMIO if CPU_BIG_ENDIAN
- select USB_OHCI_BIG_ENDIAN_MMIO if CPU_BIG_ENDIAN
- select MIPS_L1_CACHE_SHIFT_7
- + select SYS_HAS_DMA_OPS
- help
- The Cavium Octeon processor is a highly integrated chip containing
- many ethernet hardware widgets for networking tasks. The processor
- @@ -1913,6 +1914,9 @@ config MIPS_MALTA_PM
- bool
- default y
-
- +config SYS_HAS_DMA_OPS
- + bool
- +
- #
- # CPU may reorder R->R, R->W, W->R, W->W
- # Reordering beyond LL and SC is handled in WEAK_REORDERING_BEYOND_LLSC
- --- a/arch/mips/include/asm/dma-mapping.h
- +++ b/arch/mips/include/asm/dma-mapping.h
- @@ -1,9 +1,16 @@
- #ifndef _ASM_DMA_MAPPING_H
- #define _ASM_DMA_MAPPING_H
-
- +#include <linux/kmemcheck.h>
- +#include <linux/bug.h>
- #include <linux/scatterlist.h>
- +#include <linux/dma-debug.h>
- +#include <linux/dma-attrs.h>
- +
- #include <asm/dma-coherence.h>
- #include <asm/cache.h>
- +#include <asm/cpu-type.h>
- +#include <asm-generic/dma-coherent.h>
-
- #ifndef CONFIG_SGI_IP27 /* Kludge to fix 2.6.39 build for IP27 */
- #include <dma-coherence.h>
- @@ -11,12 +18,53 @@
-
- extern struct dma_map_ops *mips_dma_map_ops;
-
- +void __dma_sync(struct page *page, unsigned long offset, size_t size,
- + enum dma_data_direction direction);
- +void *mips_dma_alloc_coherent(struct device *dev, size_t size,
- + dma_addr_t *dma_handle, gfp_t gfp,
- + struct dma_attrs *attrs);
- +void mips_dma_free_coherent(struct device *dev, size_t size, void *vaddr,
- + dma_addr_t dma_handle, struct dma_attrs *attrs);
- +
- static inline struct dma_map_ops *get_dma_ops(struct device *dev)
- {
- +#ifdef CONFIG_SYS_HAS_DMA_OPS
- if (dev && dev->archdata.dma_ops)
- return dev->archdata.dma_ops;
- else
- return mips_dma_map_ops;
- +#else
- + return NULL;
- +#endif
- +}
- +
- +/*
- + * The affected CPUs below in 'cpu_needs_post_dma_flush()' can
- + * speculatively fill random cachelines with stale data at any time,
- + * requiring an extra flush post-DMA.
- + *
- + * Warning on the terminology - Linux calls an uncached area coherent;
- + * MIPS terminology calls memory areas with hardware maintained coherency
- + * coherent.
- + *
- + * Note that the R14000 and R16000 should also be checked for in this
- + * condition. However this function is only called on non-I/O-coherent
- + * systems and only the R10000 and R12000 are used in such systems, the
- + * SGI IP28 Indigo² rsp. SGI IP32 aka O2.
- + */
- +static inline int cpu_needs_post_dma_flush(struct device *dev)
- +{
- + return !plat_device_is_coherent(dev) &&
- + (boot_cpu_type() == CPU_R10000 ||
- + boot_cpu_type() == CPU_R12000 ||
- + boot_cpu_type() == CPU_BMIPS5000);
- +}
- +
- +static inline struct page *dma_addr_to_page(struct device *dev,
- + dma_addr_t dma_addr)
- +{
- + return pfn_to_page(
- + plat_dma_addr_to_phys(dev, dma_addr) >> PAGE_SHIFT);
- }
-
- static inline bool dma_capable(struct device *dev, dma_addr_t addr, size_t size)
- @@ -29,9 +77,399 @@ static inline bool dma_capable(struct de
-
- static inline void dma_mark_clean(void *addr, size_t size) {}
-
- -#include <asm-generic/dma-mapping-common.h>
- +static inline dma_addr_t dma_map_single_attrs(struct device *dev, void *ptr,
- + size_t size,
- + enum dma_data_direction dir,
- + struct dma_attrs *attrs)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- + unsigned long offset = (unsigned long)ptr & ~PAGE_MASK;
- + struct page *page = virt_to_page(ptr);
- + dma_addr_t addr;
- +
- + kmemcheck_mark_initialized(ptr, size);
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops) {
- + addr = ops->map_page(dev, page, offset, size, dir, attrs);
- + } else {
- + if (!plat_device_is_coherent(dev))
- + __dma_sync(page, offset, size, dir);
- +
- + addr = plat_map_dma_mem_page(dev, page) + offset;
- + }
- + debug_dma_map_page(dev, page, offset, size, dir, addr, true);
- + return addr;
- +}
- +
- +static inline void dma_unmap_single_attrs(struct device *dev, dma_addr_t addr,
- + size_t size,
- + enum dma_data_direction dir,
- + struct dma_attrs *attrs)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops) {
- + ops->unmap_page(dev, addr, size, dir, attrs);
- + } else {
- + if (cpu_needs_post_dma_flush(dev))
- + __dma_sync(dma_addr_to_page(dev, addr),
- + addr & ~PAGE_MASK, size, dir);
- + plat_post_dma_flush(dev);
- + plat_unmap_dma_mem(dev, addr, size, dir);
- + }
- + debug_dma_unmap_page(dev, addr, size, dir, true);
- +}
- +
- +/*
- + * dma_maps_sg_attrs returns 0 on error and > 0 on success.
- + * It should never return a value < 0.
- + */
- +static inline int dma_map_sg_attrs(struct device *dev, struct scatterlist *sg,
- + int nents, enum dma_data_direction dir,
- + struct dma_attrs *attrs)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- + int i, ents;
- + struct scatterlist *s;
- +
- + for_each_sg(sg, s, nents, i)
- + kmemcheck_mark_initialized(sg_virt(s), s->length);
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops) {
- + ents = ops->map_sg(dev, sg, nents, dir, attrs);
- + } else {
- + for_each_sg(sg, s, nents, i) {
- + struct page *page = sg_page(s);
- +
- + if (!plat_device_is_coherent(dev))
- + __dma_sync(page, s->offset, s->length, dir);
- +#ifdef CONFIG_NEED_SG_DMA_LENGTH
- + s->dma_length = s->length;
- +#endif
- + s->dma_address =
- + plat_map_dma_mem_page(dev, page) + s->offset;
- + }
- + ents = nents;
- + }
- + BUG_ON(ents < 0);
- + debug_dma_map_sg(dev, sg, nents, ents, dir);
- +
- + return ents;
- +}
- +
- +static inline void dma_unmap_sg_attrs(struct device *dev, struct scatterlist *sg,
- + int nents, enum dma_data_direction dir,
- + struct dma_attrs *attrs)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- + struct scatterlist *s;
- + int i;
- +
- + BUG_ON(!valid_dma_direction(dir));
- + debug_dma_unmap_sg(dev, sg, nents, dir);
- + if (ops) {
- + ops->unmap_sg(dev, sg, nents, dir, attrs);
- + return;
- + }
- + for_each_sg(sg, s, nents, i) {
- + if (!plat_device_is_coherent(dev) && dir != DMA_TO_DEVICE)
- + __dma_sync(sg_page(s), s->offset, s->length, dir);
- + plat_unmap_dma_mem(dev, s->dma_address, s->length, dir);
- + }
- +}
- +
- +static inline dma_addr_t dma_map_page(struct device *dev, struct page *page,
- + size_t offset, size_t size,
- + enum dma_data_direction dir)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- + dma_addr_t addr;
- +
- + kmemcheck_mark_initialized(page_address(page) + offset, size);
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops) {
- + addr = ops->map_page(dev, page, offset, size, dir, NULL);
- + } else {
- + if (!plat_device_is_coherent(dev))
- + __dma_sync(page, offset, size, dir);
- +
- + addr = plat_map_dma_mem_page(dev, page) + offset;
- + }
- + debug_dma_map_page(dev, page, offset, size, dir, addr, false);
- +
- + return addr;
- +}
- +
- +static inline void dma_unmap_page(struct device *dev, dma_addr_t addr,
- + size_t size, enum dma_data_direction dir)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops) {
- + ops->unmap_page(dev, addr, size, dir, NULL);
- + } else {
- + if (cpu_needs_post_dma_flush(dev))
- + __dma_sync(dma_addr_to_page(dev, addr),
- + addr & ~PAGE_MASK, size, dir);
- + plat_post_dma_flush(dev);
- + plat_unmap_dma_mem(dev, addr, size, dir);
- + }
- + debug_dma_unmap_page(dev, addr, size, dir, false);
- +}
- +
- +static inline void dma_sync_single_for_cpu(struct device *dev, dma_addr_t addr,
- + size_t size,
- + enum dma_data_direction dir)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops) {
- + ops->sync_single_for_cpu(dev, addr, size, dir);
- + } else {
- + if (cpu_needs_post_dma_flush(dev))
- + __dma_sync(dma_addr_to_page(dev, addr),
- + addr & ~PAGE_MASK, size, dir);
- + plat_post_dma_flush(dev);
- + }
- + debug_dma_sync_single_for_cpu(dev, addr, size, dir);
- +}
- +
- +static inline void dma_sync_single_for_device(struct device *dev,
- + dma_addr_t addr, size_t size,
- + enum dma_data_direction dir)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops)
- + ops->sync_single_for_device(dev, addr, size, dir);
- + else if (!plat_device_is_coherent(dev))
- + __dma_sync(dma_addr_to_page(dev, addr),
- + addr & ~PAGE_MASK, size, dir);
- + debug_dma_sync_single_for_device(dev, addr, size, dir);
- +}
- +
- +static inline void dma_sync_single_range_for_cpu(struct device *dev,
- + dma_addr_t addr,
- + unsigned long offset,
- + size_t size,
- + enum dma_data_direction dir)
- +{
- + const struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops) {
- + ops->sync_single_for_cpu(dev, addr + offset, size, dir);
- + } else {
- + if (cpu_needs_post_dma_flush(dev))
- + __dma_sync(dma_addr_to_page(dev, addr + offset),
- + (addr + offset) & ~PAGE_MASK, size, dir);
- + plat_post_dma_flush(dev);
- + }
- +
- + debug_dma_sync_single_range_for_cpu(dev, addr, offset, size, dir);
- +}
- +
- +static inline void dma_sync_single_range_for_device(struct device *dev,
- + dma_addr_t addr,
- + unsigned long offset,
- + size_t size,
- + enum dma_data_direction dir)
- +{
- + const struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops)
- + ops->sync_single_for_device(dev, addr + offset, size, dir);
- + else if (!plat_device_is_coherent(dev))
- + __dma_sync(dma_addr_to_page(dev, addr + offset),
- + (addr + offset) & ~PAGE_MASK, size, dir);
- + debug_dma_sync_single_range_for_device(dev, addr, offset, size, dir);
- +}
- +
- +static inline void
- +dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg,
- + int nelems, enum dma_data_direction dir)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- + struct scatterlist *s;
- + int i;
- +
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops) {
- + ops->sync_sg_for_cpu(dev, sg, nelems, dir);
- + } else if (cpu_needs_post_dma_flush(dev)) {
- + for_each_sg(sg, s, nelems, i)
- + __dma_sync(sg_page(s), s->offset, s->length, dir);
- + }
- + plat_post_dma_flush(dev);
- + debug_dma_sync_sg_for_cpu(dev, sg, nelems, dir);
- +}
- +
- +static inline void
- +dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg,
- + int nelems, enum dma_data_direction dir)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- + struct scatterlist *s;
- + int i;
- +
- + BUG_ON(!valid_dma_direction(dir));
- + if (ops) {
- + ops->sync_sg_for_device(dev, sg, nelems, dir);
- + } else if (!plat_device_is_coherent(dev)) {
- + for_each_sg(sg, s, nelems, i)
- + __dma_sync(sg_page(s), s->offset, s->length, dir);
- + }
- + debug_dma_sync_sg_for_device(dev, sg, nelems, dir);
- +
- +}
- +
- +#define dma_map_single(d, a, s, r) dma_map_single_attrs(d, a, s, r, NULL)
- +#define dma_unmap_single(d, a, s, r) dma_unmap_single_attrs(d, a, s, r, NULL)
- +#define dma_map_sg(d, s, n, r) dma_map_sg_attrs(d, s, n, r, NULL)
- +#define dma_unmap_sg(d, s, n, r) dma_unmap_sg_attrs(d, s, n, r, NULL)
- +
- +extern int dma_common_mmap(struct device *dev, struct vm_area_struct *vma,
- + void *cpu_addr, dma_addr_t dma_addr, size_t size);
- +
- +/**
- + * dma_mmap_attrs - map a coherent DMA allocation into user space
- + * @dev: valid struct device pointer, or NULL for ISA and EISA-like devices
- + * @vma: vm_area_struct describing requested user mapping
- + * @cpu_addr: kernel CPU-view address returned from dma_alloc_attrs
- + * @handle: device-view address returned from dma_alloc_attrs
- + * @size: size of memory originally requested in dma_alloc_attrs
- + * @attrs: attributes of mapping properties requested in dma_alloc_attrs
- + *
- + * Map a coherent DMA buffer previously allocated by dma_alloc_attrs
- + * into user space. The coherent DMA buffer must not be freed by the
- + * driver until the user space mapping has been released.
- + */
- +static inline int
- +dma_mmap_attrs(struct device *dev, struct vm_area_struct *vma, void *cpu_addr,
- + dma_addr_t dma_addr, size_t size, struct dma_attrs *attrs)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- + BUG_ON(!ops);
- + if (ops && ops->mmap)
- + return ops->mmap(dev, vma, cpu_addr, dma_addr, size, attrs);
- + return dma_common_mmap(dev, vma, cpu_addr, dma_addr, size);
- +}
- +
- +#define dma_mmap_coherent(d, v, c, h, s) dma_mmap_attrs(d, v, c, h, s, NULL)
- +
- +int
- +dma_common_get_sgtable(struct device *dev, struct sg_table *sgt,
- + void *cpu_addr, dma_addr_t dma_addr, size_t size);
- +
- +static inline int
- +dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, void *cpu_addr,
- + dma_addr_t dma_addr, size_t size, struct dma_attrs *attrs)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- + BUG_ON(!ops);
- + if (ops && ops->get_sgtable)
- + return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size,
- + attrs);
- + return dma_common_get_sgtable(dev, sgt, cpu_addr, dma_addr, size);
- +}
- +
- +#define dma_get_sgtable(d, t, v, h, s) dma_get_sgtable_attrs(d, t, v, h, s, NULL)
- +
- +static inline int dma_supported(struct device *dev, u64 mask)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- + if (ops)
- + return ops->dma_supported(dev, mask);
- + return plat_dma_supported(dev, mask);
- +}
- +
- +static inline int dma_mapping_error(struct device *dev, u64 mask)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + debug_dma_mapping_error(dev, mask);
- + if (ops)
- + return ops->mapping_error(dev, mask);
- + return 0;
- +}
- +
- +static inline int
- +dma_set_mask(struct device *dev, u64 mask)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + if(!dev->dma_mask || !dma_supported(dev, mask))
- + return -EIO;
- +
- + if (ops && ops->set_dma_mask)
- + return ops->set_dma_mask(dev, mask);
- +
- + *dev->dma_mask = mask;
- +
- + return 0;
- +}
-
- extern void dma_cache_sync(struct device *dev, void *vaddr, size_t size,
- enum dma_data_direction direction);
-
- +#define dma_alloc_coherent(d,s,h,f) dma_alloc_attrs(d,s,h,f,NULL)
- +
- +static inline void *dma_alloc_attrs(struct device *dev, size_t size,
- + dma_addr_t *dma_handle, gfp_t gfp,
- + struct dma_attrs *attrs)
- +{
- + void *ret;
- + struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + if (ops)
- + ret = ops->alloc(dev, size, dma_handle, gfp, attrs);
- + else
- + ret = mips_dma_alloc_coherent(dev, size, dma_handle, gfp,
- + attrs);
- +
- + debug_dma_alloc_coherent(dev, size, *dma_handle, ret);
- +
- + return ret;
- +}
- +
- +#define dma_free_coherent(d,s,c,h) dma_free_attrs(d,s,c,h,NULL)
- +
- +static inline void dma_free_attrs(struct device *dev, size_t size,
- + void *vaddr, dma_addr_t dma_handle,
- + struct dma_attrs *attrs)
- +{
- + struct dma_map_ops *ops = get_dma_ops(dev);
- +
- + if (ops)
- + ops->free(dev, size, vaddr, dma_handle, attrs);
- + else
- + mips_dma_free_coherent(dev, size, vaddr, dma_handle, attrs);
- +
- + debug_dma_free_coherent(dev, size, vaddr, dma_handle);
- +}
- +
- +static inline void *dma_alloc_noncoherent(struct device *dev, size_t size,
- + dma_addr_t *dma_handle, gfp_t gfp)
- +{
- + DEFINE_DMA_ATTRS(attrs);
- +
- + dma_set_attr(DMA_ATTR_NON_CONSISTENT, &attrs);
- + return dma_alloc_attrs(dev, size, dma_handle, gfp, &attrs);
- +}
- +
- +static inline void dma_free_noncoherent(struct device *dev, size_t size,
- + void *cpu_addr, dma_addr_t dma_handle)
- +{
- + DEFINE_DMA_ATTRS(attrs);
- +
- + dma_set_attr(DMA_ATTR_NON_CONSISTENT, &attrs);
- + dma_free_attrs(dev, size, cpu_addr, dma_handle, &attrs);
- +}
- +
- +
- #endif /* _ASM_DMA_MAPPING_H */
- --- a/arch/mips/mm/dma-default.c
- +++ b/arch/mips/mm/dma-default.c
- @@ -46,35 +46,6 @@ static int __init setnocoherentio(char *
- early_param("nocoherentio", setnocoherentio);
- #endif
-
- -static inline struct page *dma_addr_to_page(struct device *dev,
- - dma_addr_t dma_addr)
- -{
- - return pfn_to_page(
- - plat_dma_addr_to_phys(dev, dma_addr) >> PAGE_SHIFT);
- -}
- -
- -/*
- - * The affected CPUs below in 'cpu_needs_post_dma_flush()' can
- - * speculatively fill random cachelines with stale data at any time,
- - * requiring an extra flush post-DMA.
- - *
- - * Warning on the terminology - Linux calls an uncached area coherent;
- - * MIPS terminology calls memory areas with hardware maintained coherency
- - * coherent.
- - *
- - * Note that the R14000 and R16000 should also be checked for in this
- - * condition. However this function is only called on non-I/O-coherent
- - * systems and only the R10000 and R12000 are used in such systems, the
- - * SGI IP28 Indigo² rsp. SGI IP32 aka O2.
- - */
- -static inline int cpu_needs_post_dma_flush(struct device *dev)
- -{
- - return !plat_device_is_coherent(dev) &&
- - (boot_cpu_type() == CPU_R10000 ||
- - boot_cpu_type() == CPU_R12000 ||
- - boot_cpu_type() == CPU_BMIPS5000);
- -}
- -
- static gfp_t massage_gfp_flags(const struct device *dev, gfp_t gfp)
- {
- gfp_t dma_flag;
- @@ -129,7 +100,7 @@ static void *mips_dma_alloc_noncoherent(
- return ret;
- }
-
- -static void *mips_dma_alloc_coherent(struct device *dev, size_t size,
- +void *mips_dma_alloc_coherent(struct device *dev, size_t size,
- dma_addr_t * dma_handle, gfp_t gfp, struct dma_attrs *attrs)
- {
- void *ret;
- @@ -165,6 +136,7 @@ static void *mips_dma_alloc_coherent(str
-
- return ret;
- }
- +EXPORT_SYMBOL(mips_dma_alloc_coherent);
-
-
- static void mips_dma_free_noncoherent(struct device *dev, size_t size,
- @@ -174,7 +146,7 @@ static void mips_dma_free_noncoherent(st
- free_pages((unsigned long) vaddr, get_order(size));
- }
-
- -static void mips_dma_free_coherent(struct device *dev, size_t size, void *vaddr,
- +void mips_dma_free_coherent(struct device *dev, size_t size, void *vaddr,
- dma_addr_t dma_handle, struct dma_attrs *attrs)
- {
- unsigned long addr = (unsigned long) vaddr;
- @@ -196,40 +168,7 @@ static void mips_dma_free_coherent(struc
- if (!dma_release_from_contiguous(dev, page, count))
- __free_pages(page, get_order(size));
- }
- -
- -static int mips_dma_mmap(struct device *dev, struct vm_area_struct *vma,
- - void *cpu_addr, dma_addr_t dma_addr, size_t size,
- - struct dma_attrs *attrs)
- -{
- - unsigned long user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT;
- - unsigned long count = PAGE_ALIGN(size) >> PAGE_SHIFT;
- - unsigned long addr = (unsigned long)cpu_addr;
- - unsigned long off = vma->vm_pgoff;
- - unsigned long pfn;
- - int ret = -ENXIO;
- -
- - if (!plat_device_is_coherent(dev) && !hw_coherentio)
- - addr = CAC_ADDR(addr);
- -
- - pfn = page_to_pfn(virt_to_page((void *)addr));
- -
- - if (dma_get_attr(DMA_ATTR_WRITE_COMBINE, attrs))
- - vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot);
- - else
- - vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
- -
- - if (dma_mmap_from_coherent(dev, vma, cpu_addr, size, &ret))
- - return ret;
- -
- - if (off < count && user_count <= (count - off)) {
- - ret = remap_pfn_range(vma, vma->vm_start,
- - pfn + off,
- - user_count << PAGE_SHIFT,
- - vma->vm_page_prot);
- - }
- -
- - return ret;
- -}
- +EXPORT_SYMBOL(mips_dma_free_coherent);
-
- static inline void __dma_sync_virtual(void *addr, size_t size,
- enum dma_data_direction direction)
- @@ -258,7 +197,7 @@ static inline void __dma_sync_virtual(vo
- * If highmem is not configured then the bulk of this loop gets
- * optimized out.
- */
- -static inline void __dma_sync(struct page *page,
- +void __dma_sync(struct page *page,
- unsigned long offset, size_t size, enum dma_data_direction direction)
- {
- size_t left = size;
- @@ -288,120 +227,7 @@ static inline void __dma_sync(struct pag
- left -= len;
- } while (left);
- }
- -
- -static void mips_dma_unmap_page(struct device *dev, dma_addr_t dma_addr,
- - size_t size, enum dma_data_direction direction, struct dma_attrs *attrs)
- -{
- - if (cpu_needs_post_dma_flush(dev))
- - __dma_sync(dma_addr_to_page(dev, dma_addr),
- - dma_addr & ~PAGE_MASK, size, direction);
- - plat_post_dma_flush(dev);
- - plat_unmap_dma_mem(dev, dma_addr, size, direction);
- -}
- -
- -static int mips_dma_map_sg(struct device *dev, struct scatterlist *sglist,
- - int nents, enum dma_data_direction direction, struct dma_attrs *attrs)
- -{
- - int i;
- - struct scatterlist *sg;
- -
- - for_each_sg(sglist, sg, nents, i) {
- - if (!plat_device_is_coherent(dev))
- - __dma_sync(sg_page(sg), sg->offset, sg->length,
- - direction);
- -#ifdef CONFIG_NEED_SG_DMA_LENGTH
- - sg->dma_length = sg->length;
- -#endif
- - sg->dma_address = plat_map_dma_mem_page(dev, sg_page(sg)) +
- - sg->offset;
- - }
- -
- - return nents;
- -}
- -
- -static dma_addr_t mips_dma_map_page(struct device *dev, struct page *page,
- - unsigned long offset, size_t size, enum dma_data_direction direction,
- - struct dma_attrs *attrs)
- -{
- - if (!plat_device_is_coherent(dev))
- - __dma_sync(page, offset, size, direction);
- -
- - return plat_map_dma_mem_page(dev, page) + offset;
- -}
- -
- -static void mips_dma_unmap_sg(struct device *dev, struct scatterlist *sglist,
- - int nhwentries, enum dma_data_direction direction,
- - struct dma_attrs *attrs)
- -{
- - int i;
- - struct scatterlist *sg;
- -
- - for_each_sg(sglist, sg, nhwentries, i) {
- - if (!plat_device_is_coherent(dev) &&
- - direction != DMA_TO_DEVICE)
- - __dma_sync(sg_page(sg), sg->offset, sg->length,
- - direction);
- - plat_unmap_dma_mem(dev, sg->dma_address, sg->length, direction);
- - }
- -}
- -
- -static void mips_dma_sync_single_for_cpu(struct device *dev,
- - dma_addr_t dma_handle, size_t size, enum dma_data_direction direction)
- -{
- - if (cpu_needs_post_dma_flush(dev))
- - __dma_sync(dma_addr_to_page(dev, dma_handle),
- - dma_handle & ~PAGE_MASK, size, direction);
- - plat_post_dma_flush(dev);
- -}
- -
- -static void mips_dma_sync_single_for_device(struct device *dev,
- - dma_addr_t dma_handle, size_t size, enum dma_data_direction direction)
- -{
- - if (!plat_device_is_coherent(dev))
- - __dma_sync(dma_addr_to_page(dev, dma_handle),
- - dma_handle & ~PAGE_MASK, size, direction);
- -}
- -
- -static void mips_dma_sync_sg_for_cpu(struct device *dev,
- - struct scatterlist *sglist, int nelems,
- - enum dma_data_direction direction)
- -{
- - int i;
- - struct scatterlist *sg;
- -
- - if (cpu_needs_post_dma_flush(dev)) {
- - for_each_sg(sglist, sg, nelems, i) {
- - __dma_sync(sg_page(sg), sg->offset, sg->length,
- - direction);
- - }
- - }
- - plat_post_dma_flush(dev);
- -}
- -
- -static void mips_dma_sync_sg_for_device(struct device *dev,
- - struct scatterlist *sglist, int nelems,
- - enum dma_data_direction direction)
- -{
- - int i;
- - struct scatterlist *sg;
- -
- - if (!plat_device_is_coherent(dev)) {
- - for_each_sg(sglist, sg, nelems, i) {
- - __dma_sync(sg_page(sg), sg->offset, sg->length,
- - direction);
- - }
- - }
- -}
- -
- -int mips_dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
- -{
- - return 0;
- -}
- -
- -int mips_dma_supported(struct device *dev, u64 mask)
- -{
- - return plat_dma_supported(dev, mask);
- -}
- +EXPORT_SYMBOL(__dma_sync);
-
- void dma_cache_sync(struct device *dev, void *vaddr, size_t size,
- enum dma_data_direction direction)
- @@ -414,24 +240,10 @@ void dma_cache_sync(struct device *dev,
-
- EXPORT_SYMBOL(dma_cache_sync);
-
- -static struct dma_map_ops mips_default_dma_map_ops = {
- - .alloc = mips_dma_alloc_coherent,
- - .free = mips_dma_free_coherent,
- - .mmap = mips_dma_mmap,
- - .map_page = mips_dma_map_page,
- - .unmap_page = mips_dma_unmap_page,
- - .map_sg = mips_dma_map_sg,
- - .unmap_sg = mips_dma_unmap_sg,
- - .sync_single_for_cpu = mips_dma_sync_single_for_cpu,
- - .sync_single_for_device = mips_dma_sync_single_for_device,
- - .sync_sg_for_cpu = mips_dma_sync_sg_for_cpu,
- - .sync_sg_for_device = mips_dma_sync_sg_for_device,
- - .mapping_error = mips_dma_mapping_error,
- - .dma_supported = mips_dma_supported
- -};
- -
- -struct dma_map_ops *mips_dma_map_ops = &mips_default_dma_map_ops;
- +#ifdef CONFIG_SYS_HAS_DMA_OPS
- +struct dma_map_ops *mips_dma_map_ops = NULL;
- EXPORT_SYMBOL(mips_dma_map_ops);
- +#endif
-
- #define PREALLOC_DMA_DEBUG_ENTRIES (1 << 16)
-
|