* refs/heads/tmp-f960b38: Linux 4.14.159 of: unittest: fix memory leak in attach_node_and_children raid5: need to set STRIPE_HANDLE for batch head gpiolib: acpi: Add Terra Pad 1061 to the run_edge_events_on_boot_blacklist kernel/module.c: wakeup processes in module_wq on module unload gfs2: fix glock reference problem in gfs2_trans_remove_revoke net/mlx5e: Fix SFF 8472 eeprom length sunrpc: fix crash when cache_head become valid before update workqueue: Fix missing kfree(rescuer) in destroy_workqueue() blk-mq: make sure that line break can be printed mfd: rk808: Fix RK818 ID template ext4: fix a bug in ext4_wait_for_tail_page_commit mm/shmem.c: cast the type of unmap_start to u64 firmware: qcom: scm: Ensure 'a0' status code is treated as signed ext4: work around deleting a file with i_nlink == 0 safely powerpc: Fix vDSO clock_getres() powerpc: Avoid clang warnings around setjmp and longjmp ath10k: fix fw crash by moving chip reset after napi disabled media: vimc: fix component match compare mlxsw: spectrum_router: Refresh nexthop neighbour when it becomes dead power: supply: cpcap-battery: Fix signed counter sample register x86/MCE/AMD: Carve out the MC4_MISC thresholding quirk x86/MCE/AMD: Turn off MC4_MISC thresholding on all family 0x15 models e100: Fix passing zero to 'PTR_ERR' warning in e100_load_ucode_wait drbd: Change drbd_request_detach_interruptible's return type to int scsi: lpfc: Correct code setting non existent bits in sli4 ABORT WQE scsi: lpfc: Cap NPIV vports to 256 omap: pdata-quirks: remove openpandora quirks for mmc3 and wl1251 phy: renesas: rcar-gen3-usb2: Fix sysfs interface of "role" iio: adis16480: Add debugfs_reg_access entry xhci: make sure interrupts are restored to correct state xhci: Fix memory leak in xhci_add_in_port() scsi: qla2xxx: Fix message indicating vectors used by driver scsi: qla2xxx: Always check the qla2x00_wait_for_hba_online() return value scsi: qla2xxx: Fix qla24xx_process_bidir_cmd() scsi: qla2xxx: Fix session lookup in qlt_abort_work() scsi: qla2xxx: Fix DMA unmap leak scsi: zfcp: trace channel log even for FCP command responses block: fix single range discard merge reiserfs: fix extended attributes on the root directory ext4: Fix credit estimate for final inode freeing quota: fix livelock in dquot_writeback_dquots ext2: check err when partial != NULL quota: Check that quota is not dirty before release video/hdmi: Fix AVI bar unpack powerpc/xive: Skip ioremap() of ESB pages for LSI interrupts powerpc: Allow flush_icache_range to work across ranges >4GB powerpc/xive: Prevent page fault issues in the machine crash handler powerpc: Allow 64bit VDSO __kernel_sync_dicache to work across ranges >4GB ppdev: fix PPGETTIME/PPSETTIME ioctls ARM: dts: omap3-tao3530: Fix incorrect MMC card detection GPIO polarity mmc: host: omap_hsmmc: add code for special init of wl1251 to get rid of pandora_wl1251_init_card pinctrl: samsung: Fix device node refcount leaks in S3C64xx wakeup controller init pinctrl: samsung: Fix device node refcount leaks in init code pinctrl: samsung: Fix device node refcount leaks in S3C24xx wakeup controller init pinctrl: samsung: Add of_node_put() before return in error path ACPI: PM: Avoid attaching ACPI PM domain to certain devices ACPI: bus: Fix NULL pointer check in acpi_bus_get_private_data() ACPI: OSL: only free map once in osl.c cpufreq: powernv: fix stack bloat and hard limit on number of CPUs PM / devfreq: Lock devfreq in trans_stat_show intel_th: pci: Add Tiger Lake CPU support intel_th: pci: Add Ice Lake CPU support intel_th: Fix a double put_device() in error path cpuidle: Do not unset the driver if it is there already media: cec.h: CEC_OP_REC_FLAG_ values were swapped media: radio: wl1273: fix interrupt masking on release media: bdisp: fix memleak on release s390/mm: properly clear _PAGE_NOEXEC bit when it is not supported ar5523: check NULL before memcpy() in ar5523_cmd() cgroup: pids: use atomic64_t for pids->limit blk-mq: avoid sysfs buffer overflow with too many CPU cores ASoC: Jack: Fix NULL pointer dereference in snd_soc_jack_report workqueue: Fix pwq ref leak in rescuer_thread() workqueue: Fix spurious sanity check failures in destroy_workqueue() dm zoned: reduce overhead of backing device checks hwrng: omap - Fix RNG wait loop timeout watchdog: aspeed: Fix clock behaviour for ast2600 md/raid0: Fix an error message in raid0_make_request() ALSA: hda - Fix pending unsol events at shutdown ovl: relax WARN_ON() on rename to self lib: raid6: fix awk build warnings rtlwifi: rtl8192de: Fix missing enable interrupt flag rtlwifi: rtl8192de: Fix missing callback that tests for hw release of buffer rtlwifi: rtl8192de: Fix missing code to retrieve RX buffer address btrfs: record all roots for rename exchange on a subvol Btrfs: send, skip backreference walking for extents with many references btrfs: Remove btrfs_bio::flags member Btrfs: fix negative subv_writers counter and data space leak after buffered write btrfs: use refcount_inc_not_zero in kill_all_nodes btrfs: check page->mapping when loading free space cache usb: dwc3: ep0: Clear started flag on completion virtio-balloon: fix managed page counts when migrating pages between zones mtd: spear_smi: Fix Write Burst mode tpm: add check after commands attribs tab allocation usb: mon: Fix a deadlock in usbmon between mmap and read usb: core: urb: fix URB structure initialization function USB: adutux: fix interface sanity check USB: serial: io_edgeport: fix epic endpoint lookup USB: idmouse: fix interface sanity checks USB: atm: ueagle-atm: add missing endpoint check iio: humidity: hdc100x: fix IIO_HUMIDITYRELATIVE channel reporting ARM: dts: pandora-common: define wl1251 as child node of mmc3 xhci: handle some XHCI_TRUST_TX_LENGTH quirks cases as default behaviour. xhci: Increase STS_HALT timeout in xhci_suspend() usb: xhci: only set D3hot for pci device staging: gigaset: add endpoint-type sanity check staging: gigaset: fix illegal free on probe errors staging: gigaset: fix general protection fault on probe staging: rtl8712: fix interface sanity check staging: rtl8188eu: fix interface sanity check usb: Allow USB device to be warm reset in suspended state USB: documentation: flags on usb-storage versus UAS USB: uas: heed CAPACITY_HEURISTICS USB: uas: honor flag to avoid CAPACITY16 media: venus: remove invalid compat_ioctl32 handler scsi: qla2xxx: Fix driver unload hang usb: gadget: pch_udc: fix use after free usb: gadget: configfs: Fix missing spin_lock_init() appletalk: Set error code if register_snap_client failed appletalk: Fix potential NULL pointer dereference in unregister_snap_client KVM: x86: fix out-of-bounds write in KVM_GET_EMULATED_CPUID (CVE-2019-19332) ASoC: rsnd: fixup MIX kctrl registration binder: Handle start==NULL in binder_update_page_range() thermal: Fix deadlock in thermal thermal_zone_device_check iomap: Fix pipe page leakage during splicing RDMA/qib: Validate ->show()/store() callbacks before calling them spi: atmel: Fix CS high support crypto: user - fix memory leak in crypto_report crypto: ecdh - fix big endian bug in ECC library crypto: ccp - fix uninitialized list head crypto: af_alg - cast ki_complete ternary op to int crypto: crypto4xx - fix double-free in crypto4xx_destroy_sdr KVM: x86: fix presentation of TSX feature in ARCH_CAPABILITIES KVM: x86: do not modify masked bits of shared MSRs KVM: arm/arm64: vgic: Don't rely on the wrong pending table drm/i810: Prevent underflow in ioctl jbd2: Fix possible overflow in jbd2_log_space_left() kernfs: fix ino wrap-around detection can: slcan: Fix use-after-free Read in slcan_open tty: vt: keyboard: reject invalid keycodes CIFS: Fix SMB2 oplock break processing CIFS: Fix NULL-pointer dereference in smb2_push_mandatory_locks x86/PCI: Avoid AMD FCH XHCI USB PME# from D0 defect Input: Fix memory leak in psxpad_spi_probe coresight: etm4x: Fix input validation for sysfs. Input: goodix - add upside-down quirk for Teclast X89 tablet Input: synaptics-rmi4 - don't increment rmiaddr for SMBus transfers Input: synaptics-rmi4 - re-enable IRQs in f34v7_do_reflash Input: synaptics - switch another X1 Carbon 6 to RMI/SMbus ALSA: hda - Add mute led support for HP ProBook 645 G4 ALSA: pcm: oss: Avoid potential buffer overflows ALSA: hda/realtek - Dell headphone has noise on unmute for ALC236 fuse: verify attributes fuse: verify nlink sched/fair: Scale bandwidth quota and period without losing quota/period ratio precision tcp: exit if nothing to retransmit on RTO timeout net: aquantia: fix RSS table and key sizes media: vimc: fix start stream when link is disabled ARM: dts: sunxi: Fix PMU compatible strings usb: mtu3: fix dbginfo in qmu_tx_zlp_error_handler mlx4: Use snprintf instead of complicated strcpy IB/hfi1: Close VNIC sdma_progress sleep window IB/hfi1: Ignore LNI errors before DC8051 transitions to Polling state mlxsw: spectrum_router: Relax GRE decap matching check firmware: qcom: scm: fix compilation error when disabled media: stkwebcam: Bugfix for wrong return values tty: Don't block on IO when ldisc change is pending nfsd: Return EPERM, not EACCES, in some SETATTR cases MIPS: OCTEON: cvmx_pko_mem_debug8: use oldest forward compatible definition clk: renesas: r8a77995: Correct parent clock of DU powerpc/math-emu: Update macros from GCC pstore/ram: Avoid NULL deref in ftrace merging failure path net/mlx4_core: Fix return codes of unsupported operations dlm: fix invalid cluster name warning ARM: dts: realview: Fix some more duplicate regulator nodes clk: sunxi-ng: h3/h5: Fix CSI_MCLK parent ARM: dts: pxa: clean up USB controller nodes mtd: fix mtd_oobavail() incoherent returned value kbuild: fix single target build for external module modpost: skip ELF local symbols during section mismatch check tcp: fix SNMP TCP timeout under-estimation tcp: fix SNMP under-estimation on failed retransmission tcp: fix off-by-one bug on aborting window-probing socket ARM: dts: realview-pbx: Fix duplicate regulator nodes ARM: dts: mmp2: fix the gpio interrupt cell number net/x25: fix null_x25_address handling net/x25: fix called/calling length calculation in x25_parse_address_block arm64: dts: meson-gxl-khadas-vim: fix GPIO lines names arm64: dts: meson-gxbb-odroidc2: fix GPIO lines names arm64: dts: meson-gxbb-nanopi-k2: fix GPIO lines names arm64: dts: meson-gxl-libretech-cc: fix GPIO lines names ARM: OMAP1/2: fix SoC name printing ASoC: au8540: use 64-bit arithmetic instead of 32-bit nfsd: fix a warning in __cld_pipe_upcall() ARM: debug: enable UART1 for socfpga Cyclone5 dlm: NULL check before kmem_cache_destroy is not needed ARM: dts: sun8i: v3s: Change pinctrl nodes to avoid warning ARM: dts: sun5i: a10s: Fix HDMI output DTC warning ASoC: rsnd: tidyup registering method for rsnd_kctrl_new() lockd: fix decoding of TEST results i2c: imx: don't print error message on probe defer serial: imx: fix error handling in console_setup altera-stapl: check for a null key before strcasecmp'ing it dma-mapping: fix return type of dma_set_max_seg_size() sparc: Correct ctx->saw_frame_pointer logic. f2fs: fix to allow node segment for GC by ioctl path ARM: dts: rockchip: Assign the proper GPIO clocks for rv1108 ARM: dts: rockchip: Fix the PMU interrupt number for rv1108 f2fs: change segment to section in f2fs_ioc_gc_range f2fs: fix count of seg_freed to make sec_freed correct ACPI: fix acpi_find_child_device() invocation in acpi_preset_companion() usb: dwc3: don't log probe deferrals; but do log other error codes usb: dwc3: debugfs: Properly print/set link state for HS dmaengine: dw-dmac: implement dma protection control setting dmaengine: coh901318: Remove unused variable dmaengine: coh901318: Fix a double-lock bug media: cec: report Vendor ID after initialization media: pulse8-cec: return 0 when invalidating the logical address ARM: dts: exynos: Use Samsung SoC specific compatible for DWC2 module rtc: dt-binding: abx80x: fix resistance scale rtc: max8997: Fix the returned value in case of error in 'max8997_rtc_read_alarm()' math-emu/soft-fp.h: (_FP_ROUND_ZERO) cast 0 to void to fix warning net/smc: use after free fix in smc_wr_tx_put_slot() MIPS: OCTEON: octeon-platform: fix typing iomap: sub-block dio needs to zeroout beyond EOF net-next/hinic:fix a bug in set mac address regulator: Fix return value of _set_load() stub clk: rockchip: fix ID of 8ch clock of I2S1 for rk3328 clk: rockchip: fix I2S1 clock gate register for rk3328 mm/vmstat.c: fix NUMA statistics updates Staging: iio: adt7316: Fix i2c data reading, set the data field pinctrl: qcom: ssbi-gpio: fix gpio-hog related boot issues crypto: bcm - fix normal/non key hash algorithm failure crypto: ecc - check for invalid values in the key verification test scsi: zfcp: drop default switch case which might paper over missing case net: dsa: mv88e6xxx: Work around mv886e6161 SERDES missing MII_PHYSID2 MIPS: SiByte: Enable ZONE_DMA32 for LittleSur dlm: fix missing idr_destroy for recover_idr ARM: dts: rockchip: Fix rk3288-rock2 vcc_flash name clk: rockchip: fix rk3188 sclk_mac_lbtest parameter ordering clk: rockchip: fix rk3188 sclk_smc gate data i40e: don't restart nway if autoneg not supported rtc: s3c-rtc: Avoid using broken ALMYEAR register net: ethernet: ti: cpts: correct debug for expired txq skb extcon: max8997: Fix lack of path setting in USB device mode dlm: fix possible call to kfree() for non-initialized pointer clk: sunxi-ng: a64: Fix gate bit of DSI DPHY net/mlx5: Release resource on error flow ARM: 8813/1: Make aligned 2-byte getuser()/putuser() atomic on ARMv6+ iwlwifi: mvm: Send non offchannel traffic via AP sta iwlwifi: mvm: synchronize TID queue removal cxgb4vf: fix memleak in mac_hlist initialization serial: core: Allow processing sysrq at port unlock time i2c: core: fix use after free in of_i2c_notify net: ep93xx_eth: fix mismatch of request_mem_region in remove rsxx: add missed destroy_workqueue calls in remove ALSA: pcm: Fix stream lock usage in snd_pcm_period_elapsed() sched/core: Avoid spurious lock dependencies Input: cyttsp4_core - fix use after free bug xfrm: release device reference for invalid state NFC: nxp-nci: Fix NULL pointer dereference after I2C communication error audit_get_nd(): don't unlock parent too early exportfs_decode_fh(): negative pinned may become positive without the parent locked iwlwifi: pcie: don't consider IV len in A-MSDU RDMA/hns: Correct the value of HNS_ROCE_HEM_CHUNK_LEN autofs: fix a leak in autofs_expire_indirect() serial: ifx6x60: add missed pm_runtime_disable serial: serial_core: Perform NULL checks for break_ctl ops serial: pl011: Fix DMA ->flush_buffer() tty: serial: msm_serial: Fix flow control tty: serial: fsl_lpuart: use the sg count from dma_map_sg usb: gadget: u_serial: add missing port entry locking arm64: tegra: Fix 'active-low' warning for Jetson TX1 regulator rsi: release skb if rsi_prepare_beacon fails ANDROID: staging: android: ion: Fix build when CONFIG_ION_SYSTEM_HEAP=n ANDROID: staging: android: ion: Expose total heap and pool sizes via sysfs UPSTREAM: include/linux/slab.h: fix sparse warning in kmalloc_type() UPSTREAM: mm, slab: shorten kmalloc cache names for large sizes UPSTREAM: mm, proc: add KReclaimable to /proc/meminfo BACKPORT: mm: rename and change semantics of nr_indirectly_reclaimable_bytes UPSTREAM: dcache: allocate external names from reclaimable kmalloc caches BACKPORT: mm, slab/slub: introduce kmalloc-reclaimable caches UPSTREAM: mm, slab: combine kmalloc_caches and kmalloc_dma_caches ANDROID: kbuild: disable SCS by default in allmodconfig ANDROID: arm64: cuttlefish_defconfig: enable LTO, CFI, and SCS BACKPORT: FROMLIST: arm64: implement Shadow Call Stack FROMLIST: arm64: disable SCS for hypervisor code BACKPORT: FROMLIST: arm64: vdso: disable Shadow Call Stack FROMLIST: arm64: preserve x18 when CPU is suspended FROMLIST: arm64: reserve x18 from general allocation with SCS FROMLIST: arm64: disable function graph tracing with SCS FROMLIST: scs: add support for stack usage debugging FROMLIST: scs: add accounting FROMLIST: add support for Clang's Shadow Call Stack (SCS) FROMLIST: arm64: kernel: avoid x18 in __cpu_soft_restart FROMLIST: arm64: kvm: stop treating register x18 as caller save FROMLIST: arm64/lib: copy_page: avoid x18 register in assembler code FROMLIST: arm64: mm: avoid x18 in idmap_kpti_install_ng_mappings ANDROID: use non-canonical CFI jump tables ANDROID: arm64: add __nocfi to __apply_alternatives ANDROID: arm64: add __pa_function ANDROID: arm64: allow ThinLTO to be selected ANDROID: soc/tegra: disable ARCH_TEGRA_210_SOC with LTO FROMLIST: arm64: fix alternatives with LLVM's integrated assembler ANDROID: irqchip/gic-v3: rename gic_of_init to work around a ThinLTO+CFI bug ANDROID: kbuild: limit LTO inlining ANDROID: kbuild: merge module sections with LTO ANDROID: init: ensure initcall ordering with LTO Revert "ANDROID: HACK: init: ensure initcall ordering with LTO" ANDROID: add support for ThinLTO ANDROID: Switch to LLD ANDROID: clang: update to 10.0.1 ANDROID: arm64: add atomic_ll_sc.o to obj-y if using lld ANDROID: enable ARM64_ERRATUM_843419 by default with LTO_CLANG ANDROID: kbuild: allow lld to be used with CONFIG_LTO_CLANG ANDROID: Makefile: set -Qunused-arguments sooner BACKPORT: FROMLIST: Makefile: lld: tell clang to use lld BACKPORT: FROMLIST: Makefile: lld: set -O2 linker flag when linking with LLD ANDROID: scripts/Kbuild: add ld-name support for ld.lld UPSTREAM: bpf: permit multiple bpf attachments for a single perf event UPSTREAM: bpf: use the same condition in perf event set/free bpf handler UPSTREAM: bpf: multi program support for cgroup+bpf BACKPORT: serdev: make synchronous write return bytes written UPSTREAM: gnss: serial: fix synchronous write timeout UPSTREAM: gnss: fix potential error pointer dereference BACKPORT: gnss: add receiver type support UPSTREAM: dt-bindings: add generic gnss binding UPSTREAM: gnss: add generic serial driver ANDROID: cuttlefish_defconfig: Enable CONFIG_SERIAL_DEV_BUS ANDROID: cuttlefish_defconfig: Enable CONFIG_GNSS BACKPORT: gnss: add GNSS receiver subsystem UPSTREAM: arm64: Validate tagged addresses in access_ok() called from kernel threads BACKPORT: ARM: 8905/1: Emit __gnu_mcount_nc when using Clang 10.0.0 or newer fs/lock: skip lock owner pid translation in case we are in init_pid_ns f2fs: stop GC when the victim becomes fully valid f2fs: expose main_blkaddr in sysfs f2fs: choose hardlimit when softlimit is larger than hardlimit in f2fs_statfs_project() f2fs: Fix deadlock in f2fs_gc() context during atomic files handling f2fs: show f2fs instance in printk_ratelimited f2fs: fix potential overflow f2fs: fix to update dir's i_pino during cross_rename f2fs: support aligned pinned file f2fs: avoid kernel panic on corruption test f2fs: fix wrong description in document f2fs: cache global IPU bio f2fs: fix to avoid memory leakage in f2fs_listxattr f2fs: check total_segments from devices in raw_super f2fs: update multi-dev metadata in resize_fs f2fs: mark recovery flag correctly in read_raw_super_block() f2fs: fix to update time in lazytime mode vfs: don't allow writes to swap files mm: set S_SWAPFILE on blockdev swap devices Conflicts: drivers/Makefile drivers/staging/android/ion/ion.c drivers/staging/android/ion/ion.h drivers/staging/android/ion/ion_page_pool.c drivers/usb/dwc3/core.c drivers/usb/dwc3/debugfs.c drivers/usb/dwc3/ep0.c fs/f2fs/data.c include/linux/mmzone.h mm/vmstat.c Discarded below patches, as usb patches not applicable and block patch causing stability issues: usb: dwc3: ep0: Clear started flag on completion usb: dwc3: don't log probe deferrals; but do log other error codes block: fix single range discard merge Fixed build errors in below files: drivers/gpu/msm/kgsl_pool.c drivers/staging/android/ion/ion_page_pool.c kernel/taskstats.c Fixed bootup issue in: arch/arm64/mm/proc.s Change-Id: I0a16824c251c14c63af78f9cfd9ede5e82c427fc Signed-off-by: Srinivasarao P <spathi@codeaurora.org> Signed-off-by: Blagovest Kolenichev <bkolenichev@codeaurora.org>
952 lines
28 KiB
C
952 lines
28 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _LINUX_DMA_MAPPING_H
|
|
#define _LINUX_DMA_MAPPING_H
|
|
|
|
#include <linux/sizes.h>
|
|
#include <linux/string.h>
|
|
#include <linux/device.h>
|
|
#include <linux/err.h>
|
|
#include <linux/dma-debug.h>
|
|
#include <linux/dma-direction.h>
|
|
#include <linux/scatterlist.h>
|
|
#include <linux/bug.h>
|
|
#include <linux/mem_encrypt.h>
|
|
|
|
/**
|
|
* List of possible attributes associated with a DMA mapping. The semantics
|
|
* of each attribute should be defined in Documentation/DMA-attributes.txt.
|
|
*
|
|
* DMA_ATTR_WRITE_BARRIER: DMA to a memory region with this attribute
|
|
* forces all pending DMA writes to complete.
|
|
*/
|
|
#define DMA_ATTR_WRITE_BARRIER (1UL << 0)
|
|
/*
|
|
* DMA_ATTR_WEAK_ORDERING: Specifies that reads and writes to the mapping
|
|
* may be weakly ordered, that is that reads and writes may pass each other.
|
|
*/
|
|
#define DMA_ATTR_WEAK_ORDERING (1UL << 1)
|
|
/*
|
|
* DMA_ATTR_WRITE_COMBINE: Specifies that writes to the mapping may be
|
|
* buffered to improve performance.
|
|
*/
|
|
#define DMA_ATTR_WRITE_COMBINE (1UL << 2)
|
|
/*
|
|
* DMA_ATTR_NON_CONSISTENT: Lets the platform to choose to return either
|
|
* consistent or non-consistent memory as it sees fit.
|
|
*/
|
|
#define DMA_ATTR_NON_CONSISTENT (1UL << 3)
|
|
/*
|
|
* DMA_ATTR_NO_KERNEL_MAPPING: Lets the platform to avoid creating a kernel
|
|
* virtual mapping for the allocated buffer.
|
|
*/
|
|
#define DMA_ATTR_NO_KERNEL_MAPPING (1UL << 4)
|
|
/*
|
|
* DMA_ATTR_SKIP_CPU_SYNC: Allows platform code to skip synchronization of
|
|
* the CPU cache for the given buffer assuming that it has been already
|
|
* transferred to 'device' domain.
|
|
*/
|
|
#define DMA_ATTR_SKIP_CPU_SYNC (1UL << 5)
|
|
/*
|
|
* DMA_ATTR_FORCE_CONTIGUOUS: Forces contiguous allocation of the buffer
|
|
* in physical memory.
|
|
*/
|
|
#define DMA_ATTR_FORCE_CONTIGUOUS (1UL << 6)
|
|
/*
|
|
* DMA_ATTR_ALLOC_SINGLE_PAGES: This is a hint to the DMA-mapping subsystem
|
|
* that it's probably not worth the time to try to allocate memory to in a way
|
|
* that gives better TLB efficiency.
|
|
*/
|
|
#define DMA_ATTR_ALLOC_SINGLE_PAGES (1UL << 7)
|
|
/*
|
|
* DMA_ATTR_NO_WARN: This tells the DMA-mapping subsystem to suppress
|
|
* allocation failure reports (similarly to __GFP_NOWARN).
|
|
*/
|
|
#define DMA_ATTR_NO_WARN (1UL << 8)
|
|
/*
|
|
* DMA_ATTR_STRONGLY_ORDERED: Specifies that accesses to the mapping must
|
|
* not be buffered, reordered, merged with other accesses, or unaligned.
|
|
* No speculative access may occur in this mapping.
|
|
*/
|
|
#define DMA_ATTR_STRONGLY_ORDERED (1UL << 9)
|
|
/*
|
|
* DMA_ATTR_SKIP_ZEROING: Do not zero mapping.
|
|
*/
|
|
#define DMA_ATTR_SKIP_ZEROING (1UL << 10)
|
|
/*
|
|
* DMA_ATTR_NO_DELAYED_UNMAP: Used by msm specific lazy mapping to indicate
|
|
* that the mapping can be freed on unmap, rather than when the ion_buffer
|
|
* is freed.
|
|
*/
|
|
#define DMA_ATTR_NO_DELAYED_UNMAP (1UL << 11)
|
|
/*
|
|
* DMA_ATTR_EXEC_MAPPING: The mapping has executable permissions.
|
|
*/
|
|
#define DMA_ATTR_EXEC_MAPPING (1UL << 12)
|
|
/*
|
|
* DMA_ATTR_IOMMU_USE_UPSTREAM_HINT: Normally an smmu will override any bus
|
|
* attributes (i.e cacheablilty) provided by the client device. Some hardware
|
|
* may be designed to use the original attributes instead.
|
|
*/
|
|
#define DMA_ATTR_IOMMU_USE_UPSTREAM_HINT (1UL << 13)
|
|
/*
|
|
* When passed to a DMA map call the DMA_ATTR_FORCE_COHERENT DMA
|
|
* attribute can be used to force a buffer to be mapped as IO coherent.
|
|
*/
|
|
#define DMA_ATTR_FORCE_COHERENT (1UL << 14)
|
|
/*
|
|
* When passed to a DMA map call the DMA_ATTR_FORCE_NON_COHERENT DMA
|
|
* attribute can be used to force a buffer to not be mapped as IO
|
|
* coherent.
|
|
*/
|
|
#define DMA_ATTR_FORCE_NON_COHERENT (1UL << 15)
|
|
/*
|
|
* DMA_ATTR_DELAYED_UNMAP: Used by ION, it will ensure that mappings are not
|
|
* removed on unmap but instead are removed when the ion_buffer is freed.
|
|
*/
|
|
#define DMA_ATTR_DELAYED_UNMAP (1UL << 16)
|
|
|
|
/*
|
|
* DMA_ATTR_PRIVILEGED: used to indicate that the buffer is fully
|
|
* accessible at an elevated privilege level (and ideally inaccessible or
|
|
* at least read-only at lesser-privileged levels).
|
|
*/
|
|
#define DMA_ATTR_PRIVILEGED (1UL << 17)
|
|
|
|
/*
|
|
* DMA_ATTR_IOMMU_USE_LLC_NWA: Overrides the bus attributes to use the System
|
|
* Cache(LLC) with allocation policy as Inner Non-Cacheable, Outer Cacheable:
|
|
* Write-Back, Read-Allocate, No Write-Allocate policy.
|
|
*/
|
|
#define DMA_ATTR_IOMMU_USE_LLC_NWA (1UL << 18)
|
|
|
|
#define DMA_ERROR_CODE (~(dma_addr_t)0)
|
|
|
|
/*
|
|
* A dma_addr_t can hold any valid DMA or bus address for the platform.
|
|
* It can be given to a device to use as a DMA source or target. A CPU cannot
|
|
* reference a dma_addr_t directly because there may be translation between
|
|
* its physical address space and the bus address space.
|
|
*/
|
|
struct dma_map_ops {
|
|
void* (*alloc)(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, gfp_t gfp,
|
|
unsigned long attrs);
|
|
void (*free)(struct device *dev, size_t size,
|
|
void *vaddr, dma_addr_t dma_handle,
|
|
unsigned long attrs);
|
|
int (*mmap)(struct device *, struct vm_area_struct *,
|
|
void *, dma_addr_t, size_t,
|
|
unsigned long attrs);
|
|
|
|
int (*get_sgtable)(struct device *dev, struct sg_table *sgt, void *,
|
|
dma_addr_t, size_t, unsigned long attrs);
|
|
|
|
dma_addr_t (*map_page)(struct device *dev, struct page *page,
|
|
unsigned long offset, size_t size,
|
|
enum dma_data_direction dir,
|
|
unsigned long attrs);
|
|
void (*unmap_page)(struct device *dev, dma_addr_t dma_handle,
|
|
size_t size, enum dma_data_direction dir,
|
|
unsigned long attrs);
|
|
/*
|
|
* map_sg returns 0 on error and a value > 0 on success.
|
|
* It should never return a value < 0.
|
|
*/
|
|
int (*map_sg)(struct device *dev, struct scatterlist *sg,
|
|
int nents, enum dma_data_direction dir,
|
|
unsigned long attrs);
|
|
void (*unmap_sg)(struct device *dev,
|
|
struct scatterlist *sg, int nents,
|
|
enum dma_data_direction dir,
|
|
unsigned long attrs);
|
|
dma_addr_t (*map_resource)(struct device *dev, phys_addr_t phys_addr,
|
|
size_t size, enum dma_data_direction dir,
|
|
unsigned long attrs);
|
|
void (*unmap_resource)(struct device *dev, dma_addr_t dma_handle,
|
|
size_t size, enum dma_data_direction dir,
|
|
unsigned long attrs);
|
|
void (*sync_single_for_cpu)(struct device *dev,
|
|
dma_addr_t dma_handle, size_t size,
|
|
enum dma_data_direction dir);
|
|
void (*sync_single_for_device)(struct device *dev,
|
|
dma_addr_t dma_handle, size_t size,
|
|
enum dma_data_direction dir);
|
|
void (*sync_sg_for_cpu)(struct device *dev,
|
|
struct scatterlist *sg, int nents,
|
|
enum dma_data_direction dir);
|
|
void (*sync_sg_for_device)(struct device *dev,
|
|
struct scatterlist *sg, int nents,
|
|
enum dma_data_direction dir);
|
|
int (*mapping_error)(struct device *dev, dma_addr_t dma_addr);
|
|
int (*dma_supported)(struct device *dev, u64 mask);
|
|
int (*set_dma_mask)(struct device *dev, u64 mask);
|
|
void *(*remap)(struct device *dev, void *cpu_addr, dma_addr_t handle,
|
|
size_t size, unsigned long attrs);
|
|
void (*unremap)(struct device *dev, void *remapped_address,
|
|
size_t size);
|
|
#ifdef ARCH_HAS_DMA_GET_REQUIRED_MASK
|
|
u64 (*get_required_mask)(struct device *dev);
|
|
#endif
|
|
int is_phys;
|
|
};
|
|
|
|
extern const struct dma_map_ops dma_noop_ops;
|
|
extern const struct dma_map_ops dma_virt_ops;
|
|
|
|
#define DMA_BIT_MASK(n) (((n) == 64) ? ~0ULL : ((1ULL<<(n))-1))
|
|
|
|
#define DMA_MASK_NONE 0x0ULL
|
|
|
|
static inline int valid_dma_direction(int dma_direction)
|
|
{
|
|
return ((dma_direction == DMA_BIDIRECTIONAL) ||
|
|
(dma_direction == DMA_TO_DEVICE) ||
|
|
(dma_direction == DMA_FROM_DEVICE));
|
|
}
|
|
|
|
static inline int is_device_dma_capable(struct device *dev)
|
|
{
|
|
return dev->dma_mask != NULL && *dev->dma_mask != DMA_MASK_NONE;
|
|
}
|
|
|
|
#ifdef CONFIG_HAVE_GENERIC_DMA_COHERENT
|
|
/*
|
|
* These three functions are only for dma allocator.
|
|
* Don't use them in device drivers.
|
|
*/
|
|
int dma_alloc_from_dev_coherent(struct device *dev, ssize_t size,
|
|
dma_addr_t *dma_handle, void **ret);
|
|
int dma_release_from_dev_coherent(struct device *dev, int order, void *vaddr);
|
|
|
|
int dma_mmap_from_dev_coherent(struct device *dev, struct vm_area_struct *vma,
|
|
void *cpu_addr, size_t size, int *ret);
|
|
|
|
void *dma_alloc_from_global_coherent(ssize_t size, dma_addr_t *dma_handle);
|
|
int dma_release_from_global_coherent(int order, void *vaddr);
|
|
int dma_mmap_from_global_coherent(struct vm_area_struct *vma, void *cpu_addr,
|
|
size_t size, int *ret);
|
|
|
|
#else
|
|
#define dma_alloc_from_dev_coherent(dev, size, handle, ret) (0)
|
|
#define dma_release_from_dev_coherent(dev, order, vaddr) (0)
|
|
#define dma_mmap_from_dev_coherent(dev, vma, vaddr, order, ret) (0)
|
|
|
|
static inline void *dma_alloc_from_global_coherent(ssize_t size,
|
|
dma_addr_t *dma_handle)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
static inline int dma_release_from_global_coherent(int order, void *vaddr)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int dma_mmap_from_global_coherent(struct vm_area_struct *vma,
|
|
void *cpu_addr, size_t size,
|
|
int *ret)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif /* CONFIG_HAVE_GENERIC_DMA_COHERENT */
|
|
|
|
#ifdef CONFIG_HAS_DMA
|
|
#include <asm/dma-mapping.h>
|
|
static inline const struct dma_map_ops *get_dma_ops(struct device *dev)
|
|
{
|
|
if (dev && dev->dma_ops)
|
|
return dev->dma_ops;
|
|
return get_arch_dma_ops(dev ? dev->bus : NULL);
|
|
}
|
|
|
|
static inline void set_dma_ops(struct device *dev,
|
|
const struct dma_map_ops *dma_ops)
|
|
{
|
|
dev->dma_ops = dma_ops;
|
|
}
|
|
#else
|
|
/*
|
|
* Define the dma api to allow compilation but not linking of
|
|
* dma dependent code. Code that depends on the dma-mapping
|
|
* API needs to set 'depends on HAS_DMA' in its Kconfig
|
|
*/
|
|
extern const struct dma_map_ops bad_dma_ops;
|
|
static inline const struct dma_map_ops *get_dma_ops(struct device *dev)
|
|
{
|
|
return &bad_dma_ops;
|
|
}
|
|
#endif
|
|
|
|
static inline dma_addr_t dma_map_single_attrs(struct device *dev, void *ptr,
|
|
size_t size,
|
|
enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
dma_addr_t addr;
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
addr = ops->map_page(dev, virt_to_page(ptr),
|
|
offset_in_page(ptr), size,
|
|
dir, attrs);
|
|
debug_dma_map_page(dev, virt_to_page(ptr),
|
|
offset_in_page(ptr), size,
|
|
dir, addr, true);
|
|
return addr;
|
|
}
|
|
|
|
static inline void dma_unmap_single_attrs(struct device *dev, dma_addr_t addr,
|
|
size_t size,
|
|
enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
if (ops->unmap_page)
|
|
ops->unmap_page(dev, addr, size, dir, attrs);
|
|
debug_dma_unmap_page(dev, addr, size, dir, true);
|
|
}
|
|
|
|
/*
|
|
* dma_maps_sg_attrs returns 0 on error and > 0 on success.
|
|
* It should never return a value < 0.
|
|
*/
|
|
static inline int dma_map_sg_attrs(struct device *dev, struct scatterlist *sg,
|
|
int nents, enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
int ents;
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
ents = ops->map_sg(dev, sg, nents, dir, attrs);
|
|
BUG_ON(ents < 0);
|
|
debug_dma_map_sg(dev, sg, nents, ents, dir);
|
|
|
|
return ents;
|
|
}
|
|
|
|
static inline void dma_unmap_sg_attrs(struct device *dev, struct scatterlist *sg,
|
|
int nents, enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
debug_dma_unmap_sg(dev, sg, nents, dir);
|
|
if (ops->unmap_sg)
|
|
ops->unmap_sg(dev, sg, nents, dir, attrs);
|
|
}
|
|
|
|
static inline dma_addr_t dma_map_page_attrs(struct device *dev,
|
|
struct page *page,
|
|
size_t offset, size_t size,
|
|
enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
dma_addr_t addr;
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
addr = ops->map_page(dev, page, offset, size, dir, attrs);
|
|
debug_dma_map_page(dev, page, offset, size, dir, addr, false);
|
|
|
|
return addr;
|
|
}
|
|
|
|
static inline void dma_unmap_page_attrs(struct device *dev,
|
|
dma_addr_t addr, size_t size,
|
|
enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
if (ops->unmap_page)
|
|
ops->unmap_page(dev, addr, size, dir, attrs);
|
|
debug_dma_unmap_page(dev, addr, size, dir, false);
|
|
}
|
|
|
|
static inline dma_addr_t dma_map_resource(struct device *dev,
|
|
phys_addr_t phys_addr,
|
|
size_t size,
|
|
enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
dma_addr_t addr;
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
|
|
/* Don't allow RAM to be mapped */
|
|
BUG_ON(pfn_valid(PHYS_PFN(phys_addr)));
|
|
|
|
addr = phys_addr;
|
|
if (ops->map_resource)
|
|
addr = ops->map_resource(dev, phys_addr, size, dir, attrs);
|
|
|
|
debug_dma_map_resource(dev, phys_addr, size, dir, addr);
|
|
|
|
return addr;
|
|
}
|
|
|
|
static inline void dma_unmap_resource(struct device *dev, dma_addr_t addr,
|
|
size_t size, enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
if (ops->unmap_resource)
|
|
ops->unmap_resource(dev, addr, size, dir, attrs);
|
|
debug_dma_unmap_resource(dev, addr, size, dir);
|
|
}
|
|
|
|
static inline void dma_sync_single_for_cpu(struct device *dev, dma_addr_t addr,
|
|
size_t size,
|
|
enum dma_data_direction dir)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
if (ops->sync_single_for_cpu)
|
|
ops->sync_single_for_cpu(dev, addr, size, dir);
|
|
debug_dma_sync_single_for_cpu(dev, addr, size, dir);
|
|
}
|
|
|
|
static inline void dma_sync_single_for_device(struct device *dev,
|
|
dma_addr_t addr, size_t size,
|
|
enum dma_data_direction dir)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
if (ops->sync_single_for_device)
|
|
ops->sync_single_for_device(dev, addr, size, dir);
|
|
debug_dma_sync_single_for_device(dev, addr, size, dir);
|
|
}
|
|
|
|
static inline void dma_sync_single_range_for_cpu(struct device *dev,
|
|
dma_addr_t addr,
|
|
unsigned long offset,
|
|
size_t size,
|
|
enum dma_data_direction dir)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
if (ops->sync_single_for_cpu)
|
|
ops->sync_single_for_cpu(dev, addr + offset, size, dir);
|
|
debug_dma_sync_single_range_for_cpu(dev, addr, offset, size, dir);
|
|
}
|
|
|
|
static inline void dma_sync_single_range_for_device(struct device *dev,
|
|
dma_addr_t addr,
|
|
unsigned long offset,
|
|
size_t size,
|
|
enum dma_data_direction dir)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
if (ops->sync_single_for_device)
|
|
ops->sync_single_for_device(dev, addr + offset, size, dir);
|
|
debug_dma_sync_single_range_for_device(dev, addr, offset, size, dir);
|
|
}
|
|
|
|
static inline void
|
|
dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg,
|
|
int nelems, enum dma_data_direction dir)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
if (ops->sync_sg_for_cpu)
|
|
ops->sync_sg_for_cpu(dev, sg, nelems, dir);
|
|
debug_dma_sync_sg_for_cpu(dev, sg, nelems, dir);
|
|
}
|
|
|
|
static inline void
|
|
dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg,
|
|
int nelems, enum dma_data_direction dir)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!valid_dma_direction(dir));
|
|
if (ops->sync_sg_for_device)
|
|
ops->sync_sg_for_device(dev, sg, nelems, dir);
|
|
debug_dma_sync_sg_for_device(dev, sg, nelems, dir);
|
|
|
|
}
|
|
|
|
#define dma_map_single(d, a, s, r) dma_map_single_attrs(d, a, s, r, 0)
|
|
#define dma_unmap_single(d, a, s, r) dma_unmap_single_attrs(d, a, s, r, 0)
|
|
#define dma_map_sg(d, s, n, r) dma_map_sg_attrs(d, s, n, r, 0)
|
|
#define dma_unmap_sg(d, s, n, r) dma_unmap_sg_attrs(d, s, n, r, 0)
|
|
#define dma_map_page(d, p, o, s, r) dma_map_page_attrs(d, p, o, s, r, 0)
|
|
#define dma_unmap_page(d, a, s, r) dma_unmap_page_attrs(d, a, s, r, 0)
|
|
|
|
extern int dma_common_mmap(struct device *dev, struct vm_area_struct *vma,
|
|
void *cpu_addr, dma_addr_t dma_addr, size_t size);
|
|
|
|
void *dma_common_contiguous_remap(struct page *page, size_t size,
|
|
unsigned long vm_flags,
|
|
pgprot_t prot, const void *caller);
|
|
|
|
void *dma_common_pages_remap(struct page **pages, size_t size,
|
|
unsigned long vm_flags, pgprot_t prot,
|
|
const void *caller);
|
|
void dma_common_free_remap(void *cpu_addr, size_t size, unsigned long vm_flags,
|
|
bool nowarn);
|
|
|
|
/**
|
|
* dma_mmap_attrs - map a coherent DMA allocation into user space
|
|
* @dev: valid struct device pointer, or NULL for ISA and EISA-like devices
|
|
* @vma: vm_area_struct describing requested user mapping
|
|
* @cpu_addr: kernel CPU-view address returned from dma_alloc_attrs
|
|
* @handle: device-view address returned from dma_alloc_attrs
|
|
* @size: size of memory originally requested in dma_alloc_attrs
|
|
* @attrs: attributes of mapping properties requested in dma_alloc_attrs
|
|
*
|
|
* Map a coherent DMA buffer previously allocated by dma_alloc_attrs
|
|
* into user space. The coherent DMA buffer must not be freed by the
|
|
* driver until the user space mapping has been released.
|
|
*/
|
|
static inline int
|
|
dma_mmap_attrs(struct device *dev, struct vm_area_struct *vma, void *cpu_addr,
|
|
dma_addr_t dma_addr, size_t size, unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
BUG_ON(!ops);
|
|
if (ops->mmap)
|
|
return ops->mmap(dev, vma, cpu_addr, dma_addr, size, attrs);
|
|
return dma_common_mmap(dev, vma, cpu_addr, dma_addr, size);
|
|
}
|
|
|
|
#define dma_mmap_coherent(d, v, c, h, s) dma_mmap_attrs(d, v, c, h, s, 0)
|
|
|
|
int
|
|
dma_common_get_sgtable(struct device *dev, struct sg_table *sgt,
|
|
void *cpu_addr, dma_addr_t dma_addr, size_t size);
|
|
|
|
static inline int
|
|
dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, void *cpu_addr,
|
|
dma_addr_t dma_addr, size_t size,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
BUG_ON(!ops);
|
|
if (ops->get_sgtable)
|
|
return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size,
|
|
attrs);
|
|
return dma_common_get_sgtable(dev, sgt, cpu_addr, dma_addr, size);
|
|
}
|
|
|
|
#define dma_get_sgtable(d, t, v, h, s) dma_get_sgtable_attrs(d, t, v, h, s, 0)
|
|
|
|
#ifndef arch_dma_alloc_attrs
|
|
#define arch_dma_alloc_attrs(dev, flag) (true)
|
|
#endif
|
|
|
|
static inline void *dma_alloc_attrs(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, gfp_t flag,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
void *cpu_addr;
|
|
|
|
BUG_ON(!ops);
|
|
|
|
if (dma_alloc_from_dev_coherent(dev, size, dma_handle, &cpu_addr))
|
|
return cpu_addr;
|
|
|
|
if (!arch_dma_alloc_attrs(&dev, &flag))
|
|
return NULL;
|
|
if (!ops->alloc)
|
|
return NULL;
|
|
|
|
cpu_addr = ops->alloc(dev, size, dma_handle, flag, attrs);
|
|
debug_dma_alloc_coherent(dev, size, *dma_handle, cpu_addr);
|
|
return cpu_addr;
|
|
}
|
|
|
|
static inline void dma_free_attrs(struct device *dev, size_t size,
|
|
void *cpu_addr, dma_addr_t dma_handle,
|
|
unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
BUG_ON(!ops);
|
|
WARN_ON(irqs_disabled());
|
|
|
|
if (dma_release_from_dev_coherent(dev, get_order(size), cpu_addr))
|
|
return;
|
|
|
|
if (!ops->free || !cpu_addr)
|
|
return;
|
|
|
|
debug_dma_free_coherent(dev, size, cpu_addr, dma_handle);
|
|
ops->free(dev, size, cpu_addr, dma_handle, attrs);
|
|
}
|
|
|
|
static inline void *dma_alloc_coherent(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, gfp_t flag)
|
|
{
|
|
return dma_alloc_attrs(dev, size, dma_handle, flag, 0);
|
|
}
|
|
|
|
static inline void dma_free_coherent(struct device *dev, size_t size,
|
|
void *cpu_addr, dma_addr_t dma_handle)
|
|
{
|
|
return dma_free_attrs(dev, size, cpu_addr, dma_handle, 0);
|
|
}
|
|
|
|
static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
debug_dma_mapping_error(dev, dma_addr);
|
|
if (ops->mapping_error)
|
|
return ops->mapping_error(dev, dma_addr);
|
|
return 0;
|
|
}
|
|
|
|
static inline void dma_check_mask(struct device *dev, u64 mask)
|
|
{
|
|
if (sme_active() && (mask < (((u64)sme_get_me_mask() << 1) - 1)))
|
|
dev_warn(dev, "SME is active, device will require DMA bounce buffers\n");
|
|
}
|
|
|
|
static inline int dma_supported(struct device *dev, u64 mask)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
if (!ops)
|
|
return 0;
|
|
if (!ops->dma_supported)
|
|
return 1;
|
|
return ops->dma_supported(dev, mask);
|
|
}
|
|
|
|
#ifndef HAVE_ARCH_DMA_SET_MASK
|
|
static inline int dma_set_mask(struct device *dev, u64 mask)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
if (ops->set_dma_mask)
|
|
return ops->set_dma_mask(dev, mask);
|
|
|
|
if (!dev->dma_mask || !dma_supported(dev, mask))
|
|
return -EIO;
|
|
|
|
dma_check_mask(dev, mask);
|
|
|
|
*dev->dma_mask = mask;
|
|
return 0;
|
|
}
|
|
#endif
|
|
static inline void *dma_remap(struct device *dev, void *cpu_addr,
|
|
dma_addr_t dma_handle, size_t size, unsigned long attrs)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
if (!ops->remap) {
|
|
WARN_ONCE(1, "Remap function not implemented for %pS\n",
|
|
ops->remap);
|
|
return NULL;
|
|
}
|
|
|
|
return ops->remap(dev, cpu_addr, dma_handle, size, attrs);
|
|
}
|
|
|
|
|
|
static inline void dma_unremap(struct device *dev, void *remapped_addr,
|
|
size_t size)
|
|
{
|
|
const struct dma_map_ops *ops = get_dma_ops(dev);
|
|
|
|
if (!ops->unremap) {
|
|
WARN_ONCE(1, "unremap function not implemented for %pS\n",
|
|
ops->unremap);
|
|
return;
|
|
}
|
|
|
|
return ops->unremap(dev, remapped_addr, size);
|
|
}
|
|
|
|
|
|
static inline u64 dma_get_mask(struct device *dev)
|
|
{
|
|
if (dev && dev->dma_mask && *dev->dma_mask)
|
|
return *dev->dma_mask;
|
|
return DMA_BIT_MASK(32);
|
|
}
|
|
|
|
#ifdef CONFIG_ARCH_HAS_DMA_SET_COHERENT_MASK
|
|
int dma_set_coherent_mask(struct device *dev, u64 mask);
|
|
#else
|
|
static inline int dma_set_coherent_mask(struct device *dev, u64 mask)
|
|
{
|
|
if (!dma_supported(dev, mask))
|
|
return -EIO;
|
|
|
|
dma_check_mask(dev, mask);
|
|
|
|
dev->coherent_dma_mask = mask;
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* Set both the DMA mask and the coherent DMA mask to the same thing.
|
|
* Note that we don't check the return value from dma_set_coherent_mask()
|
|
* as the DMA API guarantees that the coherent DMA mask can be set to
|
|
* the same or smaller than the streaming DMA mask.
|
|
*/
|
|
static inline int dma_set_mask_and_coherent(struct device *dev, u64 mask)
|
|
{
|
|
int rc = dma_set_mask(dev, mask);
|
|
if (rc == 0)
|
|
dma_set_coherent_mask(dev, mask);
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* Similar to the above, except it deals with the case where the device
|
|
* does not have dev->dma_mask appropriately setup.
|
|
*/
|
|
static inline int dma_coerce_mask_and_coherent(struct device *dev, u64 mask)
|
|
{
|
|
dev->dma_mask = &dev->coherent_dma_mask;
|
|
return dma_set_mask_and_coherent(dev, mask);
|
|
}
|
|
|
|
extern u64 dma_get_required_mask(struct device *dev);
|
|
|
|
#ifndef arch_setup_dma_ops
|
|
static inline void arch_setup_dma_ops(struct device *dev, u64 dma_base,
|
|
u64 size, const struct iommu_ops *iommu,
|
|
bool coherent) { }
|
|
#endif
|
|
|
|
#ifndef arch_teardown_dma_ops
|
|
static inline void arch_teardown_dma_ops(struct device *dev) { }
|
|
#endif
|
|
|
|
static inline unsigned int dma_get_max_seg_size(struct device *dev)
|
|
{
|
|
if (dev->dma_parms && dev->dma_parms->max_segment_size)
|
|
return dev->dma_parms->max_segment_size;
|
|
return SZ_64K;
|
|
}
|
|
|
|
static inline int dma_set_max_seg_size(struct device *dev, unsigned int size)
|
|
{
|
|
if (dev->dma_parms) {
|
|
dev->dma_parms->max_segment_size = size;
|
|
return 0;
|
|
}
|
|
return -EIO;
|
|
}
|
|
|
|
static inline unsigned long dma_get_seg_boundary(struct device *dev)
|
|
{
|
|
if (dev->dma_parms && dev->dma_parms->segment_boundary_mask)
|
|
return dev->dma_parms->segment_boundary_mask;
|
|
return DMA_BIT_MASK(32);
|
|
}
|
|
|
|
static inline int dma_set_seg_boundary(struct device *dev, unsigned long mask)
|
|
{
|
|
if (dev->dma_parms) {
|
|
dev->dma_parms->segment_boundary_mask = mask;
|
|
return 0;
|
|
}
|
|
return -EIO;
|
|
}
|
|
|
|
#ifndef dma_max_pfn
|
|
static inline unsigned long dma_max_pfn(struct device *dev)
|
|
{
|
|
return *dev->dma_mask >> PAGE_SHIFT;
|
|
}
|
|
#endif
|
|
|
|
static inline void *dma_zalloc_coherent(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, gfp_t flag)
|
|
{
|
|
void *ret = dma_alloc_coherent(dev, size, dma_handle,
|
|
flag | __GFP_ZERO);
|
|
return ret;
|
|
}
|
|
|
|
static inline int dma_get_cache_alignment(void)
|
|
{
|
|
#ifdef ARCH_DMA_MINALIGN
|
|
return ARCH_DMA_MINALIGN;
|
|
#endif
|
|
return 1;
|
|
}
|
|
|
|
/* flags for the coherent memory api */
|
|
#define DMA_MEMORY_EXCLUSIVE 0x01
|
|
|
|
#ifdef CONFIG_HAVE_GENERIC_DMA_COHERENT
|
|
int dma_declare_coherent_memory(struct device *dev, phys_addr_t phys_addr,
|
|
dma_addr_t device_addr, size_t size, int flags);
|
|
void dma_release_declared_memory(struct device *dev);
|
|
void *dma_mark_declared_memory_occupied(struct device *dev,
|
|
dma_addr_t device_addr, size_t size);
|
|
dma_addr_t dma_get_device_base(struct device *dev,
|
|
struct dma_coherent_mem *mem);
|
|
unsigned long dma_get_size(struct dma_coherent_mem *mem);
|
|
|
|
#else
|
|
static inline int
|
|
dma_declare_coherent_memory(struct device *dev, phys_addr_t phys_addr,
|
|
dma_addr_t device_addr, size_t size, int flags)
|
|
{
|
|
return -ENOSYS;
|
|
}
|
|
|
|
static inline void
|
|
dma_release_declared_memory(struct device *dev)
|
|
{
|
|
}
|
|
|
|
static inline void *
|
|
dma_mark_declared_memory_occupied(struct device *dev,
|
|
dma_addr_t device_addr, size_t size)
|
|
{
|
|
return ERR_PTR(-EBUSY);
|
|
}
|
|
static inline dma_addr_t
|
|
dma_get_device_base(struct device *dev, struct dma_coherent_mem *mem)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline unsigned long dma_get_size(struct dma_coherent_mem *mem)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
#endif /* CONFIG_HAVE_GENERIC_DMA_COHERENT */
|
|
|
|
#ifdef CONFIG_HAS_DMA
|
|
int dma_configure(struct device *dev);
|
|
void dma_deconfigure(struct device *dev);
|
|
#else
|
|
static inline int dma_configure(struct device *dev)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline void dma_deconfigure(struct device *dev) {}
|
|
#endif
|
|
|
|
/*
|
|
* Managed DMA API
|
|
*/
|
|
extern void *dmam_alloc_coherent(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, gfp_t gfp);
|
|
extern void dmam_free_coherent(struct device *dev, size_t size, void *vaddr,
|
|
dma_addr_t dma_handle);
|
|
extern void *dmam_alloc_attrs(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, gfp_t gfp,
|
|
unsigned long attrs);
|
|
#ifdef CONFIG_HAVE_GENERIC_DMA_COHERENT
|
|
extern int dmam_declare_coherent_memory(struct device *dev,
|
|
phys_addr_t phys_addr,
|
|
dma_addr_t device_addr, size_t size,
|
|
int flags);
|
|
extern void dmam_release_declared_memory(struct device *dev);
|
|
#else /* CONFIG_HAVE_GENERIC_DMA_COHERENT */
|
|
static inline int dmam_declare_coherent_memory(struct device *dev,
|
|
phys_addr_t phys_addr, dma_addr_t device_addr,
|
|
size_t size, gfp_t gfp)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline void dmam_release_declared_memory(struct device *dev)
|
|
{
|
|
}
|
|
#endif /* CONFIG_HAVE_GENERIC_DMA_COHERENT */
|
|
|
|
static inline void *dma_alloc_wc(struct device *dev, size_t size,
|
|
dma_addr_t *dma_addr, gfp_t gfp)
|
|
{
|
|
return dma_alloc_attrs(dev, size, dma_addr, gfp,
|
|
DMA_ATTR_WRITE_COMBINE);
|
|
}
|
|
#ifndef dma_alloc_writecombine
|
|
#define dma_alloc_writecombine dma_alloc_wc
|
|
#endif
|
|
|
|
static inline void dma_free_wc(struct device *dev, size_t size,
|
|
void *cpu_addr, dma_addr_t dma_addr)
|
|
{
|
|
return dma_free_attrs(dev, size, cpu_addr, dma_addr,
|
|
DMA_ATTR_WRITE_COMBINE);
|
|
}
|
|
#ifndef dma_free_writecombine
|
|
#define dma_free_writecombine dma_free_wc
|
|
#endif
|
|
|
|
static inline int dma_mmap_wc(struct device *dev,
|
|
struct vm_area_struct *vma,
|
|
void *cpu_addr, dma_addr_t dma_addr,
|
|
size_t size)
|
|
{
|
|
return dma_mmap_attrs(dev, vma, cpu_addr, dma_addr, size,
|
|
DMA_ATTR_WRITE_COMBINE);
|
|
}
|
|
#ifndef dma_mmap_writecombine
|
|
#define dma_mmap_writecombine dma_mmap_wc
|
|
#endif
|
|
|
|
static inline void *dma_alloc_nonconsistent(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, gfp_t flag)
|
|
{
|
|
unsigned long attrs = DMA_ATTR_NON_CONSISTENT;
|
|
|
|
return dma_alloc_attrs(dev, size, dma_handle, flag, attrs);
|
|
}
|
|
|
|
static inline void dma_free_nonconsistent(struct device *dev, size_t size,
|
|
void *cpu_addr, dma_addr_t dma_handle)
|
|
{
|
|
unsigned long attrs = DMA_ATTR_NON_CONSISTENT;
|
|
|
|
return dma_free_attrs(dev, size, cpu_addr, dma_handle, attrs);
|
|
}
|
|
|
|
static inline int dma_mmap_nonconsistent(struct device *dev,
|
|
struct vm_area_struct *vma, void *cpu_addr,
|
|
dma_addr_t dma_addr, size_t size)
|
|
{
|
|
unsigned long attrs = DMA_ATTR_NON_CONSISTENT;
|
|
|
|
return dma_mmap_attrs(dev, vma, cpu_addr, dma_addr, size, attrs);
|
|
}
|
|
|
|
#if defined(CONFIG_NEED_DMA_MAP_STATE) || defined(CONFIG_DMA_API_DEBUG)
|
|
#define DEFINE_DMA_UNMAP_ADDR(ADDR_NAME) dma_addr_t ADDR_NAME
|
|
#define DEFINE_DMA_UNMAP_LEN(LEN_NAME) __u32 LEN_NAME
|
|
#define dma_unmap_addr(PTR, ADDR_NAME) ((PTR)->ADDR_NAME)
|
|
#define dma_unmap_addr_set(PTR, ADDR_NAME, VAL) (((PTR)->ADDR_NAME) = (VAL))
|
|
#define dma_unmap_len(PTR, LEN_NAME) ((PTR)->LEN_NAME)
|
|
#define dma_unmap_len_set(PTR, LEN_NAME, VAL) (((PTR)->LEN_NAME) = (VAL))
|
|
#else
|
|
#define DEFINE_DMA_UNMAP_ADDR(ADDR_NAME)
|
|
#define DEFINE_DMA_UNMAP_LEN(LEN_NAME)
|
|
#define dma_unmap_addr(PTR, ADDR_NAME) (0)
|
|
#define dma_unmap_addr_set(PTR, ADDR_NAME, VAL) do { } while (0)
|
|
#define dma_unmap_len(PTR, LEN_NAME) (0)
|
|
#define dma_unmap_len_set(PTR, LEN_NAME, VAL) do { } while (0)
|
|
#endif
|
|
|
|
#endif
|