Linux Audio

Check our new training course

Loading...
v5.4
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Copyright (C) 2012 ARM Ltd.
  4 * Copyright (c) 2014 The Linux Foundation
  5 */
  6#include <linux/dma-direct.h>
  7#include <linux/dma-noncoherent.h>
  8#include <linux/dma-contiguous.h>
  9#include <linux/init.h>
 10#include <linux/genalloc.h>
 11#include <linux/slab.h>
 12#include <linux/vmalloc.h>
 13
 14struct page **dma_common_find_pages(void *cpu_addr)
 15{
 16	struct vm_struct *area = find_vm_area(cpu_addr);
 17
 18	if (!area || area->flags != VM_DMA_COHERENT)
 19		return NULL;
 20	return area->pages;
 21}
 22
 23static struct vm_struct *__dma_common_pages_remap(struct page **pages,
 24			size_t size, pgprot_t prot, const void *caller)
 25{
 26	struct vm_struct *area;
 27
 28	area = get_vm_area_caller(size, VM_DMA_COHERENT, caller);
 29	if (!area)
 30		return NULL;
 31
 32	if (map_vm_area(area, prot, pages)) {
 33		vunmap(area->addr);
 34		return NULL;
 35	}
 36
 37	return area;
 38}
 39
 40/*
 41 * Remaps an array of PAGE_SIZE pages into another vm_area.
 42 * Cannot be used in non-sleeping contexts
 43 */
 44void *dma_common_pages_remap(struct page **pages, size_t size,
 45			 pgprot_t prot, const void *caller)
 46{
 47	struct vm_struct *area;
 48
 49	area = __dma_common_pages_remap(pages, size, prot, caller);
 50	if (!area)
 51		return NULL;
 52
 53	area->pages = pages;
 54
 55	return area->addr;
 56}
 57
 58/*
 59 * Remaps an allocated contiguous region into another vm_area.
 60 * Cannot be used in non-sleeping contexts
 61 */
 62void *dma_common_contiguous_remap(struct page *page, size_t size,
 63			pgprot_t prot, const void *caller)
 64{
 65	int i;
 66	struct page **pages;
 67	struct vm_struct *area;
 
 68
 69	pages = kmalloc(sizeof(struct page *) << get_order(size), GFP_KERNEL);
 70	if (!pages)
 71		return NULL;
 72
 73	for (i = 0; i < (size >> PAGE_SHIFT); i++)
 74		pages[i] = nth_page(page, i);
 
 
 75
 76	area = __dma_common_pages_remap(pages, size, prot, caller);
 77
 78	kfree(pages);
 79
 80	if (!area)
 81		return NULL;
 82	return area->addr;
 83}
 84
 85/*
 86 * Unmaps a range previously mapped by dma_common_*_remap
 87 */
 88void dma_common_free_remap(void *cpu_addr, size_t size)
 89{
 90	struct vm_struct *area = find_vm_area(cpu_addr);
 91
 92	if (!area || area->flags != VM_DMA_COHERENT) {
 93		WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr);
 94		return;
 95	}
 96
 97	unmap_kernel_range((unsigned long)cpu_addr, PAGE_ALIGN(size));
 98	vunmap(cpu_addr);
 99}
100
101#ifdef CONFIG_DMA_DIRECT_REMAP
102static struct gen_pool *atomic_pool __ro_after_init;
103
104#define DEFAULT_DMA_COHERENT_POOL_SIZE  SZ_256K
105static size_t atomic_pool_size __initdata = DEFAULT_DMA_COHERENT_POOL_SIZE;
106
107static int __init early_coherent_pool(char *p)
108{
109	atomic_pool_size = memparse(p, &p);
110	return 0;
111}
112early_param("coherent_pool", early_coherent_pool);
113
114static gfp_t dma_atomic_pool_gfp(void)
115{
116	if (IS_ENABLED(CONFIG_ZONE_DMA))
117		return GFP_DMA;
118	if (IS_ENABLED(CONFIG_ZONE_DMA32))
119		return GFP_DMA32;
120	return GFP_KERNEL;
121}
122
123static int __init dma_atomic_pool_init(void)
124{
125	unsigned int pool_size_order = get_order(atomic_pool_size);
126	unsigned long nr_pages = atomic_pool_size >> PAGE_SHIFT;
127	struct page *page;
128	void *addr;
129	int ret;
130
131	if (dev_get_cma_area(NULL))
132		page = dma_alloc_from_contiguous(NULL, nr_pages,
133						 pool_size_order, false);
134	else
135		page = alloc_pages(dma_atomic_pool_gfp(), pool_size_order);
136	if (!page)
137		goto out;
138
139	arch_dma_prep_coherent(page, atomic_pool_size);
140
141	atomic_pool = gen_pool_create(PAGE_SHIFT, -1);
142	if (!atomic_pool)
143		goto free_page;
144
145	addr = dma_common_contiguous_remap(page, atomic_pool_size,
146					   pgprot_dmacoherent(PAGE_KERNEL),
147					   __builtin_return_address(0));
148	if (!addr)
149		goto destroy_genpool;
150
151	ret = gen_pool_add_virt(atomic_pool, (unsigned long)addr,
152				page_to_phys(page), atomic_pool_size, -1);
153	if (ret)
154		goto remove_mapping;
155	gen_pool_set_algo(atomic_pool, gen_pool_first_fit_order_align, NULL);
156
157	pr_info("DMA: preallocated %zu KiB pool for atomic allocations\n",
158		atomic_pool_size / 1024);
159	return 0;
160
161remove_mapping:
162	dma_common_free_remap(addr, atomic_pool_size);
163destroy_genpool:
164	gen_pool_destroy(atomic_pool);
165	atomic_pool = NULL;
166free_page:
167	if (!dma_release_from_contiguous(NULL, page, nr_pages))
168		__free_pages(page, pool_size_order);
169out:
170	pr_err("DMA: failed to allocate %zu KiB pool for atomic coherent allocation\n",
171		atomic_pool_size / 1024);
172	return -ENOMEM;
173}
174postcore_initcall(dma_atomic_pool_init);
175
176bool dma_in_atomic_pool(void *start, size_t size)
177{
178	if (unlikely(!atomic_pool))
179		return false;
180
181	return addr_in_gen_pool(atomic_pool, (unsigned long)start, size);
182}
183
184void *dma_alloc_from_pool(size_t size, struct page **ret_page, gfp_t flags)
185{
186	unsigned long val;
187	void *ptr = NULL;
188
189	if (!atomic_pool) {
190		WARN(1, "coherent pool not initialised!\n");
191		return NULL;
192	}
193
194	val = gen_pool_alloc(atomic_pool, size);
195	if (val) {
196		phys_addr_t phys = gen_pool_virt_to_phys(atomic_pool, val);
197
198		*ret_page = pfn_to_page(__phys_to_pfn(phys));
199		ptr = (void *)val;
200		memset(ptr, 0, size);
201	}
202
203	return ptr;
204}
205
206bool dma_free_from_pool(void *start, size_t size)
207{
208	if (!dma_in_atomic_pool(start, size))
209		return false;
210	gen_pool_free(atomic_pool, (unsigned long)start, size);
211	return true;
212}
213
214void *arch_dma_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle,
215		gfp_t flags, unsigned long attrs)
216{
217	struct page *page = NULL;
218	void *ret;
219
220	size = PAGE_ALIGN(size);
221
222	if (!gfpflags_allow_blocking(flags)) {
223		ret = dma_alloc_from_pool(size, &page, flags);
224		if (!ret)
225			return NULL;
226		goto done;
227	}
228
229	page = __dma_direct_alloc_pages(dev, size, dma_handle, flags, attrs);
230	if (!page)
231		return NULL;
232
233	/* remove any dirty cache lines on the kernel alias */
234	arch_dma_prep_coherent(page, size);
235
236	/* create a coherent mapping */
237	ret = dma_common_contiguous_remap(page, size,
238			dma_pgprot(dev, PAGE_KERNEL, attrs),
239			__builtin_return_address(0));
240	if (!ret) {
241		__dma_direct_free_pages(dev, size, page);
242		return ret;
243	}
244
245	memset(ret, 0, size);
246done:
247	*dma_handle = phys_to_dma(dev, page_to_phys(page));
248	return ret;
249}
250
251void arch_dma_free(struct device *dev, size_t size, void *vaddr,
252		dma_addr_t dma_handle, unsigned long attrs)
253{
254	if (!dma_free_from_pool(vaddr, PAGE_ALIGN(size))) {
255		phys_addr_t phys = dma_to_phys(dev, dma_handle);
256		struct page *page = pfn_to_page(__phys_to_pfn(phys));
257
258		vunmap(vaddr);
259		__dma_direct_free_pages(dev, size, page);
260	}
261}
262
263long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr,
264		dma_addr_t dma_addr)
265{
266	return __phys_to_pfn(dma_to_phys(dev, dma_addr));
267}
268#endif /* CONFIG_DMA_DIRECT_REMAP */
v6.8
 1// SPDX-License-Identifier: GPL-2.0
 2/*
 
 3 * Copyright (c) 2014 The Linux Foundation
 4 */
 5#include <linux/dma-map-ops.h>
 
 
 
 
 6#include <linux/slab.h>
 7#include <linux/vmalloc.h>
 8
 9struct page **dma_common_find_pages(void *cpu_addr)
10{
11	struct vm_struct *area = find_vm_area(cpu_addr);
12
13	if (!area || area->flags != VM_DMA_COHERENT)
14		return NULL;
15	return area->pages;
16}
17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18/*
19 * Remaps an array of PAGE_SIZE pages into another vm_area.
20 * Cannot be used in non-sleeping contexts
21 */
22void *dma_common_pages_remap(struct page **pages, size_t size,
23			 pgprot_t prot, const void *caller)
24{
25	void *vaddr;
26
27	vaddr = vmap(pages, PAGE_ALIGN(size) >> PAGE_SHIFT,
28		     VM_DMA_COHERENT, prot);
29	if (vaddr)
30		find_vm_area(vaddr)->pages = pages;
31	return vaddr;
 
 
32}
33
34/*
35 * Remaps an allocated contiguous region into another vm_area.
36 * Cannot be used in non-sleeping contexts
37 */
38void *dma_common_contiguous_remap(struct page *page, size_t size,
39			pgprot_t prot, const void *caller)
40{
41	int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
42	struct page **pages;
43	void *vaddr;
44	int i;
45
46	pages = kvmalloc_array(count, sizeof(struct page *), GFP_KERNEL);
47	if (!pages)
48		return NULL;
49	for (i = 0; i < count; i++)
 
50		pages[i] = nth_page(page, i);
51	vaddr = vmap(pages, count, VM_DMA_COHERENT, prot);
52	kvfree(pages);
53
54	return vaddr;
 
 
 
 
 
 
55}
56
57/*
58 * Unmaps a range previously mapped by dma_common_*_remap
59 */
60void dma_common_free_remap(void *cpu_addr, size_t size)
61{
62	struct vm_struct *area = find_vm_area(cpu_addr);
63
64	if (!area || area->flags != VM_DMA_COHERENT) {
65		WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr);
66		return;
67	}
68
 
69	vunmap(cpu_addr);
70}