Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Driver for FPGA Accelerated Function Unit (AFU) DMA Region Management
4 *
5 * Copyright (C) 2017-2018 Intel Corporation, Inc.
6 *
7 * Authors:
8 * Wu Hao <hao.wu@intel.com>
9 * Xiao Guangrong <guangrong.xiao@linux.intel.com>
10 */
11
12#include <linux/dma-mapping.h>
13#include <linux/sched/signal.h>
14#include <linux/uaccess.h>
15#include <linux/mm.h>
16
17#include "dfl-afu.h"
18
19void afu_dma_region_init(struct dfl_feature_platform_data *pdata)
20{
21 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
22
23 afu->dma_regions = RB_ROOT;
24}
25
26/**
27 * afu_dma_pin_pages - pin pages of given dma memory region
28 * @pdata: feature device platform data
29 * @region: dma memory region to be pinned
30 *
31 * Pin all the pages of given dfl_afu_dma_region.
32 * Return 0 for success or negative error code.
33 */
34static int afu_dma_pin_pages(struct dfl_feature_platform_data *pdata,
35 struct dfl_afu_dma_region *region)
36{
37 int npages = region->length >> PAGE_SHIFT;
38 struct device *dev = &pdata->dev->dev;
39 int ret, pinned;
40
41 ret = account_locked_vm(current->mm, npages, true);
42 if (ret)
43 return ret;
44
45 region->pages = kcalloc(npages, sizeof(struct page *), GFP_KERNEL);
46 if (!region->pages) {
47 ret = -ENOMEM;
48 goto unlock_vm;
49 }
50
51 pinned = pin_user_pages_fast(region->user_addr, npages, FOLL_WRITE,
52 region->pages);
53 if (pinned < 0) {
54 ret = pinned;
55 goto free_pages;
56 } else if (pinned != npages) {
57 ret = -EFAULT;
58 goto unpin_pages;
59 }
60
61 dev_dbg(dev, "%d pages pinned\n", pinned);
62
63 return 0;
64
65unpin_pages:
66 unpin_user_pages(region->pages, pinned);
67free_pages:
68 kfree(region->pages);
69unlock_vm:
70 account_locked_vm(current->mm, npages, false);
71 return ret;
72}
73
74/**
75 * afu_dma_unpin_pages - unpin pages of given dma memory region
76 * @pdata: feature device platform data
77 * @region: dma memory region to be unpinned
78 *
79 * Unpin all the pages of given dfl_afu_dma_region.
80 * Return 0 for success or negative error code.
81 */
82static void afu_dma_unpin_pages(struct dfl_feature_platform_data *pdata,
83 struct dfl_afu_dma_region *region)
84{
85 long npages = region->length >> PAGE_SHIFT;
86 struct device *dev = &pdata->dev->dev;
87
88 unpin_user_pages(region->pages, npages);
89 kfree(region->pages);
90 account_locked_vm(current->mm, npages, false);
91
92 dev_dbg(dev, "%ld pages unpinned\n", npages);
93}
94
95/**
96 * afu_dma_check_continuous_pages - check if pages are continuous
97 * @region: dma memory region
98 *
99 * Return true if pages of given dma memory region have continuous physical
100 * address, otherwise return false.
101 */
102static bool afu_dma_check_continuous_pages(struct dfl_afu_dma_region *region)
103{
104 int npages = region->length >> PAGE_SHIFT;
105 int i;
106
107 for (i = 0; i < npages - 1; i++)
108 if (page_to_pfn(region->pages[i]) + 1 !=
109 page_to_pfn(region->pages[i + 1]))
110 return false;
111
112 return true;
113}
114
115/**
116 * dma_region_check_iova - check if memory area is fully contained in the region
117 * @region: dma memory region
118 * @iova: address of the dma memory area
119 * @size: size of the dma memory area
120 *
121 * Compare the dma memory area defined by @iova and @size with given dma region.
122 * Return true if memory area is fully contained in the region, otherwise false.
123 */
124static bool dma_region_check_iova(struct dfl_afu_dma_region *region,
125 u64 iova, u64 size)
126{
127 if (!size && region->iova != iova)
128 return false;
129
130 return (region->iova <= iova) &&
131 (region->length + region->iova >= iova + size);
132}
133
134/**
135 * afu_dma_region_add - add given dma region to rbtree
136 * @pdata: feature device platform data
137 * @region: dma region to be added
138 *
139 * Return 0 for success, -EEXIST if dma region has already been added.
140 *
141 * Needs to be called with pdata->lock heold.
142 */
143static int afu_dma_region_add(struct dfl_feature_platform_data *pdata,
144 struct dfl_afu_dma_region *region)
145{
146 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
147 struct rb_node **new, *parent = NULL;
148
149 dev_dbg(&pdata->dev->dev, "add region (iova = %llx)\n",
150 (unsigned long long)region->iova);
151
152 new = &afu->dma_regions.rb_node;
153
154 while (*new) {
155 struct dfl_afu_dma_region *this;
156
157 this = container_of(*new, struct dfl_afu_dma_region, node);
158
159 parent = *new;
160
161 if (dma_region_check_iova(this, region->iova, region->length))
162 return -EEXIST;
163
164 if (region->iova < this->iova)
165 new = &((*new)->rb_left);
166 else if (region->iova > this->iova)
167 new = &((*new)->rb_right);
168 else
169 return -EEXIST;
170 }
171
172 rb_link_node(®ion->node, parent, new);
173 rb_insert_color(®ion->node, &afu->dma_regions);
174
175 return 0;
176}
177
178/**
179 * afu_dma_region_remove - remove given dma region from rbtree
180 * @pdata: feature device platform data
181 * @region: dma region to be removed
182 *
183 * Needs to be called with pdata->lock heold.
184 */
185static void afu_dma_region_remove(struct dfl_feature_platform_data *pdata,
186 struct dfl_afu_dma_region *region)
187{
188 struct dfl_afu *afu;
189
190 dev_dbg(&pdata->dev->dev, "del region (iova = %llx)\n",
191 (unsigned long long)region->iova);
192
193 afu = dfl_fpga_pdata_get_private(pdata);
194 rb_erase(®ion->node, &afu->dma_regions);
195}
196
197/**
198 * afu_dma_region_destroy - destroy all regions in rbtree
199 * @pdata: feature device platform data
200 *
201 * Needs to be called with pdata->lock heold.
202 */
203void afu_dma_region_destroy(struct dfl_feature_platform_data *pdata)
204{
205 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
206 struct rb_node *node = rb_first(&afu->dma_regions);
207 struct dfl_afu_dma_region *region;
208
209 while (node) {
210 region = container_of(node, struct dfl_afu_dma_region, node);
211
212 dev_dbg(&pdata->dev->dev, "del region (iova = %llx)\n",
213 (unsigned long long)region->iova);
214
215 rb_erase(node, &afu->dma_regions);
216
217 if (region->iova)
218 dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
219 region->iova, region->length,
220 DMA_BIDIRECTIONAL);
221
222 if (region->pages)
223 afu_dma_unpin_pages(pdata, region);
224
225 node = rb_next(node);
226 kfree(region);
227 }
228}
229
230/**
231 * afu_dma_region_find - find the dma region from rbtree based on iova and size
232 * @pdata: feature device platform data
233 * @iova: address of the dma memory area
234 * @size: size of the dma memory area
235 *
236 * It finds the dma region from the rbtree based on @iova and @size:
237 * - if @size == 0, it finds the dma region which starts from @iova
238 * - otherwise, it finds the dma region which fully contains
239 * [@iova, @iova+size)
240 * If nothing is matched returns NULL.
241 *
242 * Needs to be called with pdata->lock held.
243 */
244struct dfl_afu_dma_region *
245afu_dma_region_find(struct dfl_feature_platform_data *pdata, u64 iova, u64 size)
246{
247 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
248 struct rb_node *node = afu->dma_regions.rb_node;
249 struct device *dev = &pdata->dev->dev;
250
251 while (node) {
252 struct dfl_afu_dma_region *region;
253
254 region = container_of(node, struct dfl_afu_dma_region, node);
255
256 if (dma_region_check_iova(region, iova, size)) {
257 dev_dbg(dev, "find region (iova = %llx)\n",
258 (unsigned long long)region->iova);
259 return region;
260 }
261
262 if (iova < region->iova)
263 node = node->rb_left;
264 else if (iova > region->iova)
265 node = node->rb_right;
266 else
267 /* the iova region is not fully covered. */
268 break;
269 }
270
271 dev_dbg(dev, "region with iova %llx and size %llx is not found\n",
272 (unsigned long long)iova, (unsigned long long)size);
273
274 return NULL;
275}
276
277/**
278 * afu_dma_region_find_iova - find the dma region from rbtree by iova
279 * @pdata: feature device platform data
280 * @iova: address of the dma region
281 *
282 * Needs to be called with pdata->lock held.
283 */
284static struct dfl_afu_dma_region *
285afu_dma_region_find_iova(struct dfl_feature_platform_data *pdata, u64 iova)
286{
287 return afu_dma_region_find(pdata, iova, 0);
288}
289
290/**
291 * afu_dma_map_region - map memory region for dma
292 * @pdata: feature device platform data
293 * @user_addr: address of the memory region
294 * @length: size of the memory region
295 * @iova: pointer of iova address
296 *
297 * Map memory region defined by @user_addr and @length, and return dma address
298 * of the memory region via @iova.
299 * Return 0 for success, otherwise error code.
300 */
301int afu_dma_map_region(struct dfl_feature_platform_data *pdata,
302 u64 user_addr, u64 length, u64 *iova)
303{
304 struct dfl_afu_dma_region *region;
305 int ret;
306
307 /*
308 * Check Inputs, only accept page-aligned user memory region with
309 * valid length.
310 */
311 if (!PAGE_ALIGNED(user_addr) || !PAGE_ALIGNED(length) || !length)
312 return -EINVAL;
313
314 /* Check overflow */
315 if (user_addr + length < user_addr)
316 return -EINVAL;
317
318 region = kzalloc(sizeof(*region), GFP_KERNEL);
319 if (!region)
320 return -ENOMEM;
321
322 region->user_addr = user_addr;
323 region->length = length;
324
325 /* Pin the user memory region */
326 ret = afu_dma_pin_pages(pdata, region);
327 if (ret) {
328 dev_err(&pdata->dev->dev, "failed to pin memory region\n");
329 goto free_region;
330 }
331
332 /* Only accept continuous pages, return error else */
333 if (!afu_dma_check_continuous_pages(region)) {
334 dev_err(&pdata->dev->dev, "pages are not continuous\n");
335 ret = -EINVAL;
336 goto unpin_pages;
337 }
338
339 /* As pages are continuous then start to do DMA mapping */
340 region->iova = dma_map_page(dfl_fpga_pdata_to_parent(pdata),
341 region->pages[0], 0,
342 region->length,
343 DMA_BIDIRECTIONAL);
344 if (dma_mapping_error(dfl_fpga_pdata_to_parent(pdata), region->iova)) {
345 dev_err(&pdata->dev->dev, "failed to map for dma\n");
346 ret = -EFAULT;
347 goto unpin_pages;
348 }
349
350 *iova = region->iova;
351
352 mutex_lock(&pdata->lock);
353 ret = afu_dma_region_add(pdata, region);
354 mutex_unlock(&pdata->lock);
355 if (ret) {
356 dev_err(&pdata->dev->dev, "failed to add dma region\n");
357 goto unmap_dma;
358 }
359
360 return 0;
361
362unmap_dma:
363 dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
364 region->iova, region->length, DMA_BIDIRECTIONAL);
365unpin_pages:
366 afu_dma_unpin_pages(pdata, region);
367free_region:
368 kfree(region);
369 return ret;
370}
371
372/**
373 * afu_dma_unmap_region - unmap dma memory region
374 * @pdata: feature device platform data
375 * @iova: dma address of the region
376 *
377 * Unmap dma memory region based on @iova.
378 * Return 0 for success, otherwise error code.
379 */
380int afu_dma_unmap_region(struct dfl_feature_platform_data *pdata, u64 iova)
381{
382 struct dfl_afu_dma_region *region;
383
384 mutex_lock(&pdata->lock);
385 region = afu_dma_region_find_iova(pdata, iova);
386 if (!region) {
387 mutex_unlock(&pdata->lock);
388 return -EINVAL;
389 }
390
391 if (region->in_use) {
392 mutex_unlock(&pdata->lock);
393 return -EBUSY;
394 }
395
396 afu_dma_region_remove(pdata, region);
397 mutex_unlock(&pdata->lock);
398
399 dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
400 region->iova, region->length, DMA_BIDIRECTIONAL);
401 afu_dma_unpin_pages(pdata, region);
402 kfree(region);
403
404 return 0;
405}
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Driver for FPGA Accelerated Function Unit (AFU) DMA Region Management
4 *
5 * Copyright (C) 2017-2018 Intel Corporation, Inc.
6 *
7 * Authors:
8 * Wu Hao <hao.wu@intel.com>
9 * Xiao Guangrong <guangrong.xiao@linux.intel.com>
10 */
11
12#include <linux/dma-mapping.h>
13#include <linux/sched/signal.h>
14#include <linux/uaccess.h>
15#include <linux/mm.h>
16
17#include "dfl-afu.h"
18
19static void put_all_pages(struct page **pages, int npages)
20{
21 int i;
22
23 for (i = 0; i < npages; i++)
24 if (pages[i])
25 put_page(pages[i]);
26}
27
28void afu_dma_region_init(struct dfl_feature_platform_data *pdata)
29{
30 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
31
32 afu->dma_regions = RB_ROOT;
33}
34
35/**
36 * afu_dma_pin_pages - pin pages of given dma memory region
37 * @pdata: feature device platform data
38 * @region: dma memory region to be pinned
39 *
40 * Pin all the pages of given dfl_afu_dma_region.
41 * Return 0 for success or negative error code.
42 */
43static int afu_dma_pin_pages(struct dfl_feature_platform_data *pdata,
44 struct dfl_afu_dma_region *region)
45{
46 int npages = region->length >> PAGE_SHIFT;
47 struct device *dev = &pdata->dev->dev;
48 int ret, pinned;
49
50 ret = account_locked_vm(current->mm, npages, true);
51 if (ret)
52 return ret;
53
54 region->pages = kcalloc(npages, sizeof(struct page *), GFP_KERNEL);
55 if (!region->pages) {
56 ret = -ENOMEM;
57 goto unlock_vm;
58 }
59
60 pinned = get_user_pages_fast(region->user_addr, npages, FOLL_WRITE,
61 region->pages);
62 if (pinned < 0) {
63 ret = pinned;
64 goto put_pages;
65 } else if (pinned != npages) {
66 ret = -EFAULT;
67 goto free_pages;
68 }
69
70 dev_dbg(dev, "%d pages pinned\n", pinned);
71
72 return 0;
73
74put_pages:
75 put_all_pages(region->pages, pinned);
76free_pages:
77 kfree(region->pages);
78unlock_vm:
79 account_locked_vm(current->mm, npages, false);
80 return ret;
81}
82
83/**
84 * afu_dma_unpin_pages - unpin pages of given dma memory region
85 * @pdata: feature device platform data
86 * @region: dma memory region to be unpinned
87 *
88 * Unpin all the pages of given dfl_afu_dma_region.
89 * Return 0 for success or negative error code.
90 */
91static void afu_dma_unpin_pages(struct dfl_feature_platform_data *pdata,
92 struct dfl_afu_dma_region *region)
93{
94 long npages = region->length >> PAGE_SHIFT;
95 struct device *dev = &pdata->dev->dev;
96
97 put_all_pages(region->pages, npages);
98 kfree(region->pages);
99 account_locked_vm(current->mm, npages, false);
100
101 dev_dbg(dev, "%ld pages unpinned\n", npages);
102}
103
104/**
105 * afu_dma_check_continuous_pages - check if pages are continuous
106 * @region: dma memory region
107 *
108 * Return true if pages of given dma memory region have continuous physical
109 * address, otherwise return false.
110 */
111static bool afu_dma_check_continuous_pages(struct dfl_afu_dma_region *region)
112{
113 int npages = region->length >> PAGE_SHIFT;
114 int i;
115
116 for (i = 0; i < npages - 1; i++)
117 if (page_to_pfn(region->pages[i]) + 1 !=
118 page_to_pfn(region->pages[i + 1]))
119 return false;
120
121 return true;
122}
123
124/**
125 * dma_region_check_iova - check if memory area is fully contained in the region
126 * @region: dma memory region
127 * @iova: address of the dma memory area
128 * @size: size of the dma memory area
129 *
130 * Compare the dma memory area defined by @iova and @size with given dma region.
131 * Return true if memory area is fully contained in the region, otherwise false.
132 */
133static bool dma_region_check_iova(struct dfl_afu_dma_region *region,
134 u64 iova, u64 size)
135{
136 if (!size && region->iova != iova)
137 return false;
138
139 return (region->iova <= iova) &&
140 (region->length + region->iova >= iova + size);
141}
142
143/**
144 * afu_dma_region_add - add given dma region to rbtree
145 * @pdata: feature device platform data
146 * @region: dma region to be added
147 *
148 * Return 0 for success, -EEXIST if dma region has already been added.
149 *
150 * Needs to be called with pdata->lock heold.
151 */
152static int afu_dma_region_add(struct dfl_feature_platform_data *pdata,
153 struct dfl_afu_dma_region *region)
154{
155 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
156 struct rb_node **new, *parent = NULL;
157
158 dev_dbg(&pdata->dev->dev, "add region (iova = %llx)\n",
159 (unsigned long long)region->iova);
160
161 new = &afu->dma_regions.rb_node;
162
163 while (*new) {
164 struct dfl_afu_dma_region *this;
165
166 this = container_of(*new, struct dfl_afu_dma_region, node);
167
168 parent = *new;
169
170 if (dma_region_check_iova(this, region->iova, region->length))
171 return -EEXIST;
172
173 if (region->iova < this->iova)
174 new = &((*new)->rb_left);
175 else if (region->iova > this->iova)
176 new = &((*new)->rb_right);
177 else
178 return -EEXIST;
179 }
180
181 rb_link_node(®ion->node, parent, new);
182 rb_insert_color(®ion->node, &afu->dma_regions);
183
184 return 0;
185}
186
187/**
188 * afu_dma_region_remove - remove given dma region from rbtree
189 * @pdata: feature device platform data
190 * @region: dma region to be removed
191 *
192 * Needs to be called with pdata->lock heold.
193 */
194static void afu_dma_region_remove(struct dfl_feature_platform_data *pdata,
195 struct dfl_afu_dma_region *region)
196{
197 struct dfl_afu *afu;
198
199 dev_dbg(&pdata->dev->dev, "del region (iova = %llx)\n",
200 (unsigned long long)region->iova);
201
202 afu = dfl_fpga_pdata_get_private(pdata);
203 rb_erase(®ion->node, &afu->dma_regions);
204}
205
206/**
207 * afu_dma_region_destroy - destroy all regions in rbtree
208 * @pdata: feature device platform data
209 *
210 * Needs to be called with pdata->lock heold.
211 */
212void afu_dma_region_destroy(struct dfl_feature_platform_data *pdata)
213{
214 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
215 struct rb_node *node = rb_first(&afu->dma_regions);
216 struct dfl_afu_dma_region *region;
217
218 while (node) {
219 region = container_of(node, struct dfl_afu_dma_region, node);
220
221 dev_dbg(&pdata->dev->dev, "del region (iova = %llx)\n",
222 (unsigned long long)region->iova);
223
224 rb_erase(node, &afu->dma_regions);
225
226 if (region->iova)
227 dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
228 region->iova, region->length,
229 DMA_BIDIRECTIONAL);
230
231 if (region->pages)
232 afu_dma_unpin_pages(pdata, region);
233
234 node = rb_next(node);
235 kfree(region);
236 }
237}
238
239/**
240 * afu_dma_region_find - find the dma region from rbtree based on iova and size
241 * @pdata: feature device platform data
242 * @iova: address of the dma memory area
243 * @size: size of the dma memory area
244 *
245 * It finds the dma region from the rbtree based on @iova and @size:
246 * - if @size == 0, it finds the dma region which starts from @iova
247 * - otherwise, it finds the dma region which fully contains
248 * [@iova, @iova+size)
249 * If nothing is matched returns NULL.
250 *
251 * Needs to be called with pdata->lock held.
252 */
253struct dfl_afu_dma_region *
254afu_dma_region_find(struct dfl_feature_platform_data *pdata, u64 iova, u64 size)
255{
256 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
257 struct rb_node *node = afu->dma_regions.rb_node;
258 struct device *dev = &pdata->dev->dev;
259
260 while (node) {
261 struct dfl_afu_dma_region *region;
262
263 region = container_of(node, struct dfl_afu_dma_region, node);
264
265 if (dma_region_check_iova(region, iova, size)) {
266 dev_dbg(dev, "find region (iova = %llx)\n",
267 (unsigned long long)region->iova);
268 return region;
269 }
270
271 if (iova < region->iova)
272 node = node->rb_left;
273 else if (iova > region->iova)
274 node = node->rb_right;
275 else
276 /* the iova region is not fully covered. */
277 break;
278 }
279
280 dev_dbg(dev, "region with iova %llx and size %llx is not found\n",
281 (unsigned long long)iova, (unsigned long long)size);
282
283 return NULL;
284}
285
286/**
287 * afu_dma_region_find_iova - find the dma region from rbtree by iova
288 * @pdata: feature device platform data
289 * @iova: address of the dma region
290 *
291 * Needs to be called with pdata->lock held.
292 */
293static struct dfl_afu_dma_region *
294afu_dma_region_find_iova(struct dfl_feature_platform_data *pdata, u64 iova)
295{
296 return afu_dma_region_find(pdata, iova, 0);
297}
298
299/**
300 * afu_dma_map_region - map memory region for dma
301 * @pdata: feature device platform data
302 * @user_addr: address of the memory region
303 * @length: size of the memory region
304 * @iova: pointer of iova address
305 *
306 * Map memory region defined by @user_addr and @length, and return dma address
307 * of the memory region via @iova.
308 * Return 0 for success, otherwise error code.
309 */
310int afu_dma_map_region(struct dfl_feature_platform_data *pdata,
311 u64 user_addr, u64 length, u64 *iova)
312{
313 struct dfl_afu_dma_region *region;
314 int ret;
315
316 /*
317 * Check Inputs, only accept page-aligned user memory region with
318 * valid length.
319 */
320 if (!PAGE_ALIGNED(user_addr) || !PAGE_ALIGNED(length) || !length)
321 return -EINVAL;
322
323 /* Check overflow */
324 if (user_addr + length < user_addr)
325 return -EINVAL;
326
327 if (!access_ok((void __user *)(unsigned long)user_addr,
328 length))
329 return -EINVAL;
330
331 region = kzalloc(sizeof(*region), GFP_KERNEL);
332 if (!region)
333 return -ENOMEM;
334
335 region->user_addr = user_addr;
336 region->length = length;
337
338 /* Pin the user memory region */
339 ret = afu_dma_pin_pages(pdata, region);
340 if (ret) {
341 dev_err(&pdata->dev->dev, "failed to pin memory region\n");
342 goto free_region;
343 }
344
345 /* Only accept continuous pages, return error else */
346 if (!afu_dma_check_continuous_pages(region)) {
347 dev_err(&pdata->dev->dev, "pages are not continuous\n");
348 ret = -EINVAL;
349 goto unpin_pages;
350 }
351
352 /* As pages are continuous then start to do DMA mapping */
353 region->iova = dma_map_page(dfl_fpga_pdata_to_parent(pdata),
354 region->pages[0], 0,
355 region->length,
356 DMA_BIDIRECTIONAL);
357 if (dma_mapping_error(dfl_fpga_pdata_to_parent(pdata), region->iova)) {
358 dev_err(&pdata->dev->dev, "failed to map for dma\n");
359 ret = -EFAULT;
360 goto unpin_pages;
361 }
362
363 *iova = region->iova;
364
365 mutex_lock(&pdata->lock);
366 ret = afu_dma_region_add(pdata, region);
367 mutex_unlock(&pdata->lock);
368 if (ret) {
369 dev_err(&pdata->dev->dev, "failed to add dma region\n");
370 goto unmap_dma;
371 }
372
373 return 0;
374
375unmap_dma:
376 dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
377 region->iova, region->length, DMA_BIDIRECTIONAL);
378unpin_pages:
379 afu_dma_unpin_pages(pdata, region);
380free_region:
381 kfree(region);
382 return ret;
383}
384
385/**
386 * afu_dma_unmap_region - unmap dma memory region
387 * @pdata: feature device platform data
388 * @iova: dma address of the region
389 *
390 * Unmap dma memory region based on @iova.
391 * Return 0 for success, otherwise error code.
392 */
393int afu_dma_unmap_region(struct dfl_feature_platform_data *pdata, u64 iova)
394{
395 struct dfl_afu_dma_region *region;
396
397 mutex_lock(&pdata->lock);
398 region = afu_dma_region_find_iova(pdata, iova);
399 if (!region) {
400 mutex_unlock(&pdata->lock);
401 return -EINVAL;
402 }
403
404 if (region->in_use) {
405 mutex_unlock(&pdata->lock);
406 return -EBUSY;
407 }
408
409 afu_dma_region_remove(pdata, region);
410 mutex_unlock(&pdata->lock);
411
412 dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
413 region->iova, region->length, DMA_BIDIRECTIONAL);
414 afu_dma_unpin_pages(pdata, region);
415 kfree(region);
416
417 return 0;
418}